Science.gov

Sample records for distance web service

  1. Web-Browsing Competencies of Pre-Service Adult Facilitators: Implications for Curriculum Transformation and Distance Learning

    ERIC Educational Resources Information Center

    Theresa, Ofoegbu; Ugwu, Agboeze Matthias; Ihebuzoaju, Anyanwu Joy; Uche, Asogwa

    2013-01-01

    The study investigated the Web-browsing competencies of pre-service adult facilitators in the southeast geopolitical zone of Nigeria. Survey design was adopted for the study. The population consists of all pre-service adult facilitators in all the federal universities in the southeast geopolitical zone of Nigeria. Accidental sampling technique was…

  2. Academic Library Web Sites and Distance Education: A Content Analysis

    ERIC Educational Resources Information Center

    Jurkowski, Odin L.

    2004-01-01

    This study focuses on college and university library services provided to distance education students and a content analysis of library Web sites. The Web sites chosen were categorized by the amount of distance education the institution offered to students. Results indicated that the size of the institution had the strongest correlation to library…

  3. Adding Interactivity to Web Based Distance Learning.

    ERIC Educational Resources Information Center

    Cafolla, Ralph; Knee, Richard

    Web Based Distance Learning (WBDL) is a form of distance learning based on providing instruction mainly on the World Wide Web. This paradigm has limitations, especially the lack of interactivity inherent in the Web. The purpose of this paper is to discuss some of the technologies the authors have used in their courses at Florida Atlantic…

  4. Chapter 59: Web Services

    NASA Astrophysics Data System (ADS)

    Graham, M. J.

    Web services are a cornerstone of the distributed computing infrastructure that the VO is built upon yet to the newcomer, they can appear to be a black art. This perception is not helped by the miasma of technobabble that pervades the subject and the seemingly impenetrable high priesthood of actual users. In truth, however, there is nothing conceptually difficult about web services (unsurprisingly any complexities will lie in the implementation details) nor indeed anything particularly new. A web service is a piece of software available over a network with a formal description of how it is called and what it returns that a computer can understand. Note that entities such as web servers, ftp servers and database servers do not generally qualify as they lack the standardized description of their inputs and outputs. There are prior technologies, such as RMI, CORBA, and DCOM, that have employed a similar approach but the success of web services lies predominantly in its use of standardized XML to provide a language-neutral way for representing data. In fact, the standardization goes further as web services are traditionally (or as traditionally as five years will allow) tied to a specific set of technologies (WSDL and SOAP conveyed using HTTP with an XML serialization). Alternative implementations are becoming increasingly common and we will cover some of these here. One important thing to remember in all of this, though, is that web services are meant for use by computers and not humans (unlike web pages) and this is why so much of it seems incomprehensible gobbledegook. In this chapter, we will start with an overview of the web services current in the VO and present a short guide on how to use and deploy a web service. We will then review the different approaches to web services, particularly REST and SOAP, and alternatives to XML as a data format. We will consider how web services can be formally described and discuss how advanced features such as security, state

  5. Web service performance script

    SciTech Connect

    2009-08-01

    This python script, available from ESRI and modified here, checks a server at specified intervals to ensure that web services remain up and running. If any are found to be off, they are automatically turned back on.

  6. Using Web-Based Distance Learning to Reduce Cultural Distance

    ERIC Educational Resources Information Center

    Wong, L. Fai; Trinidad, S. G.

    2004-01-01

    In recent years, Web-based distance learning (WBDL) systems have become a popular learning environment for many western learners. While it has been established as an effective learning alternative, WBDL is not flourishing in Hong Kong as expected. This paper proposes that this is because Hong Kong students are not trained to learn independently…

  7. Web Page Design in Distance Education

    ERIC Educational Resources Information Center

    Isman, Aytekin; Dabaj, Fahme; Gumus, Agah; Altinay, Fahriye; Altinay, Zehra

    2004-01-01

    Distance education is contemporary process of the education. It facilitates fast, easy delivery of information with its concrete hardware and software tools. The development of high technology, Internet and web-design delivering become impact of effective using as delivery system to the students. Within the global perspective, even the all work…

  8. Asynchronous OGC web services mechanisms

    NASA Astrophysics Data System (ADS)

    Min, Min; Di, Liping; Yu, Genong; Chen, Nengcheng; Gong, Jianya

    2008-12-01

    The Open Geospatial Consortium, Inc (OGC) Web Services (OWS) were initially primarily simple synchronous Web services based on the HTTP transport protocol, which is perfectly valid in the case of simple geoprocessing of simple data available from local sources. However, with the development of Web-based geospatial technologies, especially the development of the Sensor Web, a number of limitations have been identified with using HTTP-GET/POST binding in OGC OWS, which cannot meet the needs of asynchronous communication and operations between clients and services or in OGC services chain. Asynchronicity in Web services could be achieved in different ways. Callback pattern is widely supported in client asynchronous invocation. Message-based middleware often can be used together with the asynchronous invocation alternatives. Web Notification Service (WNS) is designed to provide asynchronous messagebased communication in OGC. This paper describes a mechanism for an asynchronous, message-based, event-driven, dynamic geospatial Web system based on OGC Web services. The addition of asynchronicity in OGC Web services has two components. One is the augmentation of OGC Web services with asynchronous message-based notification. The other is asynchronous OGC Web service orchestration based on BPEL.

  9. The EMBRACE web service collection

    PubMed Central

    Pettifer, Steve; Ison, Jon; Kalaš, Matúš; Thorne, Dave; McDermott, Philip; Jonassen, Inge; Liaquat, Ali; Fernández, José M.; Rodriguez, Jose M.; Partners, INB-; Pisano, David G.; Blanchet, Christophe; Uludag, Mahmut; Rice, Peter; Bartaseviciute, Edita; Rapacki, Kristoffer; Hekkelman, Maarten; Sand, Olivier; Stockinger, Heinz; Clegg, Andrew B.; Bongcam-Rudloff, Erik; Salzemann, Jean; Breton, Vincent; Attwood, Teresa K.; Cameron, Graham; Vriend, Gert

    2010-01-01

    The EMBRACE (European Model for Bioinformatics Research and Community Education) web service collection is the culmination of a 5-year project that set out to investigate issues involved in developing and deploying web services for use in the life sciences. The project concluded that in order for web services to achieve widespread adoption, standards must be defined for the choice of web service technology, for semantically annotating both service function and the data exchanged, and a mechanism for discovering services must be provided. Building on this, the project developed: EDAM, an ontology for describing life science web services; BioXSD, a schema for exchanging data between services; and a centralized registry (http://www.embraceregistry.net) that collects together around 1000 services developed by the consortium partners. This article presents the current status of the collection and its associated recommendations and standards definitions. PMID:20462862

  10. MedlinePlus Connect: Web Service

    MedlinePlus

    ... https://medlineplus.gov/connect/service.html MedlinePlus Connect: Web Service To use the sharing features on this ... if you implement MedlinePlus Connect by contacting us . Web Service Overview The parameters for the Web service ...

  11. Web Service: MedlinePlus

    MedlinePlus

    ... this page: https://medlineplus.gov/webservices.html MedlinePlus Web Service To use the sharing features on this ... please enable JavaScript. MedlinePlus offers a search-based Web service that provides access to MedlinePlus health topic ...

  12. RESTful Web Services at BNL

    SciTech Connect

    Casella, R.

    2011-06-14

    RESTful (REpresentational State Transfer) web services are an alternative implementation to SOAP/RPC web services in a client/server model. BNLs IT Division has started deploying RESTful Web Services for enterprise data retrieval and manipulation. Data is currently used by system administrators for tracking configuration information and as it is expanded will be used by Cyber Security for vulnerability management and as an aid to cyber investigations. This talk will describe the implementation and outstanding issues as well as some of the reasons for choosing RESTful over SOAP/RPC and future directions.

  13. Embracing a Customer Service Mindset: A Fresh Examination of Services for Distance Learners

    ERIC Educational Resources Information Center

    Steiner, Heidi

    2013-01-01

    Library literature and blogs frequently discuss customer service and user experience in physical libraries and Web sites, but little is said about this mentality toward services for distance learners specifically. This paper takes customer service best practices from well-known thinkers of the business world and makes connections to services for…

  14. Semantic Web for Manufacturing Web Services

    SciTech Connect

    Kulvatunyou, Boonserm; Ivezic, Nenad

    2002-06-01

    As markets become unexpectedly turbulent with a shortened product life cycle and a power shift towards buyers, the need for methods to rapidly and cost-effectively develop products, production facilities and supporting software is becoming urgent. The use of a virtual enterprise plays a vital role in surviving turbulent markets. However, its success requires reliable and large-scale interoperation among trading partners via a semantic web of trading partners' services whose properties, capabilities, and interfaces are encoded in an unambiguous as well as computer-understandable form. This paper demonstrates a promising approach to integration and interoperation between a design house and a manufacturer by developing semantic web services for business and engineering transactions. To this end, detailed activity and information flow diagrams are developed, in which the two trading partners exchange messages and documents. The properties and capabilities of the manufacturer sites are defined using DARPA Agent Markup Language (DAML) ontology definition language. The prototype development of semantic webs shows that enterprises can widely interoperate in an unambiguous and autonomous manner; hence, virtual enterprise is realizable at a low cost.

  15. Web-Based Communications, the Internet, and Distance Education. Readings in Distance Education, Number 7.

    ERIC Educational Resources Information Center

    Moore, Michael G., Ed.; Cozine, Geoffrey T., Ed.

    This book brings together a selection of articles published in "The American Journal of Distance Education" that are related to Web-based delivery of distance education. Articles include: "Performance and Perceptions of Distance Learners in Cyberspace" (Peter Navarro and Judy Shoemaker); "Distance Education for Dentists: Improving the Quality of…

  16. WEBCAP: Web Scheduler for Distance Learning Multimedia Documents with Web Workload Considerations

    ERIC Educational Resources Information Center

    Habib, Sami; Safar, Maytham

    2008-01-01

    In many web applications, such as the distance learning, the frequency of refreshing multimedia web documents places a heavy burden on the WWW resources. Moreover, the updated web documents may encounter inordinate delays, which make it difficult to retrieve web documents in time. Here, we present an Internet tool called WEBCAP that can schedule…

  17. Semantic Search of Web Services

    ERIC Educational Resources Information Center

    Hao, Ke

    2013-01-01

    This dissertation addresses semantic search of Web services using natural language processing. We first survey various existing approaches, focusing on the fact that the expensive costs of current semantic annotation frameworks result in limited use of semantic search for large scale applications. We then propose a vector space model based service…

  18. APPRIS WebServer and WebServices

    PubMed Central

    Rodriguez, Jose Manuel; Carro, Angel; Valencia, Alfonso; Tress, Michael L.

    2015-01-01

    This paper introduces the APPRIS WebServer (http://appris.bioinfo.cnio.es) and WebServices (http://apprisws.bioinfo.cnio.es). Both the web servers and the web services are based around the APPRIS Database, a database that presently houses annotations of splice isoforms for five different vertebrate genomes. The APPRIS WebServer and WebServices provide access to the computational methods implemented in the APPRIS Database, while the APPRIS WebServices also allows retrieval of the annotations. The APPRIS WebServer and WebServices annotate splice isoforms with protein structural and functional features, and with data from cross-species alignments. In addition they can use the annotations of structure, function and conservation to select a single reference isoform for each protein-coding gene (the principal protein isoform). APPRIS principal isoforms have been shown to agree overwhelmingly with the main protein isoform detected in proteomics experiments. The APPRIS WebServer allows for the annotation of splice isoforms for individual genes, and provides a range of visual representations and tools to allow researchers to identify the likely effect of splicing events. The APPRIS WebServices permit users to generate annotations automatically in high throughput mode and to interrogate the annotations in the APPRIS Database. The APPRIS WebServices have been implemented using REST architecture to be flexible, modular and automatic. PMID:25990727

  19. Earth Science Mining Web Services

    NASA Astrophysics Data System (ADS)

    Pham, L. B.; Lynnes, C. S.; Hegde, M.; Graves, S.; Ramachandran, R.; Maskey, M.; Keiser, K.

    2008-12-01

    To allow scientists further capabilities in the area of data mining and web services, the Goddard Earth Sciences Data and Information Services Center (GES DISC) and researchers at the University of Alabama in Huntsville (UAH) have developed a system to mine data at the source without the need of network transfers. The system has been constructed by linking together several pre-existing technologies: the Simple Scalable Script-based Science Processor for Measurements (S4PM), a processing engine at the GES DISC; the Algorithm Development and Mining (ADaM) system, a data mining toolkit from UAH that can be configured in a variety of ways to create customized mining processes; ActiveBPEL, a workflow execution engine based on BPEL (Business Process Execution Language); XBaya, a graphical workflow composer; and the EOS Clearinghouse (ECHO). XBaya is used to construct an analysis workflow at UAH using ADaM components, which are also installed remotely at the GES DISC, wrapped as Web Services. The S4PM processing engine searches ECHO for data using space-time criteria, staging them to cache, allowing the ActiveBPEL engine to remotely orchestrates the processing workflow within S4PM. As mining is completed, the output is placed in an FTP holding area for the end user. The goals are to give users control over the data they want to process, while mining data at the data source using the server's resources rather than transferring the full volume over the internet. These diverse technologies have been infused into a functioning, distributed system with only minor changes to the underlying technologies. The key to this infusion is the loosely coupled, Web- Services based architecture: All of the participating components are accessible (one way or another) through (Simple Object Access Protocol) SOAP-based Web Services.

  20. Earth Science Mining Web Services

    NASA Technical Reports Server (NTRS)

    Pham, Long; Lynnes, Christopher; Hegde, Mahabaleshwa; Graves, Sara; Ramachandran, Rahul; Maskey, Manil; Keiser, Ken

    2008-01-01

    To allow scientists further capabilities in the area of data mining and web services, the Goddard Earth Sciences Data and Information Services Center (GES DISC) and researchers at the University of Alabama in Huntsville (UAH) have developed a system to mine data at the source without the need of network transfers. The system has been constructed by linking together several pre-existing technologies: the Simple Scalable Script-based Science Processor for Measurements (S4PM), a processing engine at he GES DISC; the Algorithm Development and Mining (ADaM) system, a data mining toolkit from UAH that can be configured in a variety of ways to create customized mining processes; ActiveBPEL, a workflow execution engine based on BPEL (Business Process Execution Language); XBaya, a graphical workflow composer; and the EOS Clearinghouse (ECHO). XBaya is used to construct an analysis workflow at UAH using ADam components, which are also installed remotely at the GES DISC, wrapped as Web Services. The S4PM processing engine searches ECHO for data using space-time criteria, staging them to cache, allowing the ActiveBPEL engine to remotely orchestras the processing workflow within S4PM. As mining is completed, the output is placed in an FTP holding area for the end user. The goals are to give users control over the data they want to process, while mining data at the data source using the server's resources rather than transferring the full volume over the internet. These diverse technologies have been infused into a functioning, distributed system with only minor changes to the underlying technologies. The key to the infusion is the loosely coupled, Web-Services based architecture: All of the participating components are accessible (one way or another) through (Simple Object Access Protocol) SOAP-based Web Services.

  1. Faculty Perceptions of Web-Based Distance Education in Agriculture.

    ERIC Educational Resources Information Center

    Born, Kevin A.; Miller, Greg

    1999-01-01

    A survey of 42 agronomy faculty showed their perceptions of Web-based distance education were higher when they were familiar with the master of science in agronomy program or had viewed a lesson. Their concerns included the value and rigor of Web-based degree programs and the effectiveness of online student-teacher interaction. (SK)

  2. Effective Web Design and Core Communication Issues: The Missing Components in Web-Based Distance Education.

    ERIC Educational Resources Information Center

    Burch, Randall O.

    2001-01-01

    Discussion of Web-based distance education focuses on communication issues. Highlights include Internet communications; components of a Web site, including site architecture, user interface, information delivery method, and mode of feedback; elements of Web design, including conceptual design, sensory design, and reactive design; and a Web…

  3. Socialization of Distance Education: The Web as Enabler.

    ERIC Educational Resources Information Center

    Parker, Drew; Rossner-Merrill, Vivian

    The World Wide Web has allowed the delineation of distance versus place-based education to become a spectrum rather than a binary choice. This paper discusses a novel format of distance education, called "Virtual Seminars," and its relation to issues within cognitive flexibility theory. Virtual seminars are interactive courses offered over the…

  4. Web-based Altimeter Service

    NASA Astrophysics Data System (ADS)

    Callahan, P. S.; Wilson, B. D.; Xing, Z.; Raskin, R. G.

    2010-12-01

    We have developed a web-based system to allow updating and subsetting of TOPEX data. The Altimeter Service will be operated by PODAAC along with their other provision of oceanographic data. The Service could be easily expanded to other mission data. An Altimeter Service is crucial to the improvement and expanded use of altimeter data. A service is necessary for altimetry because the result of most interest - sea surface height anomaly (SSHA) - is composed of several components that are updated individually and irregularly by specialized experts. This makes it difficult for projects to provide the most up-to-date products. Some components are the subject of ongoing research, so the ability for investigators to make products for comparison or sharing is important. The service will allow investigators/producers to get their component models or processing into widespread use much more quickly. For coastal altimetry, the ability to subset the data to the area of interest and insert specialized models (e.g., tides) or data processing results is crucial. A key part of the Altimeter Service is having data producers provide updated or local models and data. In order for this to succeed, producers need to register their products with the Altimeter Service and to provide the product in a form consistent with the service update methods. We will describe the capabilities of the web service and the methods for providing new components. Currently the Service is providing TOPEX GDRs with Retracking (RGDRs) in netCDF format that has been coordinated with Jason data. Users can add new orbits, tide models, gridded geophysical fields such as mean sea surface, and along-track corrections as they become available and are installed by PODAAC. The updated fields are inserted into the netCDF files while the previous values are retained for comparison. The Service will also generate SSH and SSHA. In addition, the Service showcases a feature that plots any variable from files in netCDF. The

  5. Technical Services and the World Wide Web.

    ERIC Educational Resources Information Center

    Scheschy, Virginia M.

    The World Wide Web and browsers such as Netscape and Mosaic have simplified access to electronic resources. Today, technical services librarians can share in the wealth of information available on the Web. One of the premier Web sites for acquisitions librarians is AcqWeb, a cousin of the AcqNet listserv. In addition to interesting news items,…

  6. Similarity Based Semantic Web Service Match

    NASA Astrophysics Data System (ADS)

    Peng, Hui; Niu, Wenjia; Huang, Ronghuai

    Semantic web service discovery aims at returning the most matching advertised services to the service requester by comparing the semantic of the request service with an advertised service. The semantic of a web service are described in terms of inputs, outputs, preconditions and results in Ontology Web Language for Service (OWL-S) which formalized by W3C. In this paper we proposed an algorithm to calculate the semantic similarity of two services by weighted averaging their inputs and outputs similarities. Case study and applications show the effectiveness of our algorithm in service match.

  7. The Organizational Role of Web Services

    ERIC Educational Resources Information Center

    Mitchell, Erik

    2011-01-01

    The workload of Web librarians is already split between Web-related and other library tasks. But today's technological environment has created new implications for existing services and new demands for staff time. It is time to reconsider how libraries can best allocate resources to provide effective Web services. Delivering high-quality services…

  8. Finding Categories and Keywords in Web Services

    NASA Astrophysics Data System (ADS)

    Kop, Christian; Gälle, Doris; Mayr, Heinrich C.

    Nowadays web services are a common way to integrate functionality in an information system, but most of the time it is very difficult to find an appropriate service. If users of web service engines do not exactly know what they want, they often browse through categories and search with keywords. This, however, depends on the knowledge of the web service owner and his/her willingness to assign such keywords. This chapter gives a proposal to provide the user with candidates for keywords and categories which are derived directly from the web service specification itself.

  9. Distance Learning Library Services in Ugandan Universities

    ERIC Educational Resources Information Center

    Mayende, Jackline Estomihi Kiwelu; Obura, Constant Okello

    2013-01-01

    The study carried out at Makerere University and Uganda Martyrs University in 2010 aimed at providing strategies for enhanced distance learning library services in terms of convenience and adequacy. The study adopted a cross sectional descriptive survey design. The study revealed services provided in branch libraries in Ugandan universities were…

  10. Discovery and Classification of Bioinformatics Web Services

    SciTech Connect

    Rocco, D; Critchlow, T

    2002-09-02

    The transition of the World Wide Web from a paradigm of static Web pages to one of dynamic Web services provides new and exciting opportunities for bioinformatics with respect to data dissemination, transformation, and integration. However, the rapid growth of bioinformatics services, coupled with non-standardized interfaces, diminish the potential that these Web services offer. To face this challenge, we examine the notion of a Web service class that defines the functionality provided by a collection of interfaces. These descriptions are an integral part of a larger framework that can be used to discover, classify, and wrapWeb services automatically. We discuss how this framework can be used in the context of the proliferation of sites offering BLAST sequence alignment services for specialized data sets.

  11. Storage Manager and File Transfer Web Services

    SciTech Connect

    William A Watson III; Ying Chen; Jie Chen; Walt Akers

    2002-07-01

    Web services are emerging as an interesting mechanism for a wide range of grid services, particularly those focused upon information services and control. When coupled with efficient data transfer services, they provide a powerful mechanism for building a flexible, open, extensible data grid for science applications. In this paper we present our prototype work on a Java Storage Resource Manager (JSRM) web service and a Java Reliable File Transfer (JRFT) web service. A java client (Grid File Manager) on top of JSRM and is developed to demonstrate the capabilities of these web services. The purpose of this work is to show the extent to which SOAP based web services are an appropriate direction for building a grid-wide data management system, and eventually grid-based portals.

  12. An Automatic Web Service Composition Framework Using QoS-Based Web Service Ranking Algorithm

    PubMed Central

    Mallayya, Deivamani; Ramachandran, Baskaran; Viswanathan, Suganya

    2015-01-01

    Web service has become the technology of choice for service oriented computing to meet the interoperability demands in web applications. In the Internet era, the exponential addition of web services nominates the “quality of service” as essential parameter in discriminating the web services. In this paper, a user preference based web service ranking (UPWSR) algorithm is proposed to rank web services based on user preferences and QoS aspect of the web service. When the user's request cannot be fulfilled by a single atomic service, several existing services should be composed and delivered as a composition. The proposed framework allows the user to specify the local and global constraints for composite web services which improves flexibility. UPWSR algorithm identifies best fit services for each task in the user request and, by choosing the number of candidate services for each task, reduces the time to generate the composition plans. To tackle the problem of web service composition, QoS aware automatic web service composition (QAWSC) algorithm proposed in this paper is based on the QoS aspects of the web services and user preferences. The proposed framework allows user to provide feedback about the composite service which improves the reputation of the services. PMID:26504894

  13. An active registry for bioinformatics web services

    PubMed Central

    Pettifer, S.; Thorne, D.; McDermott, P.; Attwood, T.; Baran, J.; Bryne, J. C.; Hupponen, T.; Mowbray, D.; Vriend, G.

    2009-01-01

    Summary: The EMBRACE Registry is a web portal that collects and monitors web services according to test scripts provided by the their administrators. Users are able to search for, rank and annotate services, enabling them to select the most appropriate working service for inclusion in their bioinformatics analysis tasks. Availability and implementation: Web site implemented with PHP, Python, MySQL and Apache, with all major browsers supported. (www.embraceregistry.net) Contact: steve.pettifer@manchester.ac.uk PMID:19460889

  14. Web-Based Distance Learning: A Tool for Change.

    ERIC Educational Resources Information Center

    Clyatt, Bob

    1999-01-01

    Defines advanced Web-based distance learning (WBDL); examines some of its benefits, including cost effectiveness, asynchronous training, simple evaluation, and effective use of time; offers suggestions for appropriate applications of WBDL; and describes ways organizations are beginning to use WBDL, including training clients and continuing…

  15. Enriching the Web Processing Service

    NASA Astrophysics Data System (ADS)

    Wosniok, Christoph; Bensmann, Felix; Wössner, Roman; Kohlus, Jörn; Roosmann, Rainer; Heidmann, Carsten; Lehfeldt, Rainer

    2014-05-01

    The OGC Web Processing Service (WPS) provides a standard for implementing geospatial processes in service-oriented networks. In its current version 1.0.0 it allocates the operations GetCapabilities, DescribeProcess and Execute, which can be used to offer custom processes based on single or multiple sub-processes. A large range of ready to use fine granular, fundamental geospatial processes have been developed by the GIS-community in the past. However, modern use cases or whole workflow processes demand specifications of lifecycle management and service orchestration. Orchestrating smaller sub-processes is a task towards interoperability; a comprehensive documentation by using appropriate metadata is also required. Though different approaches were tested in the past, developing complex WPS applications still requires programming skills, knowledge about software libraries in use and a lot of effort for integration. Our toolset RichWPS aims at providing a better overall experience by setting up two major components. The RichWPS ModelBuilder enables the graphics-aided design of workflow processes based on existing local and distributed processes and geospatial services. Once tested by the RichWPS Server, a composition can be deployed for production use on the RichWPS Server. The ModelBuilder obtains necessary processes and services from a directory service, the RichWPS semantic proxy. It manages the lifecycle and is able to visualize results and debugging-information. One aim will be to generate reproducible results; the workflow should be documented by metadata that can be integrated in Spatial Data Infrastructures. The RichWPS Server provides a set of interfaces to the ModelBuilder for, among others, testing composed workflow sequences, estimating their performance and to publish them as common processes. Therefore the server is oriented towards the upcoming WPS 2.0 standard and its ability to transactionally deploy and undeploy processes making use of a WPS

  16. Space Physics Data Facility Web Services

    NASA Technical Reports Server (NTRS)

    Candey, Robert M.; Harris, Bernard T.; Chimiak, Reine A.

    2005-01-01

    The Space Physics Data Facility (SPDF) Web services provides a distributed programming interface to a portion of the SPDF software. (A general description of Web services is available at http://www.w3.org/ and in many current software-engineering texts and articles focused on distributed programming.) The SPDF Web services distributed programming interface enables additional collaboration and integration of the SPDF software system with other software systems, in furtherance of the SPDF mission to lead collaborative efforts in the collection and utilization of space physics data and mathematical models. This programming interface conforms to all applicable Web services specifications of the World Wide Web Consortium. The interface is specified by a Web Services Description Language (WSDL) file. The SPDF Web services software consists of the following components: 1) A server program for implementation of the Web services; and 2) A software developer s kit that consists of a WSDL file, a less formal description of the interface, a Java class library (which further eases development of Java-based client software), and Java source code for an example client program that illustrates the use of the interface.

  17. Identifying orthoimages in Web Map Services

    NASA Astrophysics Data System (ADS)

    Florczyk, A. J.; Nogueras-Iso, J.; Zarazaga-Soria, F. J.; Béjar, R.

    2012-10-01

    Orthoimages are essential in many Web applications to facilitate the background context that helps to understand other georeferenced information. Catalogues and service registries of Spatial Data Infrastructures do not necessarily register all the services providing access to imagery data on the Web, and it is not easy to automatically identify whether the data offered by a Web service are directly imagery data or not. This work presents a method for an automatic detection of the orthoimage layers offered by Web Map Services. The method combines two types of heuristics. The first one consists in analysing the text in the capabilities document. The second type is content-based heuristics, which analyse the content offered by the Web Map Service layers. These heuristics gather and analyse the colour features of a sample collection of image fragments that represent the offered content. An experiment has been performed over a set of Web Map Service layers, which have been fetched from a repository of capabilities documents gathered from the Web. This has proven the efficiency of the method (precision of 87% and recall of 60%). This functionality has been offered as a Web Processing Service, and it has been integrated within the Virtual Spain project to provide a catalogue of orthoimages and build realistic 3D views.

  18. Enhancing UCSF Chimera through web services.

    PubMed

    Huang, Conrad C; Meng, Elaine C; Morris, John H; Pettersen, Eric F; Ferrin, Thomas E

    2014-07-01

    Integrating access to web services with desktop applications allows for an expanded set of application features, including performing computationally intensive tasks and convenient searches of databases. We describe how we have enhanced UCSF Chimera (http://www.rbvi.ucsf.edu/chimera/), a program for the interactive visualization and analysis of molecular structures and related data, through the addition of several web services (http://www.rbvi.ucsf.edu/chimera/docs/webservices.html). By streamlining access to web services, including the entire job submission, monitoring and retrieval process, Chimera makes it simpler for users to focus on their science projects rather than data manipulation. Chimera uses Opal, a toolkit for wrapping scientific applications as web services, to provide scalable and transparent access to several popular software packages. We illustrate Chimera's use of web services with an example workflow that interleaves use of these services with interactive manipulation of molecular sequences and structures, and we provide an example Python program to demonstrate how easily Opal-based web services can be accessed from within an application. Web server availability: http://webservices.rbvi.ucsf.edu/opal2/dashboard?command=serviceList.

  19. Efficient Web Services Policy Combination

    NASA Technical Reports Server (NTRS)

    Vatan, Farrokh; Harman, Joseph G.

    2010-01-01

    Large-scale Web security systems usually involve cooperation between domains with non-identical policies. The network management and Web communication software used by the different organizations presents a stumbling block. Many of the tools used by the various divisions do not have the ability to communicate network management data with each other. At best, this means that manual human intervention into the communication protocols used at various network routers and endpoints is required. Developing practical, sound, and automated ways to compose policies to bridge these differences is a long-standing problem. One of the key subtleties is the need to deal with inconsistencies and defaults where one organization proposes a rule on a particular feature, and another has a different rule or expresses no rule. A general approach is to assign priorities to rules and observe the rules with the highest priorities when there are conflicts. The present methods have inherent inefficiency, which heavily restrict their practical applications. A new, efficient algorithm combines policies utilized for Web services. The method is based on an algorithm that allows an automatic and scalable composition of security policies between multiple organizations. It is based on defeasible policy composition, a promising approach for finding conflicts and resolving priorities between rules. In the general case, policy negotiation is an intractable problem. A promising method, suggested in the literature, is when policies are represented in defeasible logic, and composition is based on rules for non-monotonic inference. In this system, policy writers construct metapolicies describing both the policy that they wish to enforce and annotations describing their composition preferences. These annotations can indicate whether certain policy assertions are required by the policy writer or, if not, under what circumstances the policy writer is willing to compromise and allow other assertions to take

  20. Preservice Mathematics Teachers' Views on Distance Education and Their Web Pedagogical Content Knowledge

    ERIC Educational Resources Information Center

    Cagirgan Gulten, Dilek

    2013-01-01

    This research aims to investigate primary preservice mathematics teachers' views on distance education and web pedagogical content knowledge in terms of the subscales of general web, communicative web, pedagogical web, web pedagogical content and attitude towards web based instruction. The research was conducted with 46 senior students in the…

  1. Between Shots TRANSP Web Service

    NASA Astrophysics Data System (ADS)

    Feibush, E.; Andre, R.; Ludescher-Furth, C.; Kaye, S.; McCune, D.

    2008-11-01

    Running TRANSP between NSTX shots requires rapid data preparation and job submittal. A web service with a graphical user interface and data visualization has been developed to meet these goals. The underlying data preparation system has a command line interface written in Python and runs on a PPPL compute server. The display client is a Java program (ElVis) that sends requests to the data preparation system. As the run data is prepared, graphs are created and sent to the client for display. Flux surface plots are displayed and animated over time. The most commonly used control options are implemented in the UI as buttons and text fields. A time slice or time dependent run can be prepared. The command line interface is available in the client program for expert users to apply advanced settings, to prototype new UI buttons, and to run scripts. The client program contains a simple text editor for modifying the TRANSP namelist. When data preparation is complete the run is submitted to the TRANSP production system. The initial version has been deployed and is being tested in the control room setting. Results will be discussed in the poster presentation. Work performed at PPPL under the auspices of U.S. DOE Contract DE-AC02-76CH03073.

  2. Transimulation - protein biosynthesis web service.

    PubMed

    Siwiak, Marlena; Zielenkiewicz, Piotr

    2013-01-01

    Although translation is the key step during gene expression, it remains poorly characterized at the level of individual genes. For this reason, we developed Transimulation - a web service measuring translational activity of genes in three model organisms: Escherichia coli, Saccharomyces cerevisiae and Homo sapiens. The calculations are based on our previous computational model of translation and experimental data sets. Transimulation quantifies mean translation initiation and elongation time (expressed in SI units), and the number of proteins produced per transcript. It also approximates the number of ribosomes that typically occupy a transcript during translation, and simulates their propagation. The simulation of ribosomes' movement is interactive and allows modifying the coding sequence on the fly. It also enables uploading any coding sequence and simulating its translation in one of three model organisms. In such a case, ribosomes propagate according to mean codon elongation times of the host organism, which may prove useful for heterologous expression. Transimulation was used to examine evolutionary conservation of translational parameters of orthologous genes. Transimulation may be accessed at http://nexus.ibb.waw.pl/Transimulation (requires Java version 1.7 or higher). Its manual and source code, distributed under the GPL-2.0 license, is freely available at the website.

  3. Online Information Services. Caught in the Web?

    ERIC Educational Resources Information Center

    Green, Tim

    1995-01-01

    Provides brief reviews of the sites for several online services of the World Wide Web; the Web as a marketing tool and other aspects of interest to information professionals are highlighted. A sidebar presents information on accessing Internet locations, graphics, online forms, Telnet, saving, printing, mailing, and searching. (AEF)

  4. Domain-specific Web Service Discovery with Service Class Descriptions

    SciTech Connect

    Rocco, D; Caverlee, J; Liu, L; Critchlow, T J

    2005-02-14

    This paper presents DynaBot, a domain-specific web service discovery system. The core idea of the DynaBot service discovery system is to use domain-specific service class descriptions powered by an intelligent Deep Web crawler. In contrast to current registry-based service discovery systems--like the several available UDDI registries--DynaBot promotes focused crawling of the Deep Web of services and discovers candidate services that are relevant to the domain of interest. It uses intelligent filtering algorithms to match services found by focused crawling with the domain-specific service class descriptions. We demonstrate the capability of DynaBot through the BLAST service discovery scenario and describe our initial experience with DynaBot.

  5. CB-EMIS WEB SERVICE SOFTWARE

    2007-01-01

    This software provides CB-EMIS data to remote devices using a secure internet connection. The CB-EMIS Web Service filters and repackages data in a form suitable for resource limited devices such as a cell phone. Data transmission is filtered based on a user's authentical level. The web services acts as intermediary so that no direct connection is possible between the internet and the CB-EMIS server software.

  6. Using ESO Reflex with Web Services

    NASA Astrophysics Data System (ADS)

    Järveläinen, P.; Savolainen, V.; Oittinen, T.; Maisala, S.; Ullgrén, M. Hook, R.

    2008-08-01

    ESO Reflex is a prototype graphical workflow system, based on Taverna, and primarily intended to be a flexible way of running ESO data reduction recipes along with other legacy applications and user-written tools. ESO Reflex can also readily use the Taverna Web Services features that are based on the Apache Axis SOAP implementation. Taverna is a general purpose Web Service client, and requires no programming to use such services. However, Taverna also has some restrictions: for example, no numerical types such integers. In addition the preferred binding style is document/literal wrapped, but most astronomical services publish the Axis default WSDL using RPC/encoded style. Despite these minor limitations we have created simple but very promising test VO workflow using the Sesame name resolver service at CDS Strasbourg, the Hubble SIAP server at the Multi-Mission Archive at Space Telescope (MAST) and the WESIX image cataloging and catalogue cross-referencing service at the University of Pittsburgh. ESO Reflex can also pass files and URIs via the PLASTIC protocol to visualisation tools and has its own viewer for VOTables. We picked these three Web Services to try to set up a realistic and useful ESO Reflex workflow. They also demonstrate ESO Reflex abilities to use many kind of Web Services because each of them requires a different interface. We describe each of these services in turn and comment on how it was used

  7. Semantic Web Service Composition in Social Environments

    NASA Astrophysics Data System (ADS)

    Kuter, Ugur; Golbeck, Jennifer

    This paper describes how to generate compositions of semantic Web services using social trust information from user ratings of the services. We present a taxonomy of features, such as interoperability, availability, privacy, security, and others. We describe a way to compute social trust in OWL-S style semantic Web services. Our formalism exploits the users' ratings of the services and execution characteristics of those services. We describe our service-composition algorithm, called Trusty, that is based on this formalism. We discuss the formal properties of Trusty and our implementation of the algorithm. We present our experiments in which we compared Trusty with SHOP2, a well-known AI planning algorithm that has been successfully used for OWL-S style service composition. Our results demonstrate that Trusty generates more trustworthy compositions than SHOP2.

  8. A Method for Transforming Existing Web Service Descriptions into an Enhanced Semantic Web Service Framework

    NASA Astrophysics Data System (ADS)

    Du, Xiaofeng; Song, William; Munro, Malcolm

    Web Services as a new distributed system technology has been widely adopted by industries in the areas, such as enterprise application integration (EAI), business process management (BPM), and virtual organisation (VO). However, lack of semantics in the current Web Service standards has been a major barrier in service discovery and composition. In this chapter, we propose an enhanced context-based semantic service description framework (CbSSDF+) that tackles the problem and improves the flexibility of service discovery and the correctness of generated composite services. We also provide an agile transformation method to demonstrate how the various formats of Web Service descriptions on the Web can be managed and renovated step by step into CbSSDF+ based service description without large amount of engineering work. At the end of the chapter, we evaluate the applicability of the transformation method and the effectiveness of CbSSDF+ through a series of experiments.

  9. Web Service Architecture Framework for Embedded Devices

    ERIC Educational Resources Information Center

    Yanzick, Paul David

    2009-01-01

    The use of Service Oriented Architectures, namely web services, has become a widely adopted method for transfer of data between systems across the Internet as well as the Enterprise. Adopting a similar approach to embedded devices is also starting to emerge as personal devices and sensor networks are becoming more common in the industry. This…

  10. New Interfaces to Web Documents and Services

    NASA Technical Reports Server (NTRS)

    Carlisle, W. H.

    1996-01-01

    This paper reports on investigations into how to extend capabilities of the Virtual Research Center (VRC) for NASA's Advanced Concepts Office. The work was performed as part of NASA's 1996 Summer Faculty Fellowship program, and involved research into and prototype development of software components that provide documents and services for the World Wide Web (WWW). The WWW has become a de-facto standard for sharing resources over the internet, primarily because web browsers are freely available for the most common hardware platforms and their operating systems. As a consequence of the popularity of the internet, tools, and techniques associated with web browsers are changing rapidly. New capabilities are offered by companies that support web browsers in order to achieve or remain a dominant participant in internet services. Because a goal of the VRC is to build an environment for NASA centers, universities, and industrial partners to share information associated with Advanced Concepts Office activities, the VRC tracks new techniques and services associated with the web in order to determine the their usefulness for distributed and collaborative engineering research activities. Most recently, Java has emerged as a new tool for providing internet services. Because the major web browser providers have decided to include Java in their software, investigations into Java were conducted this summer.

  11. Opal web services for biomedical applications

    PubMed Central

    Ren, Jingyuan; Williams, Nadya; Clementi, Luca; Krishnan, Sriram; Li, Wilfred W.

    2010-01-01

    Biomedical applications have become increasingly complex, and they often require large-scale high-performance computing resources with a large number of processors and memory. The complexity of application deployment and the advances in cluster, grid and cloud computing require new modes of support for biomedical research. Scientific Software as a Service (sSaaS) enables scalable and transparent access to biomedical applications through simple standards-based Web interfaces. Towards this end, we built a production web server (http://ws.nbcr.net) in August 2007 to support the bioinformatics application called MEME. The server has grown since to include docking analysis with AutoDock and AutoDock Vina, electrostatic calculations using PDB2PQR and APBS, and off-target analysis using SMAP. All the applications on the servers are powered by Opal, a toolkit that allows users to wrap scientific applications easily as web services without any modification to the scientific codes, by writing simple XML configuration files. Opal allows both web forms-based access and programmatic access of all our applications. The Opal toolkit currently supports SOAP-based Web service access to a number of popular applications from the National Biomedical Computation Resource (NBCR) and affiliated collaborative and service projects. In addition, Opal’s programmatic access capability allows our applications to be accessed through many workflow tools, including Vision, Kepler, Nimrod/K and VisTrails. From mid-August 2007 to the end of 2009, we have successfully executed 239 814 jobs. The number of successfully executed jobs more than doubled from 205 to 411 per day between 2008 and 2009. The Opal-enabled service model is useful for a wide range of applications. It provides for interoperation with other applications with Web Service interfaces, and allows application developers to focus on the scientific tool and workflow development. Web server availability: http://ws.nbcr.net. PMID

  12. Opal web services for biomedical applications.

    PubMed

    Ren, Jingyuan; Williams, Nadya; Clementi, Luca; Krishnan, Sriram; Li, Wilfred W

    2010-07-01

    Biomedical applications have become increasingly complex, and they often require large-scale high-performance computing resources with a large number of processors and memory. The complexity of application deployment and the advances in cluster, grid and cloud computing require new modes of support for biomedical research. Scientific Software as a Service (sSaaS) enables scalable and transparent access to biomedical applications through simple standards-based Web interfaces. Towards this end, we built a production web server (http://ws.nbcr.net) in August 2007 to support the bioinformatics application called MEME. The server has grown since to include docking analysis with AutoDock and AutoDock Vina, electrostatic calculations using PDB2PQR and APBS, and off-target analysis using SMAP. All the applications on the servers are powered by Opal, a toolkit that allows users to wrap scientific applications easily as web services without any modification to the scientific codes, by writing simple XML configuration files. Opal allows both web forms-based access and programmatic access of all our applications. The Opal toolkit currently supports SOAP-based Web service access to a number of popular applications from the National Biomedical Computation Resource (NBCR) and affiliated collaborative and service projects. In addition, Opal's programmatic access capability allows our applications to be accessed through many workflow tools, including Vision, Kepler, Nimrod/K and VisTrails. From mid-August 2007 to the end of 2009, we have successfully executed 239,814 jobs. The number of successfully executed jobs more than doubled from 205 to 411 per day between 2008 and 2009. The Opal-enabled service model is useful for a wide range of applications. It provides for interoperation with other applications with Web Service interfaces, and allows application developers to focus on the scientific tool and workflow development. Web server availability: http://ws.nbcr.net.

  13. A Strategic Model of Trust Management in Web Services

    NASA Astrophysics Data System (ADS)

    Sun, Junqing; Sun, Zhaohao; Li, Yuanzhe; Zhao, Shuliang

    This article examines trust and trust management in web services and proposes a multiagent model of trust relationship in web services. It looks at the hierarchical structure of trust management in web services and proposes a strategic model of trust management in web services. The proposed approach in this article will facilitate research and development of trust management in e-commerce, web services and social networking.

  14. Web Services Security Problem in Service-oriented Architecture

    NASA Astrophysics Data System (ADS)

    Yue, Hua; Tao, Xu

    With the development and universal application of SOA technology, security issues of Web services based on heterogeneous platform have become increasingly prominent. This article first introduces two security different solutions of Microsoft Net, Apache Axis platform. At the same time for the security issues between heterogeneous platforms. A Web services security mode based on the .NET platform and the Axis2 platform is proposed in the papers. We also pointed out the direction of the future research.

  15. Efficiently Selecting the Best Web Services

    NASA Astrophysics Data System (ADS)

    Goncalves, Marlene; Vidal, Maria-Esther; Regalado, Alfredo; Yacoubi Ayadi, Nadia

    Emerging technologies and linking data initiatives have motivated the publication of a large number of datasets, and provide the basis for publishing Web services and tools to manage the available data. This wealth of resources opens a world of possibilities to satisfy user requests. However, Web services may have similar functionality and assess different performance; therefore, it is required to identify among the Web services that satisfy a user request, the ones with the best quality. In this paper we propose a hybrid approach that combines reasoning tasks with ranking techniques to aim at the selection of the Web services that best implement a user request. Web service functionalities are described in terms of input and output attributes annotated with existing ontologies, non-functionality is represented as Quality of Services (QoS) parameters, and user requests correspond to conjunctive queries whose sub-goals impose restrictions on the functionality and quality of the services to be selected. The ontology annotations are used in different reasoning tasks to infer service implicit properties and to augment the size of the service search space. Furthermore, QoS parameters are considered by a ranking metric to classify the services according to how well they meet a user non-functional condition. We assume that all the QoS parameters of the non-functional condition are equally important, and apply the Top-k Skyline approach to select the k services that best meet this condition. Our proposal relies on a two-fold solution which fires a deductive-based engine that performs different reasoning tasks to discover the services that satisfy the requested functionality, and an efficient implementation of the Top-k Skyline approach to compute the top-k services that meet the majority of the QoS constraints. Our Top-k Skyline solution exploits the properties of the Skyline Frequency metric and identifies the top-k services by just analyzing a subset of the services that

  16. Grid Enabled Geospatial Catalogue Web Service

    NASA Technical Reports Server (NTRS)

    Chen, Ai-Jun; Di, Li-Ping; Wei, Ya-Xing; Liu, Yang; Bui, Yu-Qi; Hu, Chau-Min; Mehrotra, Piyush

    2004-01-01

    Geospatial Catalogue Web Service is a vital service for sharing and interoperating volumes of distributed heterogeneous geospatial resources, such as data, services, applications, and their replicas over the web. Based on the Grid technology and the Open Geospatial Consortium (0GC) s Catalogue Service - Web Information Model, this paper proposes a new information model for Geospatial Catalogue Web Service, named as GCWS which can securely provides Grid-based publishing, managing and querying geospatial data and services, and the transparent access to the replica data and related services under the Grid environment. This information model integrates the information model of the Grid Replica Location Service (RLS)/Monitoring & Discovery Service (MDS) with the information model of OGC Catalogue Service (CSW), and refers to the geospatial data metadata standards from IS0 19115, FGDC and NASA EOS Core System and service metadata standards from IS0 191 19 to extend itself for expressing geospatial resources. Using GCWS, any valid geospatial user, who belongs to an authorized Virtual Organization (VO), can securely publish and manage geospatial resources, especially query on-demand data in the virtual community and get back it through the data-related services which provide functions such as subsetting, reformatting, reprojection etc. This work facilitates the geospatial resources sharing and interoperating under the Grid environment, and implements geospatial resources Grid enabled and Grid technologies geospatial enabled. It 2!so makes researcher to focus on science, 2nd not cn issues with computing ability, data locztic, processir,g and management. GCWS also is a key component for workflow-based virtual geospatial data producing.

  17. Web-based Service Portal in Healthcare

    NASA Astrophysics Data System (ADS)

    Silhavy, Petr; Silhavy, Radek; Prokopova, Zdenka

    Information delivery is one the most important task in healthcare. The growing sector of electronic healthcare has an important impact on the information delivery. There are two basic approaches towards information delivering. The first is web portal and second is touch-screen terminal. The aim of this paper is to investigate the web-based service portal. The most important advantage of web-based portal in the field of healthcare is an independent access for patients. This paper deals with the conditions and frameworks for healthcare portals

  18. Designing Crop Simulation Web Service with Service Oriented Architecture Principle

    NASA Astrophysics Data System (ADS)

    Chinnachodteeranun, R.; Hung, N. D.; Honda, K.

    2015-12-01

    Crop simulation models are efficient tools for simulating crop growth processes and yield. Running crop models requires data from various sources as well as time-consuming data processing, such as data quality checking and data formatting, before those data can be inputted to the model. It makes the use of crop modeling limited only to crop modelers. We aim to make running crop models convenient for various users so that the utilization of crop models will be expanded, which will directly improve agricultural applications. As the first step, we had developed a prototype that runs DSSAT on Web called as Tomorrow's Rice (v. 1). It predicts rice yields based on a planting date, rice's variety and soil characteristics using DSSAT crop model. A user only needs to select a planting location on the Web GUI then the system queried historical weather data from available sources and expected yield is returned. Currently, we are working on weather data connection via Sensor Observation Service (SOS) interface defined by Open Geospatial Consortium (OGC). Weather data can be automatically connected to a weather generator for generating weather scenarios for running the crop model. In order to expand these services further, we are designing a web service framework consisting of layers of web services to support compositions and executions for running crop simulations. This framework allows a third party application to call and cascade each service as it needs for data preparation and running DSSAT model using a dynamic web service mechanism. The framework has a module to manage data format conversion, which means users do not need to spend their time curating the data inputs. Dynamic linking of data sources and services are implemented using the Service Component Architecture (SCA). This agriculture web service platform demonstrates interoperability of weather data using SOS interface, convenient connections between weather data sources and weather generator, and connecting

  19. University Student Perceptions of the Use of the Web in Distance-Delivered Programs.

    ERIC Educational Resources Information Center

    Kanuka, Heather

    2001-01-01

    Investigated over two years Canadian students' needs when integrating the Web into distance-delivered degree programs, using Michael Moore's theory of transactional distance. Found that students considered the Web an effective technology to facilitate distance delivery; problems included working with instructors without expertise in the…

  20. Evolution of web services in bioinformatics.

    PubMed

    Neerincx, Pieter B T; Leunissen, Jack A M

    2005-06-01

    Bioinformaticians have developed large collections of tools to make sense of the rapidly growing pool of molecular biological data. Biological systems tend to be complex and in order to understand them, it is often necessary to link many data sets and use more than one tool. Therefore, bioinformaticians have experimented with several strategies to try to integrate data sets and tools. Owing to the lack of standards for data sets and the interfaces of the tools this is not a trivial task. Over the past few years building services with web-based interfaces has become a popular way of sharing the data and tools that have resulted from many bioinformatics projects. This paper discusses the interoperability problem and how web services are being used to try to solve it, resulting in the evolution of tools with web interfaces from HTML/web form-based tools not suited for automatic workflow generation to a dynamic network of XML-based web services that can easily be used to create pipelines.

  1. CMR Catalog Service for the Web

    NASA Technical Reports Server (NTRS)

    Newman, Doug; Mitchell, Andrew

    2016-01-01

    With the impending retirement of Global Change Master Directory (GCMD) Application Programming Interfaces (APIs) the Common Metadata Repository (CMR) was charged with providing a collection-level Catalog Service for the Web (CSW) that provided the same level of functionality as GCMD. This talk describes the capabilities of the CMR CSW API with particular reference to the support of the Committee on Earth Observation Satellites (CEOS) Working Group on Information Systems and Services (WGISS) Integrated Catalog (CWIC).

  2. Introducing the PRIDE Archive RESTful web services.

    PubMed

    Reisinger, Florian; del-Toro, Noemi; Ternent, Tobias; Hermjakob, Henning; Vizcaíno, Juan Antonio

    2015-07-01

    The PRIDE (PRoteomics IDEntifications) database is one of the world-leading public repositories of mass spectrometry (MS)-based proteomics data and it is a founding member of the ProteomeXchange Consortium of proteomics resources. In the original PRIDE database system, users could access data programmatically by accessing the web services provided by the PRIDE BioMart interface. New REST (REpresentational State Transfer) web services have been developed to serve the most popular functionality provided by BioMart (now discontinued due to data scalability issues) and address the data access requirements of the newly developed PRIDE Archive. Using the API (Application Programming Interface) it is now possible to programmatically query for and retrieve peptide and protein identifications, project and assay metadata and the originally submitted files. Searching and filtering is also possible by metadata information, such as sample details (e.g. species and tissues), instrumentation (mass spectrometer), keywords and other provided annotations. The PRIDE Archive web services were first made available in April 2014. The API has already been adopted by a few applications and standalone tools such as PeptideShaker, PRIDE Inspector, the Unipept web application and the Python-based BioServices package. This application is free and open to all users with no login requirement and can be accessed at http://www.ebi.ac.uk/pride/ws/archive/.

  3. Introducing the PRIDE Archive RESTful web services

    PubMed Central

    Reisinger, Florian; del-Toro, Noemi; Ternent, Tobias; Hermjakob, Henning; Vizcaíno, Juan Antonio

    2015-01-01

    The PRIDE (PRoteomics IDEntifications) database is one of the world-leading public repositories of mass spectrometry (MS)-based proteomics data and it is a founding member of the ProteomeXchange Consortium of proteomics resources. In the original PRIDE database system, users could access data programmatically by accessing the web services provided by the PRIDE BioMart interface. New REST (REpresentational State Transfer) web services have been developed to serve the most popular functionality provided by BioMart (now discontinued due to data scalability issues) and address the data access requirements of the newly developed PRIDE Archive. Using the API (Application Programming Interface) it is now possible to programmatically query for and retrieve peptide and protein identifications, project and assay metadata and the originally submitted files. Searching and filtering is also possible by metadata information, such as sample details (e.g. species and tissues), instrumentation (mass spectrometer), keywords and other provided annotations. The PRIDE Archive web services were first made available in April 2014. The API has already been adopted by a few applications and standalone tools such as PeptideShaker, PRIDE Inspector, the Unipept web application and the Python-based BioServices package. This application is free and open to all users with no login requirement and can be accessed at http://www.ebi.ac.uk/pride/ws/archive/. PMID:25904633

  4. Predicting Student Performance in Web-Based Distance Education Courses Based on Survey Instruments Measuring Personality Traits and Technical Skills

    ERIC Educational Resources Information Center

    Hall, Michael

    2008-01-01

    Two common web-based surveys, "Is Online Learning Right for Me?' and "What Technical Skills Do I Need?", were combined into a single survey instrument and given to 228 on-campus and 83 distance education students. The students were enrolled in four different classes (business, computer information services, criminal justice, and…

  5. Enhancing Data Interoperability with Web Services

    NASA Astrophysics Data System (ADS)

    Shrestha, S. R.; Zimble, D. A.; Wang, W.; Herring, D.; Halpert, M.

    2014-12-01

    In an effort to improve data access and interoperability of climate and weather data, the National Oceanic and Atmospheric Administration's (NOAA) Climate.gov and Climate Prediction Center (CPC) are exploring various platform solutions to enhance a user's ability to locate, preview, and acquire the data. The Climate.gov and CPC data team faces multiple challenges including the various kinds of data and formats, inconsistency of metadata records, variety of data service implementations, very large volumes of data and geographically distributed locations. We have created the Data Access and Interoperability project to design a web-based platform, where interoperability between systems can be leveraged to allow greater data discovery, access, visualization and delivery. In the interoperable data platform, systems can integrate with each other to support the synthesis of climate and weather data. Interoperability is the ability for users to discover the available climate and weather data, preview and interact with the data, and acquire the data in common digital formats through a simple web-based interface. The goal of the interoperable data platform is to leverage existing web services, implement the established standards and integrate with existing solutions across the earth sciences domain instead of creating new technologies. Towards this effort to improve the interoperability of the platform, we are collaborating with ESRI Inc. to provide climate and weather data via web services. In this presentation, we will discuss and demonstrate how to use ArcGIS to author RESTful based scientific web services using open standards. These web services are able to encapsulate the logic required to handle and describe scientific data through a variety of service types including, image, map, feature, geoprocessing, and their respective service methods. Combining these types of services and leveraging well-documented APIs, including the ArcGIS JavaScript API, we can afford to

  6. An Integrated Environment for Geospatial Web Service Composition

    NASA Astrophysics Data System (ADS)

    Zhao, P.; di, L.; Wei, Y.

    2006-05-01

    Assembling individual geospatial web services into more complex and more useful web processes to achieve desired results proves to be essential for complex geospatial applications. As geospatial web services embed lots of geospatial information and knowledge, the geospatial domain-specific tools are needed to help users discover, retrieve and integrate these services. We have developed an integrated environment that allows users in the semi-automatic and dynamic composition of geospatial web services. This paper illustrates the whole life cycle of a web service composition: 1) discover services in OGC Catalog Service for Web (CSW), 2) present and select matching service at each step of a composition with the assistance of intelligent autonomous interface agent, 3) compose a service chain based on domain knowledge, and 4) execute the composed service chain automatically through its BPEL and WSDL scripts. The environment has been used to demonstrate practical benefits in the context of OGC Web services for Earth science research and applications.

  7. Focused Crawling of the Deep Web Using Service Class Descriptions

    SciTech Connect

    Rocco, D; Liu, L; Critchlow, T

    2004-06-21

    Dynamic Web data sources--sometimes known collectively as the Deep Web--increase the utility of the Web by providing intuitive access to data repositories anywhere that Web access is available. Deep Web services provide access to real-time information, like entertainment event listings, or present a Web interface to large databases or other data repositories. Recent studies suggest that the size and growth rate of the dynamic Web greatly exceed that of the static Web, yet dynamic content is often ignored by existing search engine indexers owing to the technical challenges that arise when attempting to search the Deep Web. To address these challenges, we present DynaBot, a service-centric crawler for discovering and clustering Deep Web sources offering dynamic content. DynaBot has three unique characteristics. First, DynaBot utilizes a service class model of the Web implemented through the construction of service class descriptions (SCDs). Second, DynaBot employs a modular, self-tuning system architecture for focused crawling of the DeepWeb using service class descriptions. Third, DynaBot incorporates methods and algorithms for efficient probing of the Deep Web and for discovering and clustering Deep Web sources and services through SCD-based service matching analysis. Our experimental results demonstrate the effectiveness of the service class discovery, probing, and matching algorithms and suggest techniques for efficiently managing service discovery in the face of the immense scale of the Deep Web.

  8. A Web Services Data Analysis Grid

    SciTech Connect

    William A Watson III; Ian Bird; Jie Chen; Bryan Hess; Andy Kowalski; Ying Chen

    2002-07-01

    The trend in large-scale scientific data analysis is to exploit compute, storage and other resources located at multiple sites, and to make those resources accessible to the scientist as if they were a single, coherent system. Web technologies driven by the huge and rapidly growing electronic commerce industry provide valuable components to speed the deployment of such sophisticated systems. Jefferson Lab, where several hundred terabytes of experimental data are acquired each year, is in the process of developing a web-based distributed system for data analysis and management. The essential aspects of this system are a distributed data grid (site independent access to experiment, simulation and model data) and a distributed batch system, augmented with various supervisory and management capabilities, and integrated using Java and XML-based web services.

  9. User Needs of Digital Service Web Portals: A Case Study

    ERIC Educational Resources Information Center

    Heo, Misook; Song, Jung-Sook; Seol, Moon-Won

    2013-01-01

    The authors examined the needs of digital information service web portal users. More specifically, the needs of Korean cultural portal users were examined as a case study. The conceptual framework of a web-based portal is that it is a complex, web-based service application with characteristics of information systems and service agents. In…

  10. Web Services Basic Profile Version 1.0

    NASA Astrophysics Data System (ADS)

    Graham, Matthew; Schaaff, Andre; Graham, Matthew

    2010-12-01

    This document describes rules to take into account when implementing SOAP-based web services. It explains also how to check conformance to these rules. It can be read as a "Guideline to VO Web Service Interoperability" or a "How to provide interoperable VO web services".

  11. Problems with Assessment Validity and Reliability in Web-Based Distance Learning Environments and Solutions

    ERIC Educational Resources Information Center

    Wijekumar, Kay; Ferguson, Lon; Wagoner, Diane

    2006-01-01

    Assessment of learning is critical to the learners, teachers, and designers of learning environments. Current assessment techniques in web-based distance learning apply age-old techniques to a new medium and are not adequate for web-based distance learning environments (WBDLE). The goals of this article are to extend existing critiques of…

  12. Influence of Structure and Interaction on Student Achievement and Satisfaction in Web-Based Distance Learning

    ERIC Educational Resources Information Center

    Lee, Hye-Jung; Rha, Ilju

    2009-01-01

    This study examines the influence of instructional design and management style on student achievement and satisfaction in a web-based distance learning environment. From the literature review, two major instructional design and management styles in web-based distance education were conceptualized as structure and interpersonal interaction. To…

  13. Quality Assurance of Web-Based Learning in Distance Education Institutions.

    ERIC Educational Resources Information Center

    Yeung, Davey

    2001-01-01

    Discussion of quality in higher education focuses on quality assurance of Web-based learning in distance education, based on experiences at the Open University of Hong Kong. Highlights include problems with distance education; a literature review; traditional learning versus Web-based learning; and the changing role of faculty. (Contains 63…

  14. Contract Observation in Web Services Environments

    NASA Astrophysics Data System (ADS)

    Bíba, Jiří; Hodík, Jiří; Jakob, Michal; Pěchouček, Michal

    Electronic contracting, based on explicit representation of different parties' commitments, is a promising way to specifying and regulating behaviour in distributed business applications. A key part of contract-based system is a process through which the actual behaviour of individual parties is checked for conformance with contracts set to govern such behaviour. Such checking requires that relevant information on the behaviour of the parties, both with respect to the application processes they execute and to managing their contractual relationships, is captured. The process of collecting all such information, termed contract observation, is the subject of this paper. First, we describe general properties and requirements of such an observation process; afterwards, we discuss specifics of realising contract observation in web services environments. Finally, we show how contract observation has been implemented as part of the IST-CONTRACT web services framework for contract-based systems.

  15. The Reality of Web-Based Interaction in an Egyptian Distance Education Course

    ERIC Educational Resources Information Center

    Sadik, Alaa

    2006-01-01

    This paper reports the results of a study conducted to evaluate the reality of interaction in a web-based distance education course. The learners were Egyptian first-grade secondary school students (15-16 years old) and the learning subject is mathematics. To investigate students' interactions via the Web, a Web-based learning environment was…

  16. Climate Model Diagnostic Analyzer Web Service System

    NASA Astrophysics Data System (ADS)

    Lee, S.; Pan, L.; Zhai, C.; Tang, B.; Kubar, T. L.; Li, J.; Zhang, J.; Wang, W.

    2015-12-01

    Both the National Research Council Decadal Survey and the latest Intergovernmental Panel on Climate Change Assessment Report stressed the need for the comprehensive and innovative evaluation of climate models with the synergistic use of global satellite observations in order to improve our weather and climate simulation and prediction capabilities. The abundance of satellite observations for fundamental climate parameters and the availability of coordinated model outputs from CMIP5 for the same parameters offer a great opportunity to understand and diagnose model biases in climate models. In addition, the Obs4MIPs efforts have created several key global observational datasets that are readily usable for model evaluations. However, a model diagnostic evaluation process requires physics-based multi-variable comparisons that typically involve large-volume and heterogeneous datasets, making them both computationally- and data-intensive. In response, we have developed a novel methodology to diagnose model biases in contemporary climate models and implementing the methodology as a web-service based, cloud-enabled, provenance-supported climate-model evaluation system. The evaluation system is named Climate Model Diagnostic Analyzer (CMDA), which is the product of the research and technology development investments of several current and past NASA ROSES programs. The current technologies and infrastructure of CMDA are designed and selected to address several technical challenges that the Earth science modeling and model analysis community faces in evaluating and diagnosing climate models. In particular, we have three key technology components: (1) diagnostic analysis methodology; (2) web-service based, cloud-enabled technology; (3) provenance-supported technology. The diagnostic analysis methodology includes random forest feature importance ranking, conditional probability distribution function, conditional sampling, and time-lagged correlation map. We have implemented the

  17. Climate Model Diagnostic Analyzer Web Service System

    NASA Astrophysics Data System (ADS)

    Lee, S.; Pan, L.; Zhai, C.; Tang, B.; Jiang, J. H.

    2014-12-01

    We have developed a cloud-enabled web-service system that empowers physics-based, multi-variable model performance evaluations and diagnoses through the comprehensive and synergistic use of multiple observational data, reanalysis data, and model outputs. We have developed a methodology to transform an existing science application code into a web service using a Python wrapper interface and Python web service frameworks. The web-service system, called Climate Model Diagnostic Analyzer (CMDA), currently supports (1) all the observational datasets from Obs4MIPs and a few ocean datasets from NOAA and Argo, which can serve as observation-based reference data for model evaluation, (2) many of CMIP5 model outputs covering a broad range of atmosphere, ocean, and land variables from the CMIP5 specific historical runs and AMIP runs, and (3) ECMWF reanalysis outputs for several environmental variables in order to supplement observational datasets. Analysis capabilities currently supported by CMDA are (1) the calculation of annual and seasonal means of physical variables, (2) the calculation of time evolution of the means in any specified geographical region, (3) the calculation of correlation between two variables, (4) the calculation of difference between two variables, and (5) the conditional sampling of one physical variable with respect to another variable. A web user interface is chosen for CMDA because it not only lowers the learning curve and removes the adoption barrier of the tool but also enables instantaneous use, avoiding the hassle of local software installation and environment incompatibility. CMDA will be used as an educational tool for the summer school organized by JPL's Center for Climate Science in 2014. In order to support 30+ simultaneous users during the school, we have deployed CMDA to the Amazon cloud environment. The cloud-enabled CMDA will provide each student with a virtual machine while the user interaction with the system will remain the same

  18. Distance Learning: Information Access and Services for Virtual Users.

    ERIC Educational Resources Information Center

    Iyer, Hemalata, Ed.

    This volume centers broadly on information support services for distance education. The articles in this book can be categorized into two areas: access to information resources for distance learners, and studies of distance learning programs. Contents include: "The Challenges and Benefits of Asynchronous Learning Networks" (Daphne Jorgensen);…

  19. Using EMBL-EBI services via Web interface and programmatically via Web Services

    PubMed Central

    Lopez, Rodrigo; Cowley, Andrew; Li, Weizhong; McWilliam, Hamish

    2015-01-01

    The European Bioinformatics Institute (EMBL-EBI) provides access to a wide range of databases and analysis tools that are of key importance in bioinformatics. As well as providing Web interfaces to these resources, Web Services are available using SOAP and REST protocols that enable programmatic access to our resources and allow their integration into other applications and analytical workflows. This unit describes the various options available to a typical researcher or bioinformatician who wishes to use our resources via Web interface or programmatically via a range of programming languages. PMID:25501941

  20. Pragmatic Computing - A Semiotic Perspective to Web Services

    NASA Astrophysics Data System (ADS)

    Liu, Kecheng

    The web seems to have evolved from a syntactic web, a semantic web to a pragmatic web. This evolution conforms to the study of information and technology from the theory of semiotics. The pragmatics, concerning with the use of information in relation to the context and intended purposes, is extremely important in web service and applications. Much research in pragmatics has been carried out; but in the same time, attempts and solutions have led to some more questions. After reviewing the current work in pragmatic web, the paper presents a semiotic approach to website services, particularly on request decomposition and service aggregation.

  1. Climate Model Diagnostic Analyzer Web Service System

    NASA Astrophysics Data System (ADS)

    Lee, S.; Pan, L.; Zhai, C.; Tang, B.; Jiang, J. H.

    2013-12-01

    The latest Intergovernmental Panel on Climate Change (IPCC) Fourth Assessment Report stressed the need for the comprehensive and innovative evaluation of climate models with newly available global observations. The traditional approach to climate model evaluation, which compares a single parameter at a time, identifies symptomatic model biases and errors but fails to diagnose the model problems. The model diagnosis process requires physics-based multi-variable comparisons that typically involve large-volume and heterogeneous datasets, making them both computationally- and data-intensive. To address these challenges, we are developing a parallel, distributed web-service system that enables the physics-based multi-variable model performance evaluations and diagnoses through the comprehensive and synergistic use of multiple observational data, reanalysis data, and model outputs. We have developed a methodology to transform an existing science application code into a web service using a Python wrapper interface and Python web service frameworks (i.e., Flask, Gunicorn, and Tornado). The web-service system, called Climate Model Diagnostic Analyzer (CMDA), currently supports (1) all the datasets from Obs4MIPs and a few ocean datasets from NOAA and Argo, which can serve as observation-based reference data for model evaluation and (2) many of CMIP5 model outputs covering a broad range of atmosphere, ocean, and land variables from the CMIP5 specific historical runs and AMIP runs. Analysis capabilities currently supported by CMDA are (1) the calculation of annual and seasonal means of physical variables, (2) the calculation of time evolution of the means in any specified geographical region, (3) the calculation of correlation between two variables, and (4) the calculation of difference between two variables. A web user interface is chosen for CMDA because it not only lowers the learning curve and removes the adoption barrier of the tool but also enables instantaneous use

  2. A Different Web-Based Geocoding Service Using Fuzzy Techniques

    NASA Astrophysics Data System (ADS)

    Pahlavani, P.; Abbaspour, R. A.; Zare Zadiny, A.

    2015-12-01

    Geocoding - the process of finding position based on descriptive data such as address or postal code - is considered as one of the most commonly used spatial analyses. Many online map providers such as Google Maps, Bing Maps and Yahoo Maps present geocoding as one of their basic capabilities. Despite the diversity of geocoding services, users usually face some limitations when they use available online geocoding services. In existing geocoding services, proximity and nearness concept is not modelled appropriately as well as these services search address only by address matching based on descriptive data. In addition there are also some limitations in display searching results. Resolving these limitations can enhance efficiency of the existing geocoding services. This paper proposes the idea of integrating fuzzy technique with geocoding process to resolve these limitations. In order to implement the proposed method, a web-based system is designed. In proposed method, nearness to places is defined by fuzzy membership functions and multiple fuzzy distance maps are created. Then these fuzzy distance maps are integrated using fuzzy overlay technique for obtain the results. Proposed methods provides different capabilities for users such as ability to search multi-part addresses, searching places based on their location, non-point representation of results as well as displaying search results based on their priority.

  3. Web Map Services (WMS) Global Mosaic

    NASA Technical Reports Server (NTRS)

    Percivall, George; Plesea, Lucian

    2003-01-01

    The WMS Global Mosaic provides access to imagery of the global landmass using an open standard for web mapping. The seamless image is a mosaic of Landsat 7 scenes; geographically-accurate with 30 and 15 meter resolutions. By using the OpenGIS Web Map Service (WMS) interface, any organization can use the global mosaic as a layer in their geospatial applications. Based on a trade study, an implementation approach was chosen that extends a previously developed WMS hosting a Landsat 5 CONUS mosaic developed by JPL. The WMS Global Mosaic supports the NASA Geospatial Interoperability Office goal of providing an integrated digital representation of the Earth, widely accessible for humanity's critical decisions.

  4. Synthetic seismogram web service and Python tools

    NASA Astrophysics Data System (ADS)

    Heimann, Sebastian; Cesca, Simone; Kriegerowski, Marius; Dahm, Torsten

    2014-05-01

    Many geophysical methods require knowledge of Green's functions (GF) or synthetic seismograms in dependence of ranges of source and receiver coordinates. Examples include synthetic seismogram generation, moment tensor inversion, the modeling of depth phases for regional and teleseismic earthquakes, or the modeling of pressure diffusion induced static displacement and strain. Calculation of Green's functions is a computationally expensive operation and it can be of advantage to calculate them in advance: the same Green's function traces can then be reused several or many times as required in a typical application. Regarding Green's function computation as an independent step in a use-case's processing chain encourages to store these in an application independent form. They can then be shared between different applications and they can also be passed to other researchers, e.g. via a web service. Starting now, we provide such a web service to the seismological community (http://kinherd.org/), where a researcher can share Green's function stores and retrieve synthetic seismograms for various point and extended earthquake source models for many different earth models at local, regional and global scale. This web service is part of a rich new toolset for the creation and handling of Green's functions and synthetic seismograms (http://emolch.github.com/pyrocko/gf). It can be used off-line or in client mode. Its core features are: greatly simplified generation of Green's function stores supports various codes for Green's function computation extensible Green's function storage format flexible spacial indexing of Green's functions integrated travel time computation support for other types of Green's functions; e.g. poro-elastic GFs written in Python

  5. Solving the Problem of Promoting Distance Library Services

    ERIC Educational Resources Information Center

    Wyss, Paul Alan

    2007-01-01

    Promoting services is a conundrum for any organization. This is especially true for an academic library promoting distance library services. Systems thinking, process mapping, team learning, and diffusion of information practices offer ways of thinking about promoting services that help those involved find novel ways to approach promoting distance…

  6. OGC Web Services standards by example : the European Seismic Portal

    NASA Astrophysics Data System (ADS)

    Frobert, L.; Kamb, L.; Trani, L.; Spinuso, A.; Bossu, R.; Van Eck, T.

    2011-12-01

    NERIES (2006-2010) was an Integrated Infrastructure Initiative (I3) project in the Sixth Framework Program (FP6) of the European Commission (EC), aiming at networking the European seismic networks, improving access to data, allowing access to specific seismic infrastructures and pursuing targeted research developing the next generation of tools for improved service and data analysis. During this project, a web portal was developed using web services to access data and a Visual Web Applications to display them. However these web services were not conform to any standard, making them difficult to consume by any new user interface. Therefore, for the NERA project, the follow-up of NERIES, we have proposed the use of web services standards to access our data. We have decided to use standards defined by the Open Geospatial Consortium (OGC). The OGC defines standards for the Web service interfaces to access geo-tagged data. The events and seismic stations are also geo-tagged making these web services suitable for our purpose. Using standard web services gives us the opportunity to distribute our data across all conformant consumers to these standards through various programming languages and applications We have implemented a preliminary version of web services conforming to the Web Map Service (WMS) and Web Feature Service (WFS) standard to access our catalog of seismic events (nearly 200 000 events). To visualize them we have made four examples demo on our web site using different technologies (Adobe Flash, JavaScript, Java with Nasa World Wind and UDig a desktop GIS application). In the future we hope to implement other OGC Web services standard like : - Sensor Observation Service (SOS) to provide seismic waveform records; - Web Notification Service (WNS); - Catalog Service for the Web (CSW) to provide a search engine of all our web services; - Web Processing Service (WPS) to process data between different services. The power of the use of OGC standards is the easy

  7. A web services choreography scenario for interoperating bioinformatics applications

    PubMed Central

    de Knikker, Remko; Guo, Youjun; Li, Jin-long; Kwan, Albert KH; Yip, Kevin Y; Cheung, David W; Cheung, Kei-Hoi

    2004-01-01

    Background Very often genome-wide data analysis requires the interoperation of multiple databases and analytic tools. A large number of genome databases and bioinformatics applications are available through the web, but it is difficult to automate interoperation because: 1) the platforms on which the applications run are heterogeneous, 2) their web interface is not machine-friendly, 3) they use a non-standard format for data input and output, 4) they do not exploit standards to define application interface and message exchange, and 5) existing protocols for remote messaging are often not firewall-friendly. To overcome these issues, web services have emerged as a standard XML-based model for message exchange between heterogeneous applications. Web services engines have been developed to manage the configuration and execution of a web services workflow. Results To demonstrate the benefit of using web services over traditional web interfaces, we compare the two implementations of HAPI, a gene expression analysis utility developed by the University of California San Diego (UCSD) that allows visual characterization of groups or clusters of genes based on the biomedical literature. This utility takes a set of microarray spot IDs as input and outputs a hierarchy of MeSH Keywords that correlates to the input and is grouped by Medical Subject Heading (MeSH) category. While the HTML output is easy for humans to visualize, it is difficult for computer applications to interpret semantically. To facilitate the capability of machine processing, we have created a workflow of three web services that replicates the HAPI functionality. These web services use document-style messages, which means that messages are encoded in an XML-based format. We compared three approaches to the implementation of an XML-based workflow: a hard coded Java application, Collaxa BPEL Server and Taverna Workbench. The Java program functions as a web services engine and interoperates with these web

  8. Creating OGC Web Processing Service workflows using a web-based editor

    NASA Astrophysics Data System (ADS)

    de Jesus, J.; Walker, P.; Grant, M.

    2012-04-01

    The OGC WPS (Web Processing Service) specifies how geospatial algorithms may be accessed in an SOA (Service Oriented Architecture). Service providers can encode both simple and sophisticated algorithms as WPS processes and publish them as web services. These services are not only useful individually but may be built into complex processing chains (workflows) that can solve complex data analysis and/or scientific problems. The NETMAR project has extended the Web Processing Service (WPS) framework to provide transparent integration between it and the commonly used WSDL (Web Service Description Language) that describes the web services and its default SOAP (Simple Object Access Protocol) binding. The extensions allow WPS services to be orchestrated using commonly used tools (in this case Taverna Workbench, but BPEL based systems would also be an option). We have also developed a WebGUI service editor, based on HTML5 and the WireIt! Javascript API, that allows users to create these workflows using only a web browser. The editor is coded entirely in Javascript and performs all XSLT transformations needed to produce a Taverna compatible (T2FLOW) workflow description which can be exported and run on a local Taverna Workbench or uploaded to a web-based orchestration server and run there. Here we present the NETMAR WebGUI service chain editor and discuss the problems associated with the development of a WebGUI for scientific workflow editing; content transformation into the Taverna orchestration language (T2FLOW/SCUFL); final orchestration in the Taverna engine and how to deal with the large volumes of data being transferred between different WPS services (possibly running on different servers) during workflow orchestration. We will also demonstrate using the WebGUI for creating a simple workflow making use of published web processing services, showing how simple services may be chained together to produce outputs that would previously have required a GIS (Geographic

  9. Persistence and availability of Web services in computational biology.

    PubMed

    Schultheiss, Sebastian J; Münch, Marc-Christian; Andreeva, Gergana D; Rätsch, Gunnar

    2011-01-01

    We have conducted a study on the long-term availability of bioinformatics Web services: an observation of 927 Web services published in the annual Nucleic Acids Research Web Server Issues between 2003 and 2009. We found that 72% of Web sites are still available at the published addresses, only 9% of services are completely unavailable. Older addresses often redirect to new pages. We checked the functionality of all available services: for 33%, we could not test functionality because there was no example data or a related problem; 13% were truly no longer working as expected; we could positively confirm functionality only for 45% of all services. Additionally, we conducted a survey among 872 Web Server Issue corresponding authors; 274 replied. 78% of all respondents indicate their services have been developed solely by students and researchers without a permanent position. Consequently, these services are in danger of falling into disrepair after the original developers move to another institution, and indeed, for 24% of services, there is no plan for maintenance, according to the respondents. We introduce a Web service quality scoring system that correlates with the number of citations: services with a high score are cited 1.8 times more often than low-scoring services. We have identified key characteristics that are predictive of a service's survival, providing reviewers, editors, and Web service developers with the means to assess or improve Web services. A Web service conforming to these criteria receives more citations and provides more reliable service for its users. The most effective way of ensuring continued access to a service is a persistent Web address, offered either by the publishing journal, or created on the authors' own initiative, for example at http://bioweb.me. The community would benefit the most from a policy requiring any source code needed to reproduce results to be deposited in a public repository.

  10. Persistence and availability of Web services in computational biology.

    PubMed

    Schultheiss, Sebastian J; Münch, Marc-Christian; Andreeva, Gergana D; Rätsch, Gunnar

    2011-01-01

    We have conducted a study on the long-term availability of bioinformatics Web services: an observation of 927 Web services published in the annual Nucleic Acids Research Web Server Issues between 2003 and 2009. We found that 72% of Web sites are still available at the published addresses, only 9% of services are completely unavailable. Older addresses often redirect to new pages. We checked the functionality of all available services: for 33%, we could not test functionality because there was no example data or a related problem; 13% were truly no longer working as expected; we could positively confirm functionality only for 45% of all services. Additionally, we conducted a survey among 872 Web Server Issue corresponding authors; 274 replied. 78% of all respondents indicate their services have been developed solely by students and researchers without a permanent position. Consequently, these services are in danger of falling into disrepair after the original developers move to another institution, and indeed, for 24% of services, there is no plan for maintenance, according to the respondents. We introduce a Web service quality scoring system that correlates with the number of citations: services with a high score are cited 1.8 times more often than low-scoring services. We have identified key characteristics that are predictive of a service's survival, providing reviewers, editors, and Web service developers with the means to assess or improve Web services. A Web service conforming to these criteria receives more citations and provides more reliable service for its users. The most effective way of ensuring continued access to a service is a persistent Web address, offered either by the publishing journal, or created on the authors' own initiative, for example at http://bioweb.me. The community would benefit the most from a policy requiring any source code needed to reproduce results to be deposited in a public repository. PMID:21966383

  11. Library Services to Distance Learners in the Commonwealth: A Reader.

    ERIC Educational Resources Information Center

    Watson, Elizabeth F., Ed.; Jagannathan, Neela, Ed.

    The provision of good library services is a crucial factor in determining the quality of distance education. This collection of articles acquaints readers with distance librarianship as it is practiced in developed and developing countries throughout the British Commonwealth. The reader includes: "Introduction" (Michael Wooliscroft); "Distance…

  12. A web service for service composition to aid geospatial modelers

    NASA Astrophysics Data System (ADS)

    Bigagli, L.; Santoro, M.; Roncella, R.; Mazzetti, P.

    2012-04-01

    The identification of appropriate mechanisms for process reuse, chaining and composition is considered a key enabler for the effective uptake of a global Earth Observation infrastructure, currently pursued by the international geospatial research community. In the Earth and Space Sciences, such a facility could primarily enable integrated and interoperable modeling, for what several approaches have been proposed and developed, over the last years. In fact, GEOSS is specifically tasked with the development of the so-called "Model Web". At increasing levels of abstraction and generalization, the initial stove-pipe software tools have evolved to community-wide modeling frameworks, to Component-Based Architecture solution, and, more recently, started to embrace Service-Oriented Architectures technologies, such as the OGC WPS specification and the WS-* stack of W3C standards for service composition. However, so far, the level of abstraction seems too low for implementing the Model Web vision, and far too complex technological aspects must still be addressed by both providers and users, resulting in limited usability and, eventually, difficult uptake. As by the recent ICT trend of resource virtualization, it has been suggested that users in need of a particular processing capability, required by a given modeling workflow, may benefit from outsourcing the composition activities into an external first-class service, according to the Composition as a Service (CaaS) approach. A CaaS system provides the necessary interoperability service framework for adaptation, reuse and complementation of existing processing resources (including models and geospatial services in general) in the form of executable workflows. This work introduces the architecture of a CaaS system, as a distributed information system for creating, validating, editing, storing, publishing, and executing geospatial workflows. This way, the users can be freed from the need of a composition infrastructure and

  13. A Web Services Composition Design framework based on Agent Organization

    NASA Astrophysics Data System (ADS)

    Li, JiaJia; Li, Bin; Zhang, Xiaowei

    Computing environments are becoming more open, distributed and pervasive. The web services compositions build for these dynamic environments will need to become more adaptable and adaptive to unexpected event. This paper defines a way for web services composition which based on agent organization. The functions of three layers, classification of agent, and agent model and agents design in this framework are introduced in details. It realizes a reliable and flexible web services composition using this framework.

  14. Web Services and Their Use in Starlink Software

    NASA Astrophysics Data System (ADS)

    Taylor, M.; Platon, R.; Chipperfield, A.; Draper, P.; McIlwrath, B.; Giaretta, D.

    Web Services are gaining great popularity in the Grid community, and with good reason. The Starlink project is adopting Web Services as the method of interapplication communication. This is being done natively in new Java-based applications while older applications are being wrapped to provide Web Service interfaces. We are in this way providing interoperability between the generations of software in a heterogeneous, distributed manner and allowing the software to be usable in a distributed environment such as the GRID.

  15. Synchronous Distance Education: Using Web-Conferencing in an MBA Accounting Course

    ERIC Educational Resources Information Center

    Ellingson, Dee Ann; Notbohm, Matthew

    2012-01-01

    Online distance education can take many forms, from a correspondence course with materials online to fully synchronous, live instruction. This paper describes a fully synchronous, live format using web-conferencing. Some useful features of web-conferencing and the way they are employed in this course are described. Instructor observations and…

  16. Software Architecture Design of GIS Web Service Aggregation Based on Service Group

    NASA Astrophysics Data System (ADS)

    Liu, J.-C.; Yang, J.; Tan, M.-J.; Gan, Q.

    2011-08-01

    Based on the analysis of research status of domestic and international GIS web service aggregation and development tendency of public platform of GIS web service, the paper designed software architecture of GIS web service aggregation based on GIS web service group. Firstly, using heterogeneous GIS services model, the software architecture converted a variety of heterogeneous services to a unified interface of GIS services, and divided different types of GIS services into different service groups referring to description of GIS services. Secondly, a service aggregation process model was designed. This model completed the task of specific service aggregation instance, by automatically selecting member GIS Web services in the same service group. Dynamic capabilities and automatic adaptation of GIS Web services aggregation process were achieved. Thirdly, this paper designed a service evaluation model of GIS web service aggregation based on service group from three aspects, i.e. GIS Web Service itself, networking conditions and service consumer. This model implemented effective quality evaluation and performance monitoring of GIS web service aggregation. It could be used to guide the execution, monitor and service selection of aggregation process. Therefore, robustness of aggregated GIS web service was improved. Finally, the software architecture has been widely used in public platform of GIS web service and a number of geo-spatial framework constructions for digital city in Sichuan Province, and aggregated various GIS web services such as World Map(National Public Platform of Geo-spatial Service), ArcGIS, SuperMap, MapGIS, NewMap etc. Applications of items showed that this software architecture was practicability.

  17. Effectiveness of Learning Process Using "Web Technology" in the Distance Learning System

    ERIC Educational Resources Information Center

    Killedar, Manoj

    2008-01-01

    Web is a globally distributed, still highly personalized media for cost-effective delivery of multimedia information and services. Web is expected to have a strong impact on almost every aspect of how we learn. "Total Quality" is the totality of features, as perceived by the customers of the product or service. Totality of features includes stated…

  18. Geo-enabling Science through Web Services (Invited)

    NASA Astrophysics Data System (ADS)

    White, C. E.

    2010-12-01

    Sharing research is a crucial part of participating in science. The value of a dataset increases when users 1) know the dataset exists, and 2) can access the dataset and use it. Exposing data through web services allows other researchers to quickly access, overlay, and analyze data in the web-based or desktop mapping client of their choice. The ability to mash up different web services can reveal new - especially spatial - relationships between phenomena, and encourage creative uses of the data. This presentation will investigate the scientific and business value of standards-based web services for earth observation data, teach step-by-step how to expose such datasets as web services, and demonstrate tools - such as Catalog Services, REST endpoints, and GIS portals - that enable other researchers to discover and use web-accessible data resources.

  19. BioSWR--semantic web services registry for bioinformatics.

    PubMed

    Repchevsky, Dmitry; Gelpi, Josep Ll

    2014-01-01

    Despite of the variety of available Web services registries specially aimed at Life Sciences, their scope is usually restricted to a limited set of well-defined types of services. While dedicated registries are generally tied to a particular format, general-purpose ones are more adherent to standards and usually rely on Web Service Definition Language (WSDL). Although WSDL is quite flexible to support common Web services types, its lack of semantic expressiveness led to various initiatives to describe Web services via ontology languages. Nevertheless, WSDL 2.0 descriptions gained a standard representation based on Web Ontology Language (OWL). BioSWR is a novel Web services registry that provides standard Resource Description Framework (RDF) based Web services descriptions along with the traditional WSDL based ones. The registry provides Web-based interface for Web services registration, querying and annotation, and is also accessible programmatically via Representational State Transfer (REST) API or using a SPARQL Protocol and RDF Query Language. BioSWR server is located at http://inb.bsc.es/BioSWR/and its code is available at https://sourceforge.net/projects/bioswr/under the LGPL license.

  20. Distance Learning and the Web: Are Advertising Programs Missing the Target?

    ERIC Educational Resources Information Center

    Falk, Louis K.; Rehman, Sharaf; Foster, Dawn

    1999-01-01

    Discusses survey results that examined whether distance education programs at universities offering courses in advertising and/or public relations make use of the Internet/Web pages to inform potential students about courses taught via distance education. The survey and list of Association of Education in Journalism and Mass Communication (AEJMC)…

  1. Empirical Investigation into Motives for Choosing Web-Based Distance Learning Programs

    ERIC Educational Resources Information Center

    Alkhattabi, Mona

    2016-01-01

    Today, in association with rapid social and economic changes, there is an increasing level of demand for distance and online learning programs. This study will focus on identifying the main motivational factors for choosing a web-based distance-learning program. Moreover, it will investigate how these factors relate to age, gender, marital status…

  2. Innovation in Open & Distance Learning: Successful Development of Online and Web-Based Learning.

    ERIC Educational Resources Information Center

    Lockwood, Fred, Ed.; Gooley, Anne, Ed.

    This book contains 19 papers examining innovation in open and distance learning through development of online and World Wide Web-based learning. The following papers are included: "Innovation in Distributed Learning: Creating the Environment" (Fred Lockwood); "Innovation in Open and Distance Learning: Some Lessons from Experience and Research"…

  3. An Exploration of Cultural Value Orientations in Distance Education Web Marketing

    ERIC Educational Resources Information Center

    DeGaetano, Lora A.

    2013-01-01

    In the current global environment, universities seek to attract international students. The low enrollment of international students at a particular distance education institution demonstrated the competitive challenge in attracting international students. Distance education web marketing communications may be a factor influencing low enrollment…

  4. Web services for distributed and interoperable hydro-information systems

    NASA Astrophysics Data System (ADS)

    Horak, J.; Orlik, A.; Stromsky, J.

    2007-06-01

    Web services support the integration and interoperability of Web-based applications and enable machine-to-machine interaction. The concepts of web services and open distributed architecture were applied to the development of T-DSS, the prototype customised for web based hydro-information systems. T-DSS provides mapping services, database related services and access to remote components, with special emphasis placed on output flexibility (e.g. multilingualism), where SOAP web services are mainly used for communication. The remote components are represented above all by distant data and mapping services (e.g. eteorological predictions), modelling and analytical systems (currently HEC-HMS, Modflow and additional utilities), which support decision making in water management.

  5. Web services for distributed and interoperable hydro-information systems

    NASA Astrophysics Data System (ADS)

    Horak, J.; Orlik, A.; Stromsky, J.

    2008-03-01

    Web services support the integration and interoperability of Web-based applications and enable machine-to-machine interaction. The concepts of web services and open distributed architecture were applied to the development of T-DSS, the prototype customised for web based hydro-information systems. T-DSS provides mapping services, database related services and access to remote components, with special emphasis placed on the output flexibility (e.g. multilingualism), where SOAP web services are mainly used for communication. The remote components are represented above all by remote data and mapping services (e.g. meteorological predictions), modelling and analytical systems (currently HEC-HMS, MODFLOW and additional utilities), which support decision making in water management.

  6. PRONET services for distance learning in mammographic image processing.

    PubMed

    Costaridou, L; Panayiotakis, G; Efstratiou, C; Sakellaropoulos, P; Cavouras, D; Kalogeropoulou, C; Varaki, K; Giannakou, L; Dimopoulos, J

    1997-01-01

    The potential of telematics services is investigated with respect to learning needs of medical physicists and biomedical engineers. Telematics services are integrated into a system, the PRONET, which evolves around multimedia computer based courses and distance tutoring support. In addition, information database access and special interest group support are offered. System architecture is based on a component integration approach. The services are delivered in three modes: LAN, ISDN and Internet. Mammographic image processing is selected as an example content area. PMID:10179585

  7. Density estimation of small-mammal populations using a trapping web and distance sampling methods

    USGS Publications Warehouse

    Anderson, David R.; Burnham, Kenneth P.; White, Gary C.; Otis, David L.

    1983-01-01

    Distance sampling methodology is adapted to enable animal density (number per unit of area) to be estimated from capture-recapture and removal data. A trapping web design provides the link between capture data and distance sampling theory. The estimator of density is D = Mt+1f(0), where Mt+1 is the number of individuals captured and f(0) is computed from the Mt+1 distances from the web center to the traps in which those individuals were first captured. It is possible to check qualitatively the critical assumption on which the web design and the estimator are based. This is a conceptual paper outlining a new methodology, not a definitive investigation of the best specific way to implement this method. Several alternative sampling and analysis methods are possible within the general framework of distance sampling theory; a few alternatives are discussed and an example is given.

  8. Experience using web services for biological sequence analysis

    PubMed Central

    Attwood, Teresa; Chohan, Shahid Nadeem; Côté, Richard; Cudré-Mauroux, Philippe; Falquet, Laurent; Fernandes, Pedro; Finn, Robert D.; Hupponen, Taavi; Korpelainen, Eija; Labarga, Alberto; Laugraud, Aurelie; Lima, Tania; Pafilis, Evangelos; Pagni, Marco; Pettifer, Steve; Phan, Isabelle; Rahman, Nazim

    2008-01-01

    Programmatic access to data and tools through the web using so-called web services has an important role to play in bioinformatics. In this article, we discuss the most popular approaches based on SOAP/WS-I and REST and describe our, a cross section of the community, experiences with providing and using web services in the context of biological sequence analysis. We briefly review main technological approaches as well as best practice hints that are useful for both users and developers. Finally, syntactic and semantic data integration issues with multiple web services are discussed. PMID:18621748

  9. Experience using web services for biological sequence analysis.

    PubMed

    Stockinger, Heinz; Attwood, Teresa; Chohan, Shahid Nadeem; Côté, Richard; Cudré-Mauroux, Philippe; Falquet, Laurent; Fernandes, Pedro; Finn, Robert D; Hupponen, Taavi; Korpelainen, Eija; Labarga, Alberto; Laugraud, Aurelie; Lima, Tania; Pafilis, Evangelos; Pagni, Marco; Pettifer, Steve; Phan, Isabelle; Rahman, Nazim

    2008-11-01

    Programmatic access to data and tools through the web using so-called web services has an important role to play in bioinformatics. In this article, we discuss the most popular approaches based on SOAP/WS-I and REST and describe our, a cross section of the community, experiences with providing and using web services in the context of biological sequence analysis. We briefly review main technological approaches as well as best practice hints that are useful for both users and developers. Finally, syntactic and semantic data integration issues with multiple web services are discussed.

  10. A VR-Based Shared Web System for Distance Education

    ERIC Educational Resources Information Center

    Shih, Timothy K.; Chang, Ya-Fung; Hsu, Hun-Hui; Wang, Ying-Hong; Chen, Yung-Hui

    2004-01-01

    Distance education has been an important research issue of multimedia computing and communication. Since the instructional activities are implemented on cyberspace, how to control behaviors of students and to increase the degree of communication awareness have been a challenging issue. We propose a system based on the scaffolding theory. Behaviors…

  11. Web 2.0 Strategy in Libraries and Information Services

    ERIC Educational Resources Information Center

    Byrne, Alex

    2008-01-01

    Web 2.0 challenges libraries to change from their predominantly centralised service models with integrated library management systems at the hub. Implementation of Web 2.0 technologies and the accompanying attitudinal shifts will demand reconceptualisation of the nature of library and information service around a dynamic, ever changing, networked,…

  12. Compression-based aggregation model for medical web services.

    PubMed

    Al-Shammary, Dhiah; Khalil, Ibrahim

    2010-01-01

    Many organizations such as hospitals have adopted Cloud Web services in applying their network services to avoid investing heavily computing infrastructure. SOAP (Simple Object Access Protocol) is the basic communication protocol of Cloud Web services that is XML based protocol. Generally,Web services often suffer congestions and bottlenecks as a result of the high network traffic that is caused by the large XML overhead size. At the same time, the massive load on Cloud Web services in terms of the large demand of client requests has resulted in the same problem. In this paper, two XML-aware aggregation techniques that are based on exploiting the compression concepts are proposed in order to aggregate the medical Web messages and achieve higher message size reduction. PMID:21097152

  13. Analysis Tool Web Services from the EMBL-EBI.

    PubMed

    McWilliam, Hamish; Li, Weizhong; Uludag, Mahmut; Squizzato, Silvano; Park, Young Mi; Buso, Nicola; Cowley, Andrew Peter; Lopez, Rodrigo

    2013-07-01

    Since 2004 the European Bioinformatics Institute (EMBL-EBI) has provided access to a wide range of databases and analysis tools via Web Services interfaces. This comprises services to search across the databases available from the EMBL-EBI and to explore the network of cross-references present in the data (e.g. EB-eye), services to retrieve entry data in various data formats and to access the data in specific fields (e.g. dbfetch), and analysis tool services, for example, sequence similarity search (e.g. FASTA and NCBI BLAST), multiple sequence alignment (e.g. Clustal Omega and MUSCLE), pairwise sequence alignment and protein functional analysis (e.g. InterProScan and Phobius). The REST/SOAP Web Services (http://www.ebi.ac.uk/Tools/webservices/) interfaces to these databases and tools allow their integration into other tools, applications, web sites, pipeline processes and analytical workflows. To get users started using the Web Services, sample clients are provided covering a range of programming languages and popular Web Service tool kits, and a brief guide to Web Services technologies, including a set of tutorials, is available for those wishing to learn more and develop their own clients. Users of the Web Services are informed of improvements and updates via a range of methods.

  14. Enhancing the AliEn Web Service Authentication

    NASA Astrophysics Data System (ADS)

    Zhu, Jianlin; Saiz, Pablo; Carminati, Federico; Betev, Latchezar; Zhou, Daicui; Mendez Lorenzo, Patricia; Grigoras, Alina Gabriela; Grigoras, Costin; Furano, Fabrizio; Schreiner, Steffen; Vladimirovna Datskova, Olga; Sankar Banerjee, Subho; Zhang, Guoping

    2011-12-01

    Web Services are an XML based technology that allow applications to communicate with each other across disparate systems. Web Services are becoming the de facto standard that enable inter operability between heterogeneous processes and systems. AliEn2 is a grid environment based on web services. The AliEn2 services can be divided in three categories: Central services, deployed once per organization; Site services, deployed on each of the participating centers; Job Agents running on the worker nodes automatically. A security model to protect these services is essential for the whole system. Current implementations of web server, such as Apache, are not suitable to be used within the grid environment. Apache with the mod_ssl and OpenSSL only supports the X.509 certificates. But in the grid environment, the common credential is the proxy certificate for the purpose of providing restricted proxy and delegation. An Authentication framework was taken for AliEn2 web services to add the ability to accept X.509 certificates and proxy certificates from client-side to Apache Web Server. The authentication framework could also allow the generation of access control policies to limit access to the AliEn2 web services.

  15. Spatial Data Web Services Pricing Model Infrastructure

    NASA Astrophysics Data System (ADS)

    Ozmus, L.; Erkek, B.; Colak, S.; Cankurt, I.; Bakıcı, S.

    2013-08-01

    The General Directorate of Land Registry and Cadastre (TKGM) which is the leader in the field of cartography largely continues its missions which are; to keep and update land registry and cadastre system of the country under the responsibility of the treasure, to perform transactions related to real estate and to establish Turkish national spatial information system. TKGM a public agency has completed many projects. Such as; Continuously Operating GPS Reference Stations (TUSAGA-Aktif), Geo-Metadata Portal (HBB), Orthophoto-Base Map Production and web services, Completion of Initial Cadastre, Cadastral Renovation Project (TKMP), Land Registry and Cadastre Information System (TAKBIS), Turkish National Spatial Data Infrastructure Project (TNSDI), Ottoman Land Registry Archive Information System (TARBIS). TKGM provides updated map and map information to not only public institutions but also to related society in the name of social responsibility principals. Turkish National Spatial Data Infrastructure activities have been started by the motivation of Circular No. 2003/48 which was declared by Turkish Prime Ministry in 2003 within the context of e-Transformation of Turkey Short-term Action Plan. Action No. 47 in the mentioned action plan implies that "A Feasibility Study shall be made in order to establish the Turkish National Spatial Data Infrastructure" whose responsibility has been given to General Directorate of Land Registry and Cadastre. Feasibility report of NSDI has been completed in 10th of December 2010. After decision of Steering Committee, feasibility report has been send to Development Bank (old name State Planning Organization) for further evaluation. There are two main arrangements with related this project (feasibility report).First; Now there is only one Ministry which is Ministry of Environment and Urbanism responsible for establishment, operating and all national level activities of NSDI. And Second arrangement is related to institutional Level. The

  16. Web Service Reputation Evaluation Based on QoS Measurement

    PubMed Central

    Zhang, Haiteng; Shao, Zhiqing; Zheng, Hong; Zhai, Jie

    2014-01-01

    In the early service transactions, quality of service (QoS) information was published by service provider which was not always true and credible. For better verification the trust of the QoS information was provided by the Web service. In this paper, the factual QoS running data are collected by our WS-QoS measurement tool; based on these objectivity data, an algorithm compares the difference of the offered and measured quality data of the service and gives the similarity, and then a reputation evaluation method computes the reputation level of the Web service based on the similarity. The initial implementation and experiment with three Web services' example show that this approach is feasible and these values can act as the references for subsequent consumers to select the service. PMID:24977194

  17. Processing biological literature with customizable Web services supporting interoperable formats.

    PubMed

    Rak, Rafal; Batista-Navarro, Riza Theresa; Carter, Jacob; Rowley, Andrew; Ananiadou, Sophia

    2014-01-01

    Web services have become a popular means of interconnecting solutions for processing a body of scientific literature. This has fuelled research on high-level data exchange formats suitable for a given domain and ensuring the interoperability of Web services. In this article, we focus on the biological domain and consider four interoperability formats, BioC, BioNLP, XMI and RDF, that represent domain-specific and generic representations and include well-established as well as emerging specifications. We use the formats in the context of customizable Web services created in our Web-based, text-mining workbench Argo that features an ever-growing library of elementary analytics and capabilities to build and deploy Web services straight from a convenient graphical user interface. We demonstrate a 2-fold customization of Web services: by building task-specific processing pipelines from a repository of available analytics, and by configuring services to accept and produce a combination of input and output data interchange formats. We provide qualitative evaluation of the formats as well as quantitative evaluation of automatic analytics. The latter was carried out as part of our participation in the fourth edition of the BioCreative challenge. Our analytics built into Web services for recognizing biochemical concepts in BioC collections achieved the highest combined scores out of 10 participating teams. Database URL: http://argo.nactem.ac.uk.

  18. The impact of web services at the IRIS DMC

    NASA Astrophysics Data System (ADS)

    Weekly, R. T.; Trabant, C. M.; Ahern, T. K.; Stults, M.; Suleiman, Y. Y.; Van Fossen, M.; Weertman, B.

    2015-12-01

    The IRIS Data Management Center (DMC) has served the seismological community for nearly 25 years. In that time we have offered data and information from our archive using a variety of mechanisms ranging from email-based to desktop applications to web applications and web services. Of these, web services have quickly become the primary method for data extraction at the DMC. In 2011, the first full year of operation, web services accounted for over 40% of the data shipped from the DMC. In 2014, over ~450 TB of data was delivered directly to users through web services, representing nearly 70% of all shipments from the DMC that year. In addition to handling requests directly from users, the DMC switched all data extraction methods to use web services in 2014. On average the DMC now handles between 10 and 20 million requests per day submitted to web service interfaces. The rapid adoption of web services is attributed to the many advantages they bring. For users, they provide on-demand data using an interface technology, HTTP, that is widely supported in nearly every computing environment and language. These characteristics, combined with human-readable documentation and existing tools make integration of data access into existing workflows relatively easy. For the DMC, the web services provide an abstraction layer to internal repositories allowing for concentrated optimization of extraction workflow and easier evolution of those repositories. Lending further support to DMC's push in this direction, the core web services for station metadata, timeseries data and event parameters were adopted as standards by the International Federation of Digital Seismograph Networks (FDSN). We expect to continue enhancing existing services and building new capabilities for this platform. For example, the DMC has created a federation system and tools allowing researchers to discover and collect seismic data from data centers running the FDSN-standardized services. A future capability

  19. Ethical Issues in Providing Library Services to Distance Learners

    ERIC Educational Resources Information Center

    Needham, Gill; Johnson, Kay

    2007-01-01

    The authors, library practitioners from either side of the Atlantic Ocean, embarked on a dialogue about the ethical challenges encountered in providing library services to distance learners. Unable to find an existing, appropriate ethical framework for their discussion, they agreed to devise their own, informed by relevant professional codes and…

  20. Storage Viability and Optimization Web Service

    SciTech Connect

    Stadler, Michael; Marnay, Christ; Lai, Judy; Siddiqui, Afzal; Limpaitoon, Tanachai; Phan, Trucy; Megel, Olivier; Chang, Jessica; DeForest, Nicholas

    2010-10-11

    Non-residential sectors offer many promising applications for electrical storage (batteries) and photovoltaics (PVs). However, choosing and operating storage under complex tariff structures poses a daunting technical and economic problem that may discourage potential customers and result in lost carbon and economic savings. Equipment vendors are unlikely to provide adequate environmental analysis or unbiased economic results to potential clients, and are even less likely to completely describe the robustness of choices in the face of changing fuel prices and tariffs. Given these considerations, researchers at Lawrence Berkeley National Laboratory (LBNL) have designed the Storage Viability and Optimization Web Service (SVOW): a tool that helps building owners, operators and managers to decide if storage technologies and PVs merit deeper analysis. SVOW is an open access, web-based energy storage and PV analysis calculator, accessible by secure remote login. Upon first login, the user sees an overview of the parameters: load profile, tariff, technologies, and solar radiation location. Each parameter has a pull-down list of possible predefined inputs and users may upload their own as necessary. Since the non-residential sectors encompass a broad range of facilities with fundamentally different characteristics, the tool starts by asking the users to select a load profile from a limited cohort group of example facilities. The example facilities are categorized according to their North American Industry Classification System (NAICS) code. After the load profile selection, users select a predefined tariff or use the widget to create their own. The technologies and solar radiation menus operate in a similar fashion. After these four parameters have been inputted, the users have to select an optimization setting as well as an optimization objective. The analytic engine of SVOW is LBNL?s Distributed Energy Resources Customer Adoption Model (DER-CAM), which is a mixed

  1. Communicating data quality through Web Map Services

    NASA Astrophysics Data System (ADS)

    Blower, Jon; Roberts, Charles; Griffiths, Guy; Lewis, Jane; Yang, Kevin

    2013-04-01

    The sharing and visualization of environmental data through spatial data infrastructures is becoming increasingly common. However, information about the quality of data is frequently unavailable or presented in an inconsistent fashion. ("Data quality" is a phrase with many possible meanings but here we define it as "fitness for purpose" - therefore different users have different notions of what constitutes a "high quality" dataset.) The GeoViQua project (www.geoviqua.org) is developing means for eliciting, formatting, discovering and visualizing quality information using ISO and Open Geospatial Consortium (OGC) standards. Here we describe one aspect of the innovations of the GeoViQua project. In this presentation, we shall demonstrate new developments in using Web Map Services to communicate data quality at the level of datasets, variables and individual samples. We shall outline a new draft set of conventions (known as "WMS-Q"), which describe a set of rules for using WMS to convey quality information (OGC draft Engineering Report 12-160). We shall demonstrate these conventions through new prototype software, based upon the widely-used ncWMS software, that applies these rules to enable the visualization of uncertainties in raster data such as satellite products and the results of numerical simulations. Many conceptual and practical issues have arisen from these experiments. How can source data be formatted so that a WMS implementation can detect the semantic links between variables (e.g. the links between a mean field and its variance)? The visualization of uncertainty can be a complex task - how can we provide users with the power and flexibility to choose an optimal strategy? How can we maintain compatibility (as far as possible) with existing WMS clients? We explore these questions with reference to existing standards and approaches, including UncertML, NetCDF-U and Styled Layer Descriptors.

  2. Going, going, still there: using the WebCite service to permanently archive cited web pages.

    PubMed

    Eysenbach, Gunther; Trudel, Mathieu

    2005-12-30

    Scholars are increasingly citing electronic "web references" which are not preserved in libraries or full text archives. WebCite is a new standard for citing web references. To "webcite" a document involves archiving the cited Web page through www.webcitation.org and citing the WebCite permalink instead of (or in addition to) the unstable live Web page. This journal has amended its "instructions for authors" accordingly, asking authors to archive cited Web pages before submitting a manuscript. Almost 200 other journals are already using the system. We discuss the rationale for WebCite, its technology, and how scholars, editors, and publishers can benefit from the service. Citing scholars initiate an archiving process of all cited Web references, ideally before they submit a manuscript. Authors of online documents and websites which are expected to be cited by others can ensure that their work is permanently available by creating an archived copy using WebCite and providing the citation information including the WebCite link on their Web document(s). Editors should ask their authors to cache all cited Web addresses (Uniform Resource Locators, or URLs) "prospectively" before submitting their manuscripts to their journal. Editors and publishers should also instruct their copyeditors to cache cited Web material if the author has not done so already. Finally, WebCite can process publisher submitted "citing articles" (submitted for example as eXtensible Markup Language [XML] documents) to automatically archive all cited Web pages shortly before or on publication. Finally, WebCite can act as a focussed crawler, caching retrospectively references of already published articles. Copyright issues are addressed by honouring respective Internet standards (robot exclusion files, no-cache and no-archive tags). Long-term preservation is ensured by agreements with libraries and digital preservation organizations. The resulting WebCite Index may also have applications for research

  3. A Web Service and Interface for Remote Electronic Device Characterization

    ERIC Educational Resources Information Center

    Dutta, S.; Prakash, S.; Estrada, D.; Pop, E.

    2011-01-01

    A lightweight Web Service and a Web site interface have been developed, which enable remote measurements of electronic devices as a "virtual laboratory" for undergraduate engineering classes. Using standard browsers without additional plugins (such as Internet Explorer, Firefox, or even Safari on an iPhone), remote users can control a Keithley…

  4. Service Learning and Building Community with the World Wide Web

    ERIC Educational Resources Information Center

    Longan, Michael W.

    2007-01-01

    The geography education literature touts the World Wide Web (Web) as a revolutionary educational tool, yet most accounts ignore its uses for public communication and creative expression. This article argues that students can be producers of content that is of service to local audiences. Drawing inspiration from the community networking movement,…

  5. Grid-enabled Web Services for Geospatial Interoperability

    NASA Astrophysics Data System (ADS)

    Chen, A.; di, L.; Bai, Y.; Wei, Y.

    2006-05-01

    Geospatial interoperability technology makes better and easier use of the huge volume of distributed heterogeneous geospatial data and services in Earth related science research and applications . Open Geospatial Consortium (OGC) has been developing interoperable Web service specifications, such as Web Coverage Service (WCS), Web Map Service (WMS), Web Feature Service (WFS) and Catalog Service for Web (CSW), for promoting geospatial interoperability in the distributed environment. These specifications are widely used by the geospatial community for sharing data and service. Due to the complex nature of Earth related science research and applications, a geoprocessing task normally composes of many inter-related steps of computations in the web service environment. There is a need for the cooperation and security mechanism between any two geospatial web services. Grid, as a promising e-science infrastructure, promotes and facilitates the secure interoperation and collaboration of distributed heterogeneous resources. In this paper, we discuss the technology for enabling the OGC-based geospatial interoperability in a Globus- based Grid environment. Firstly, a new Grid-enabled catalogue services model for secure registry, discovery and access of geospatial data and service was developed. The model not only combines the information schemas of Grid Metadata Catalog Service (MCS)/Replica Location Service (RSL) and the OGC Catalog Service for Web (CSW), but also exploits the geospatial metadata standards including ISO 19115, ISO 19115- 2, FGDC Content Standard for Geospatial metadata, and NASA ECS Metadata. Based on the model, the Grid- enabled CSW (GCSW) service is developed. The service preserves the OGC CSW interface while providing the naming and location transparency by mapping Grid MCS/RLS information model to OGC CSW information model. Moreover, the OGC CSW model is extended to accommodate more than 40 mandatory metadata elements needed for describing the properties

  6. The Use of WebCT in Distance Learning Course in University of Manchester

    ERIC Educational Resources Information Center

    Ahmad, Rosman; Edwards, Rodger; Tomkinson, Bland

    2006-01-01

    The World Wide Web impacted the educational model and became part of distance education in this early century. There were many changes taking place in higher education for political, economic and educational reasons. New goals and educational objectives are being set within educational institutions. There were particular emphases to produce a more…

  7. The Impact of Web Conferencing Training on Peer Tutors' Attitudes toward Distance Education

    ERIC Educational Resources Information Center

    Dvorak, Johanna; Roessger, Kevin

    2012-01-01

    This study investigated the attitudes of peer tutors who received web conferencing training in preparation for synchronous online tutoring. A quasi-experimental design was employed to evaluate changes in peer tutors' attitudes toward distance learning following participation in an online tutor training program. Peer tutors were found to have: (a)…

  8. Problems of Implementing SCORM in an Enterprise Distance Learning Architecture: SCORM Incompatibility across Multiple Web Domains.

    ERIC Educational Resources Information Center

    Engelbrecht, Jeffrey C.

    2003-01-01

    Delivering content to distant users located in dispersed networks, separated by firewalls and different web domains requires extensive customization and integration. This article outlines some of the problems of implementing the Sharable Content Object Reference Model (SCORM) in the Marine Corps' Distance Learning System (MarineNet) and extends…

  9. A Comparison of Web-Based Concept Mapping Tasks for Alternative Assessment in Distance Teacher Education

    ERIC Educational Resources Information Center

    Oliver, Kevin

    2008-01-01

    Three sections of the same distance education class completed a series of Web-based concept map assessments using one of two methods. Open-ended maps applied in section 1 led students to conduct more relational thinking overall, but variance in map items was very high introducing more subjectivity in scoring. Pre-selected term mapping applied in…

  10. The Design of an Integrated System for Web-Based Distance Education.

    ERIC Educational Resources Information Center

    Wang, Hongxue; Holt, Pete

    This paper deals with the design of an integrated system for Web-based distance education (ISWBDE). At the highest level, this system can be dissected into two sub-systems: an integrated course authoring system (ICAS) and an integrated course delivery system (ICDS). The ICAS is designed to help with, or automate the creation and management of…

  11. Faculty Adoption Behaviour about Web-Based Distance Education: A Case Study from China Agricultural University

    ERIC Educational Resources Information Center

    Li, Yan; Lindner, James R.

    2007-01-01

    The purpose of this study was to determine China Agricultural University's (CAU's) faculty adoption behaviour about web-based distance education (WBDE). Rogers' (2003) model of five stages in the innovation-decision process was adopted and modified as the theoretical base for the study. Quantitative research was employed and the research design…

  12. SSWAP: A Simple Semantic Web Architecture and Protocol for Semantic Web Services

    Technology Transfer Automated Retrieval System (TEKTRAN)

    SSWAP (Simple Semantic Web Architecture and Protocol) is an architecture, protocol, and platform for using reasoning to semantically integrate heterogeneous disparate data and services on the web. SSWAP is the driving technology behind the Virtual Plant Information Network, an NSF-funded semantic w...

  13. A Source Extraction Web Service with Cross Matching Capability

    NASA Astrophysics Data System (ADS)

    Krughoff, K. S.; Connolly, A. J.; Colberg, J. M.; O'Mullane, W.; Williams, R. D.

    2005-12-01

    Web services provide an excellent means for users to take advantage of algorithms and computing time on machines that are physically displaced from their location. We can construct tools that are essentially platform and language independent by using XML, specifically SOAP, as the communication language between client and server. This allows for a large set of utilities to be built for use by the entire astronomical community via simple client programs. We present a case study in which we have deployed the popular source extraction code ``SExtractor" in a web service. ``SExtractor" takes an input FITS image and output a catalog of sources with associated parameters. Our web service provides the additional option of cross matching the resultant catalog with other previously published data sets (i.e. SDSS, 2MASS, FIRST, HDF, etc.) through the SkyPortal web service provided by OpenSkyQuery. This is an important step in the development of web services as we begin to use pre-existing web services to build increasing utility into the set of tools available.

  14. Building asynchronous geospatial processing workflows with web services

    NASA Astrophysics Data System (ADS)

    Zhao, Peisheng; Di, Liping; Yu, Genong

    2012-02-01

    Geoscience research and applications often involve a geospatial processing workflow. This workflow includes a sequence of operations that use a variety of tools to collect, translate, and analyze distributed heterogeneous geospatial data. Asynchronous mechanisms, by which clients initiate a request and then resume their processing without waiting for a response, are very useful for complicated workflows that take a long time to run. Geospatial contents and capabilities are increasingly becoming available online as interoperable Web services. This online availability significantly enhances the ability to use Web service chains to build distributed geospatial processing workflows. This paper focuses on how to orchestrate Web services for implementing asynchronous geospatial processing workflows. The theoretical bases for asynchronous Web services and workflows, including asynchrony patterns and message transmission, are examined to explore different asynchronous approaches to and architecture of workflow code for the support of asynchronous behavior. A sample geospatial processing workflow, issued by the Open Geospatial Consortium (OGC) Web Service, Phase 6 (OWS-6), is provided to illustrate the implementation of asynchronous geospatial processing workflows and the challenges in using Web Services Business Process Execution Language (WS-BPEL) to develop them.

  15. Provenance-Based Approaches to Semantic Web Service Discovery and Usage

    ERIC Educational Resources Information Center

    Narock, Thomas William

    2012-01-01

    The World Wide Web Consortium defines a Web Service as "a software system designed to support interoperable machine-to-machine interaction over a network." Web Services have become increasingly important both within and across organizational boundaries. With the recent advent of the Semantic Web, web services have evolved into semantic…

  16. Reinforcement Learning Based Web Service Compositions for Mobile Business

    NASA Astrophysics Data System (ADS)

    Zhou, Juan; Chen, Shouming

    In this paper, we propose a new solution to Reactive Web Service Composition, via molding with Reinforcement Learning, and introducing modified (alterable) QoS variables into the model as elements in the Markov Decision Process tuple. Moreover, we give an example of Reactive-WSC-based mobile banking, to demonstrate the intrinsic capability of the solution in question of obtaining the optimized service composition, characterized by (alterable) target QoS variable sets with optimized values. Consequently, we come to the conclusion that the solution has decent potentials in boosting customer experiences and qualities of services in Web Services, and those in applications in the whole electronic commerce and business sector.

  17. Scalable web services for the PSIPRED Protein Analysis Workbench.

    PubMed

    Buchan, Daniel W A; Minneci, Federico; Nugent, Tim C O; Bryson, Kevin; Jones, David T

    2013-07-01

    Here, we present the new UCL Bioinformatics Group's PSIPRED Protein Analysis Workbench. The Workbench unites all of our previously available analysis methods into a single web-based framework. The new web portal provides a greatly streamlined user interface with a number of new features to allow users to better explore their results. We offer a number of additional services to enable computationally scalable execution of our prediction methods; these include SOAP and XML-RPC web server access and new HADOOP packages. All software and services are available via the UCL Bioinformatics Group website at http://bioinf.cs.ucl.ac.uk/. PMID:23748958

  18. Business Systems Branch Abilities, Capabilities, and Services Web Page

    NASA Technical Reports Server (NTRS)

    Cortes-Pena, Aida Yoguely

    2009-01-01

    During the INSPIRE summer internship I acted as the Business Systems Branch Capability Owner for the Kennedy Web-based Initiative for Communicating Capabilities System (KWICC), with the responsibility of creating a portal that describes the services provided by this Branch. This project will help others achieve a clear view ofthe services that the Business System Branch provides to NASA and the Kennedy Space Center. After collecting the data through the interviews with subject matter experts and the literature in Business World and other web sites I identified discrepancies, made the necessary corrections to the sites and placed the information from the report into the KWICC web page.

  19. Guiding Principles for Facilitating Higher Levels of Web-Based Distance Teaching and Learning in Post-Secondary Settings

    ERIC Educational Resources Information Center

    Kanuka, Heather

    2002-01-01

    The purpose of this study was to develop guiding principles to encourage higher levels of teaching and learning in Web-based distance education. The research framework used Zetterberg's (1962) model for change. Data from semi-structured interviews with university instructors who had experience in teaching Web-based distance education courses, a…

  20. A SOAP Web Service for accessing MODIS land product subsets

    SciTech Connect

    SanthanaVannan, Suresh K; Cook, Robert B; Pan, Jerry Yun; Wilson, Bruce E

    2011-01-01

    Remote sensing data from satellites have provided valuable information on the state of the earth for several decades. Since March 2000, the Moderate Resolution Imaging Spectroradiometer (MODIS) sensor on board NASA s Terra and Aqua satellites have been providing estimates of several land parameters useful in understanding earth system processes at global, continental, and regional scales. However, the HDF-EOS file format, specialized software needed to process the HDF-EOS files, data volume, and the high spatial and temporal resolution of MODIS data make it difficult for users wanting to extract small but valuable amounts of information from the MODIS record. To overcome this usability issue, the NASA-funded Distributed Active Archive Center (DAAC) for Biogeochemical Dynamics at Oak Ridge National Laboratory (ORNL) developed a Web service that provides subsets of MODIS land products using Simple Object Access Protocol (SOAP). The ORNL DAAC MODIS subsetting Web service is a unique way of serving satellite data that exploits a fairly established and popular Internet protocol to allow users access to massive amounts of remote sensing data. The Web service provides MODIS land product subsets up to 201 x 201 km in a non-proprietary comma delimited text file format. Users can programmatically query the Web service to extract MODIS land parameters for real time data integration into models, decision support tools or connect to workflow software. Information regarding the MODIS SOAP subsetting Web service is available on the World Wide Web (WWW) at http://daac.ornl.gov/modiswebservice.

  1. Video in Distance Education: ITFS vs. Web-Streaming--Evaluation of Student Attitudes

    ERIC Educational Resources Information Center

    Reisslein, Jana; Seeling, Patrick; Reisslein, Martin

    2005-01-01

    The use of video in distance education courses has a long tradition, with many colleges and universities having been delivering distance education courses with video since the 80's using the Instructional Television Fixed Service (ITFS) and cable television. With the emergence of the Internet and the increased access bandwidths from private homes…

  2. 78 FR 42537 - Agency Information Collection Activities: Online Survey of Web Services Employers; New...

    Federal Register 2010, 2011, 2012, 2013, 2014

    2013-07-16

    ... of Web Services Employers; New Information Collection; Comment Request ACTION: 60-Day Notice. The... information collection. (2) Title of the Form/Collection: Online Survey of Web Services Employers. (3) Agency... obtains data on the E-Verify Program Web Services. Gaining an understanding of the Web Services...

  3. BPELPower—A BPEL execution engine for geospatial web services

    NASA Astrophysics Data System (ADS)

    Yu, Genong (Eugene); Zhao, Peisheng; Di, Liping; Chen, Aijun; Deng, Meixia; Bai, Yuqi

    2012-10-01

    The Business Process Execution Language (BPEL) has become a popular choice for orchestrating and executing workflows in the Web environment. As one special kind of scientific workflow, geospatial Web processing workflows are data-intensive, deal with complex structures in data and geographic features, and execute automatically with limited human intervention. To enable the proper execution and coordination of geospatial workflows, a specially enhanced BPEL execution engine is required. BPELPower was designed, developed, and implemented as a generic BPEL execution engine with enhancements for executing geospatial workflows. The enhancements are especially in its capabilities in handling Geography Markup Language (GML) and standard geospatial Web services, such as the Web Processing Service (WPS) and the Web Feature Service (WFS). BPELPower has been used in several demonstrations over the decade. Two scenarios were discussed in detail to demonstrate the capabilities of BPELPower. That study showed a standard-compliant, Web-based approach for properly supporting geospatial processing, with the only enhancement at the implementation level. Pattern-based evaluation and performance improvement of the engine are discussed: BPELPower directly supports 22 workflow control patterns and 17 workflow data patterns. In the future, the engine will be enhanced with high performance parallel processing and broad Web paradigms.

  4. Architecture-Based Reliability Analysis of Web Services

    ERIC Educational Resources Information Center

    Rahmani, Cobra Mariam

    2012-01-01

    In a Service Oriented Architecture (SOA), the hierarchical complexity of Web Services (WS) and their interactions with the underlying Application Server (AS) create new challenges in providing a realistic estimate of WS performance and reliability. The current approaches often treat the entire WS environment as a black-box. Thus, the sensitivity…

  5. Advancing Your Library's Web-Based Services. ERIC Digest.

    ERIC Educational Resources Information Center

    Feldman, Sari; Strobel, Tracy

    This digest discusses the development of World Wide Web-based services for libraries and provides examples from the Cleveland Public Library (CPL). The first section highlights the importance of developing such services, steps to be followed for a successful project, and the importance of having the goal of replicating and enhancing traditional…

  6. SSWAP: A Simple Semantic Web Architecture and Protocol for semantic web services

    PubMed Central

    Gessler, Damian DG; Schiltz, Gary S; May, Greg D; Avraham, Shulamit; Town, Christopher D; Grant, David; Nelson, Rex T

    2009-01-01

    Background SSWAP (Simple Semantic Web Architecture and Protocol; pronounced "swap") is an architecture, protocol, and platform for using reasoning to semantically integrate heterogeneous disparate data and services on the web. SSWAP was developed as a hybrid semantic web services technology to overcome limitations found in both pure web service technologies and pure semantic web technologies. Results There are currently over 2400 resources published in SSWAP. Approximately two dozen are custom-written services for QTL (Quantitative Trait Loci) and mapping data for legumes and grasses (grains). The remaining are wrappers to Nucleic Acids Research Database and Web Server entries. As an architecture, SSWAP establishes how clients (users of data, services, and ontologies), providers (suppliers of data, services, and ontologies), and discovery servers (semantic search engines) interact to allow for the description, querying, discovery, invocation, and response of semantic web services. As a protocol, SSWAP provides the vocabulary and semantics to allow clients, providers, and discovery servers to engage in semantic web services. The protocol is based on the W3C-sanctioned first-order description logic language OWL DL. As an open source platform, a discovery server running at (as in to "swap info") uses the description logic reasoner Pellet to integrate semantic resources. The platform hosts an interactive guide to the protocol at , developer tools at , and a portal to third-party ontologies at (a "swap meet"). Conclusion SSWAP addresses the three basic requirements of a semantic web services architecture (i.e., a common syntax, shared semantic, and semantic discovery) while addressing three technology limitations common in distributed service systems: i.e., i) the fatal mutability of traditional interfaces, ii) the rigidity and fragility of static subsumption hierarchies, and iii) the confounding of content, structure, and presentation. SSWAP is novel by establishing

  7. OntoGene web services for biomedical text mining

    PubMed Central

    2014-01-01

    Text mining services are rapidly becoming a crucial component of various knowledge management pipelines, for example in the process of database curation, or for exploration and enrichment of biomedical data within the pharmaceutical industry. Traditional architectures, based on monolithic applications, do not offer sufficient flexibility for a wide range of use case scenarios, and therefore open architectures, as provided by web services, are attracting increased interest. We present an approach towards providing advanced text mining capabilities through web services, using a recently proposed standard for textual data interchange (BioC). The web services leverage a state-of-the-art platform for text mining (OntoGene) which has been tested in several community-organized evaluation challenges, with top ranked results in several of them. PMID:25472638

  8. Finding, Browsing and Getting Data Easily Using SPDF Web Services

    NASA Technical Reports Server (NTRS)

    Candey, R.; Chimiak, R.; Harris, B.; Johnson, R.; Kovalick, T.; Lal, N.; Leckner, H.; Liu, M.; McGuire, R.; Papitashvili, N.; Roberts, A.

    2010-01-01

    The NASA GSFC Space Physics Data Facility (5PDF) provides heliophysics science-enabling information services for enhancing scientific research and enabling integration of these services into the Heliophysics Data Environment paradigm, via standards-based approach (SOAP) and Representational State Transfer (REST) web services in addition to web browser, FTP, and OPeNDAP interfaces. We describe these interfaces and the philosophies behind these web services, and show how to call them from various languages, such as IDL and Perl. We are working towards a "one simple line to call" philosophy extolled in the recent VxO discussions. Combining data from many instruments and missions enables broad research analysis and correlation and coordination with other experiments and missions.

  9. Operational Use of OGC Web Services at the Met Office

    NASA Astrophysics Data System (ADS)

    Wright, Bruce

    2010-05-01

    The Met Office has adopted the Service-Orientated Architecture paradigm to deliver services to a range of customers through Rich Internet Applications (RIAs). The approach uses standard Open Geospatial Consortium (OGC) web services to provide information to web-based applications through a range of generic data services. "Invent", the Met Office beta site, is used to showcase Met Office future plans for presenting web-based weather forecasts, product and information to the public. This currently hosts a freely accessible Weather Map Viewer, written in JavaScript, which accesses a Web Map Service (WMS), to deliver innovative web-based visualizations of weather and its potential impacts to the public. The intention is to engage the public in the development of new web-based services that more accurately meet their needs. As the service is intended for public use within the UK, it has been designed to support a user base of 5 million, the analysed level of UK web traffic reaching the Met Office's public weather information site. The required scalability has been realised through the use of multi-tier tile caching: - WMS requests are made for 256x256 tiles for fixed areas and zoom levels; - a Tile Cache, developed in house, efficiently serves tiles on demand, managing WMS request for the new tiles; - Edge Servers, externally hosted by Akamai, provide a highly scalable (UK-centric) service for pre-cached tiles, passing new requests to the Tile Cache; - the Invent Weather Map Viewer uses the Google Maps API to request tiles from Edge Servers. (We would expect to make use of the Web Map Tiling Service, when it becomes an OGC standard.) The Met Office delivers specialist commercial products to market sectors such as transport, utilities and defence, which exploit a Web Feature Service (WFS) for data relating forecasts and observations to specific geographic features, and a Web Coverage Service (WCS) for sub-selections of gridded data. These are locally rendered as maps or

  10. Proposal for a Web Encoding Service (wes) for Spatial Data Transactio

    NASA Astrophysics Data System (ADS)

    Siew, C. B.; Peters, S.; Rahman, A. A.

    2015-10-01

    Web services utilizations in Spatial Data Infrastructure (SDI) have been well established and standardized by Open Geospatial Consortium (OGC). Similar web services for 3D SDI are also being established in recent years, with extended capabilities to handle 3D spatial data. The increasing popularity of using City Geographic Markup Language (CityGML) for 3D city modelling applications leads to the needs for large spatial data handling for data delivery. This paper revisits the available web services in OGC Web Services (OWS), and propose the background concepts and requirements for encoding spatial data via Web Encoding Service (WES). Furthermore, the paper discusses the data flow of the encoder within web service, e.g. possible integration with Web Processing Service (WPS) or Web 3D Services (W3DS). The integration with available web service could be extended to other available web services for efficient handling of spatial data, especially 3D spatial data.

  11. Web Service for Positional Quality Assessment: the Wps Tier

    NASA Astrophysics Data System (ADS)

    Xavier, E. M. A.; Ariza-López, F. J.; Ureña-Cámara, M. A.

    2015-08-01

    In the field of spatial data every day we have more and more information available, but we still have little or very little information about the quality of spatial data. We consider that the automation of the spatial data quality assessment is a true need for the geomatic sector, and that automation is possible by means of web processing services (WPS), and the application of specific assessment procedures. In this paper we propose and develop a WPS tier centered on the automation of the positional quality assessment. An experiment using the NSSDA positional accuracy method is presented. The experiment involves the uploading by the client of two datasets (reference and evaluation data). The processing is to determine homologous pairs of points (by distance) and calculate the value of positional accuracy under the NSSDA standard. The process generates a small report that is sent to the client. From our experiment, we reached some conclusions on the advantages and disadvantages of WPSs when applied to the automation of spatial data accuracy assessments.

  12. Semantic Web Services Challenge, Results from the First Year. Series: Semantic Web And Beyond, Volume 8.

    NASA Astrophysics Data System (ADS)

    Petrie, C.; Margaria, T.; Lausen, H.; Zaremba, M.

    Explores trade-offs among existing approaches. Reveals strengths and weaknesses of proposed approaches, as well as which aspects of the problem are not yet covered. Introduces software engineering approach to evaluating semantic web services. Service-Oriented Computing is one of the most promising software engineering trends because of the potential to reduce the programming effort for future distributed industrial systems. However, only a small part of this potential rests on the standardization of tools offered by the web services stack. The larger part of this potential rests upon the development of sufficient semantics to automate service orchestration. Currently there are many different approaches to semantic web service descriptions and many frameworks built around them. A common understanding, evaluation scheme, and test bed to compare and classify these frameworks in terms of their capabilities and shortcomings, is necessary to make progress in developing the full potential of Service-Oriented Computing. The Semantic Web Services Challenge is an open source initiative that provides a public evaluation and certification of multiple frameworks on common industrially-relevant problem sets. This edited volume reports on the first results in developing common understanding of the various technologies intended to facilitate the automation of mediation, choreography and discovery for Web Services using semantic annotations. Semantic Web Services Challenge: Results from the First Year is designed for a professional audience composed of practitioners and researchers in industry. Professionals can use this book to evaluate SWS technology for their potential practical use. The book is also suitable for advanced-level students in computer science.

  13. The Role of Libraries in Web-Based Distance Education: An Account and an Analysis of the Impact of Web Technology on Distance Learning--What Remains Unchanged, What Is Changing

    ERIC Educational Resources Information Center

    Cooke, Nicole A.

    2004-01-01

    Even though distance education has a long and diverse history, dating back to 1840, in the last ten-to-fifteen years it has been completely transformed by the emergence of Web-based technology. This technology has had an enormous impact on all aspects of distance education (or distance learning as it is increasingly called). In addition to…

  14. CMS data quality monitoring web service

    NASA Astrophysics Data System (ADS)

    Tuura, L.; Eulisse, G.; Meyer, A.

    2010-04-01

    A central component of the data quality monitoring system of the CMS experiment at the Large Hadron Collider is a web site for browsing data quality histograms. The production servers in data taking provide access to several hundred thousand histograms per run, both live in online as well as for up to several terabytes of archived histograms for the online data taking, Tier-0 prompt reconstruction, prompt calibration and analysis activities, for re-reconstruction at Tier-1s and for release validation. At the present usage level the servers currently handle in total around a million authenticated HTTP requests per day. We describe the main features and components of the system, our implementation for web-based interactive rendering, and the server design. We give an overview of the deployment and maintenance procedures. We discuss the main technical challenges and our solutions to them, with emphasis on functionality, long-term robustness and performance.

  15. A web-services approach to modeling global fluid flux

    NASA Astrophysics Data System (ADS)

    Wood, W. T.; Martin, K. M.; Sample, J.; Owens, R.; Jung, W.; Calantoni, J.; Boyd, T. J.; Coffin, R. B.

    2013-12-01

    Regional and global estimates of fluid and solute accumulation and transport require aspects of physics, chemistry, and biology that manifest as geospatial data sets. Recent developments in computer sciences known as web-services allow a new approach to earth modeling that allows data and models from different sources and on different platforms to be linked - allowing continual updates as data are acquired or models upgraded. NRL has begun constructing a such a generic earth modeling system (GEMS) based on OGC and WD3 compliant web-services data exchange, and designed to take advantage of newly developing web-service based data repositories funded by NSF, NOAA, USGS and others. Specifically, global data sets such as those found in the World Ocean Atlas (2009), sediment thickness, crust age, and seafloor temperature that have been used previously to estimate methane hydrate content in the World's ocean can now be combined with models of sedimentary accretion to generate estimates of fluid flux along margins. The data are served up as independent web coverage services, so they can be independently updated as new data arrives. We anticipate the web services approach to will allow a more flexible, dynamic, and resilient global model of fluid and solute flux than was previously possible.

  16. Creating Web-based, multimedia, and interactive courses for distance learning.

    PubMed

    Mills, A C

    2000-01-01

    A case study describes how faculty at Saint Louis University School of Nursing have developed computer-based, multimedia courses for master's and post-master's nursing education. Lectures with slide presentations are recorded digitally and encoded for multimedia streaming over the World Wide Web for distance learning. The technology is explained in detail in terms of the specific technologies used for lecture and course development. How this technology supports graduate student learning is also presented. PMID:10835811

  17. Real-time medical collaboration services over the web.

    PubMed

    Andrikos, Christos; Rassias, Georgios; Tsanakas, Panayiotis; Maglogiannis, Ilias

    2015-08-01

    The gradual shift in modern medical practice, from working alone clinical doctors to MDTs (Multi-Disciplinary Teams), raises the need of online real-time collaboration among geographically distributed medical personnel. The paper presents a Web-based platform, featuring an efficient medical data management and exchange, for hosting real-time collaborative services. The presented work leverages state-of-the-art features of the web (technologies and APIs) to support client-side medical data processing. Moreover, to address the typical bandwidth bottleneck and known scalability issues of centralized data sharing, an indirect RPC (Remote Process Call) scheme is introduced through object synchronization over the WebRTC paradigm.

  18. WebAUGUSTUS—a web service for training AUGUSTUS and predicting genes in eukaryotes

    PubMed Central

    Hoff, Katharina J.; Stanke, Mario

    2013-01-01

    The prediction of protein coding genes is an important step in the annotation of newly sequenced and assembled genomes. AUGUSTUS is one of the most accurate tools for eukaryotic gene prediction. Here, we present WebAUGUSTUS, a web interface for training AUGUSTUS and predicting genes with AUGUSTUS. Depending on the needs of the user, WebAUGUSTUS generates training gene structures automatically. Besides a genome file, either a file with expressed sequence tags or a file with protein sequences is required for this step. Alternatively, it is possible to submit an externally generated training gene structure file and a genome file. The web service optimizes AUGUSTUS parameters and predicts genes with those parameters. WebAUGUSTUS is available at http://bioinf.uni-greifswald.de/webaugustus. PMID:23700307

  19. Exploring Education Major Focused Adult Learners' Perspectives and Practices of Web-Based Distance Education in Sixteen Universities

    ERIC Educational Resources Information Center

    Zhang, Jing

    2009-01-01

    Distance education is not a new concept for all kinds of learners in the modern societies. Many researchers have studied traditional distance education programs for adult learners in the past, but little research has been done on Web-based distance education (WBDE) for adult learners. There are also many popular online universities in the U.S. or…

  20. A Method of EC Model Implementation Using Web Service Functions

    NASA Astrophysics Data System (ADS)

    Kurihara, Jun; Koizumi, Hisao; Ishikawa, Toshiyuki; Dasai, Takashi

    In recent years, advances in computer and communication technology and the associated rapid increase in the number of Internet users are encouraging advances in Electronic Commerce (EC). Business models of EC are being actively developed by many different enterprises and engineers, and implemented in many kinds of fields. Meanwhile Web services that reuse remote components over the Internet are drawing attention. Web services are based on SOAP/WSDL/UDDI and are given an important position as the infrastructure of the EC systems. The article analyzes the functions and structures of various business models, establishing the patterns of their distinctive and common features, and proposes a method of determining the implementation specifications of business models utilizing these patterns and Web service functions. This method has been applied to a parts purchasing system, which is a typical pattern of the B to B (Business to Business) EC applications. The article also discusses the results of evaluating this prototype system.

  1. Web Service Enabled Data Mining Architecture for Digital Observatories

    NASA Astrophysics Data System (ADS)

    Sinha, P.; Kumar, P.; Zaslavsky, I.; Maidment, D.; Mehra, V.; Bajcsy, P.

    2005-12-01

    Several agencies such as USGS provide data through various web-sites. These websites are individually complex to navigate, provide data in incongruous format and require data to be downloaded locally for any processing. Web services have emerged as new tool that provide a means of transferring the data over the Internet using standard messaging system (XML). Web services are independent of operating system or programming language used. One of the potentials of web services lie in getting various heterogeneous data from different data sources directly into the application, such as a data mining system. We have designed a scientific workflow in Data to Knowledge (D2K) platform developed by National Center for Supercomputing Application (NCSA) which given a study domain, retrieves all streamflow and various water quality data using the CUAHSI (Consortium of University for Advancement of Hydrologic Science, Inc.) NWIS web services. Time series data mining in the D2K environment can be performed on a selection of these data. The analyses functionalities are extensive and include exploratory visualization (such as probability density function and box plot) as well as sophisticated pattern recognition such as similarity search in the time series. The goal of the project is to provide a robust tool for digital observatories that will enable us to better understand our natural environment system.

  2. A resource oriented webs service for environmental modeling

    NASA Astrophysics Data System (ADS)

    Ferencik, Ioan

    2013-04-01

    Environmental modeling is a largely adopted practice in the study of natural phenomena. Environmental models can be difficult to build and use and thus sharing them within the community is an important aspect. The most common approach to share a model is to expose it as a web service. In practice the interaction with this web service is cumbersome due to lack of standardized contract and the complexity of the model being exposed. In this work we investigate the use of a resource oriented approach in exposing environmental models as web services. We view a model as a layered resource build atop the object concept from Object Oriented Programming, augmented with persistence capabilities provided by an embedded object database to keep track of its state and implementing the four basic principles of resource oriented architectures: addressability, statelessness, representation and uniform interface. For implementation we use exclusively open source software: Django framework, dyBase object oriented database and Python programming language. We developed a generic framework of resources structured into a hierarchy of types and consequently extended this typology with recurses specific to the domain of environmental modeling. To test our web service we used cURL, a robust command-line based web client.

  3. Web services for open meteorological data in British Columbia

    NASA Astrophysics Data System (ADS)

    Hiebert, J.; Anslow, F. S.

    2012-12-01

    Until recently, British Columbia suffered from a dearth of publicly and easily accessible (open) meteorological data. While Environment Canada (EC) maintains approximately 250 active in situ weather stations, the remaining meteorological and climate data -- which represent the majority of observations made in the province -- have been gathered by the provincial government within several disparate, ministry-specific networks. Those observations have traditionally been either inaccessible to non-government employees or only available on a network-by-network basis by contacting network managers and requesting custom data queries. Under a collaborative agreement between several provincial ministries, private industry and the Pacific Climate Impacts Consortium (PCIC) and with support from EC, the entire province's meteorological archive has been collected into a single database at PCIC and made publicly accessible via web services and open data protocols. In this paper, we describe our web services, built on open-source software, which provide users access to the full catalogue of BC's meteorological observations through a simple user interface. Our geographic web services provide users access to station locations using Open Geospatial Consortium's Web Mapping Service and Web Feature Service protocols. We use OpenDAP to provide users download access to over a century of weather observations through a variety of open formats such as NetCDF, HDF, ASCII, and others. The goals of these web services are twofold. We primarily aim to provide planners, scientists and researchers with timely and comprehensive climate data as conveniently and efficiently as possible. A natural consequence of this is to enable the flexibility to expand the volume and types of data served and to facilitate more sophisticated analysis regarding past and future climate.

  4. Modeling and Analyzing Web Service Behavior with Regular Flow Nets

    NASA Astrophysics Data System (ADS)

    Xie, Jingang; Tan, Qingping; Cao, Guorong

    Web services are emerging as a promising technology for the development of next generation distributed heterogeneous software systems. To support automated service composition and adaptation, there should be a formal approach for modeling Web service behavior. In this paper we present a novel methodology of modeling and analyzing based on regular flow nets—extended from Petri nets and YAWL. Firstly, we motivate the formal definition of regular flow nets. Secondly, the formalism for dealing with symbolic marking is developed and it is used to define symbolic coverability tree. Finally, an algorithm for generating symbolic coverability tree is presented. Using symbolic coverability tree we can analyze the properties of regular flow nets we concerned. The technology of modeling and analyzing we proposed allows us to deal with cyclic services and data dependence among services.

  5. Deploying and sharing U-Compare workflows as web services

    PubMed Central

    2013-01-01

    Background U-Compare is a text mining platform that allows the construction, evaluation and comparison of text mining workflows. U-Compare contains a large library of components that are tuned to the biomedical domain. Users can rapidly develop biomedical text mining workflows by mixing and matching U-Compare’s components. Workflows developed using U-Compare can be exported and sent to other users who, in turn, can import and re-use them. However, the resulting workflows are standalone applications, i.e., software tools that run and are accessible only via a local machine, and that can only be run with the U-Compare platform. Results We address the above issues by extending U-Compare to convert standalone workflows into web services automatically, via a two-click process. The resulting web services can be registered on a central server and made publicly available. Alternatively, users can make web services available on their own servers, after installing the web application framework, which is part of the extension to U-Compare. We have performed a user-oriented evaluation of the proposed extension, by asking users who have tested the enhanced functionality of U-Compare to complete questionnaires that assess its functionality, reliability, usability, efficiency and maintainability. The results obtained reveal that the new functionality is well received by users. Conclusions The web services produced by U-Compare are built on top of open standards, i.e., REST and SOAP protocols, and therefore, they are decoupled from the underlying platform. Exported workflows can be integrated with any application that supports these open standards. We demonstrate how the newly extended U-Compare enhances the cross-platform interoperability of workflows, by seamlessly importing a number of text mining workflow web services exported from U-Compare into Taverna, i.e., a generic scientific workflow construction platform. PMID:23419017

  6. Magallanes: a web services discovery and automatic workflow composition tool

    PubMed Central

    Ríos, Javier; Karlsson, Johan; Trelles, Oswaldo

    2009-01-01

    Background To aid in bioinformatics data processing and analysis, an increasing number of web-based applications are being deployed. Although this is a positive circumstance in general, the proliferation of tools makes it difficult to find the right tool, or more importantly, the right set of tools that can work together to solve real complex problems. Results Magallanes (Magellan) is a versatile, platform-independent Java library of algorithms aimed at discovering bioinformatics web services and associated data types. A second important feature of Magallanes is its ability to connect available and compatible web services into workflows that can process data sequentially to reach a desired output given a particular input. Magallanes' capabilities can be exploited both as an API or directly accessed through a graphic user interface. The Magallanes' API is freely available for academic use, and together with Magallanes application has been tested in MS-Windows™ XP and Unix-like operating systems. Detailed implementation information, including user manuals and tutorials, is available at . Conclusion Different implementations of the same client (web page, desktop applications, web services, etc.) have been deployed and are currently in use in real installations such as the National Institute of Bioinformatics (Spain) and the ACGT-EU project. This shows the potential utility and versatility of the software library, including the integration of novel tools in the domain and with strong evidences in the line of facilitate the automatic discovering and composition of workflows. PMID:19832968

  7. Facility Energy Management via a Commercial Web Service

    SciTech Connect

    Brambley, Michael R.; Katipamula, Srinivas; O'Neill, Patrick

    2005-08-09

    The World Wide Web (the Web) is rapidly transforming many business practices. The purpose of this chapter is to show readers how facility management tools provided by application service providers (ASPs) via the Web may represent a simple, cost-effective solution to their facility and energy management problems. Accessing tools simply by using a web browser and an Internet connection, all participants in the management of a facility can get convenient access to the information and tools to get their needs met more effectively. Further, energy tools that are integrated into these facility management software environments become readily accessible and can be linked to other tools for managing work orders for getting energy-wasting problems fixed. The chapter also describes how the Federal Energy Management Program is working to make such energy tools readily available.

  8. Consuming Web Services: A Yahoo! Newsfeed Reader

    ERIC Educational Resources Information Center

    Dadashzadeh, Mohammad

    2010-01-01

    Service Oriented Architecture (SOA) shows demonstrable signs of simplifying software integration. It provides the necessary framework for building applications that can be integrated and can reduce the cost of integration significantly. Organizations are beginning to architect new integration solutions following the SOA approach. As such,…

  9. A demanding web-based PACS supported by web services technology

    NASA Astrophysics Data System (ADS)

    Costa, Carlos M. A.; Silva, Augusto; Oliveira, José L.; Ribeiro, Vasco G.; Ribeiro, José

    2006-03-01

    During the last years, the ubiquity of web interfaces have pushed practically all PACS suppliers to develop client applications in which clinical practitioners can receive and analyze medical images, using conventional personal computers and Web browsers. However, due to security and performance issues, the utilization of these software packages has been restricted to Intranets. Paradigmatically, one of the most important advantages of digital image systems is to simplify the widespread sharing and remote access of medical data between healthcare institutions. This paper analyses the traditional PACS drawbacks that contribute to their reduced usage in the Internet and describes a PACS based on Web Services technology that supports a customized DICOM encoding syntax and a specific compression scheme providing all historical patient data in a unique Web interface.

  10. Academic Public Service Web Sites and the Future of Virtual Academic Public Service

    ERIC Educational Resources Information Center

    Cohn, Ellen; Hibbitts, Bernard

    2005-01-01

    Some faculty have started to use the Internet as a bridge to the public instead of merely to each other. Leveraging their specialist knowledge and their academic authority against perceived public needs, they have created another type of academic Web site on their institutional servers--the academic public service Web site (APSWS). APSWS is an…

  11. Web services at the European Bioinformatics Institute-2009

    PubMed Central

    Mcwilliam, Hamish; Valentin, Franck; Goujon, Mickael; Li, Weizhong; Narayanasamy, Menaka; Martin, Jenny; Miyar, Teresa; Lopez, Rodrigo

    2009-01-01

    The European Bioinformatics Institute (EMBL-EBI) has been providing access to mainstream databases and tools in bioinformatics since 1997. In addition to the traditional web form based interfaces, APIs exist for core data resources such as EMBL-Bank, Ensembl, UniProt, InterPro, PDB and ArrayExpress. These APIs are based on Web Services (SOAP/REST) interfaces that allow users to systematically access databases and analytical tools. From the user's point of view, these Web Services provide the same functionality as the browser-based forms. However, using the APIs frees the user from web page constraints and are ideal for the analysis of large batches of data, performing text-mining tasks and the casual or systematic evaluation of mathematical models in regulatory networks. Furthermore, these services are widespread and easy to use; require no prior knowledge of the technology and no more than basic experience in programming. In the following we wish to inform of new and updated services as well as briefly describe planned developments to be made available during the course of 2009–2010. PMID:19435877

  12. 3D Medical Volume Reconstruction Using Web Services

    PubMed Central

    Kooper, Rob; Shirk, Andrew; Lee, Sang-Chul; Lin, Amy; Folberg, Robert; Bajcsy, Peter

    2008-01-01

    We address the problem of 3D medical volume reconstruction using web services. The use of proposed web services is motivated by the fact that the problem of 3D medical volume reconstruction requires significant computer resources and human expertise in medical and computer science areas. Web services are implemented as an additional layer to a dataflow framework called Data to Knowledge. In the collaboration between UIC and NCSA, pre-processed input images at NCSA are made accessible to medical collaborators for registration. Every time UIC medical collaborators inspected images and selected corresponding features for registration, the web service at NCSA is contacted and the registration processing query is executed using the Image to Knowledge library of registration methods. Co-registered frames are returned for verification by medical collaborators in a new window. In this paper, we present 3D volume reconstruction problem requirements and the architecture of the developed prototype system at http://isda.ncsa.uiuc.edu/MedVolume. We also explain the tradeoffs of our system design and provide experimental data to support our system implementation. The prototype system has been used for multiple 3D volume reconstructions of blood vessels and vasculogenic mimicry patterns in histological sections of uveal melanoma studied by fluorescent confocal laser scanning microscope. PMID:18336808

  13. SWS: accessing SRS sites contents through Web Services

    PubMed Central

    Romano, Paolo; Marra, Domenico

    2008-01-01

    Background Web Services and Workflow Management Systems can support creation and deployment of network systems, able to automate data analysis and retrieval processes in biomedical research. Web Services have been implemented at bioinformatics centres and workflow systems have been proposed for biological data analysis. New databanks are often developed by taking into account these technologies, but many existing databases do not allow a programmatic access. Only a fraction of available databanks can thus be queried through programmatic interfaces. SRS is a well know indexing and search engine for biomedical databanks offering public access to many databanks and analysis tools. Unfortunately, these data are not easily and efficiently accessible through Web Services. Results We have developed ‘SRS by WS’ (SWS), a tool that makes information available in SRS sites accessible through Web Services. Information on known sites is maintained in a database, srsdb. SWS consists in a suite of WS that can query both srsdb, for information on sites and databases, and SRS sites. SWS returns results in a text-only format and can be accessed through a WSDL compliant client. SWS enables interoperability between workflow systems and SRS implementations, by also managing access to alternative sites, in order to cope with network and maintenance problems, and selecting the most up-to-date among available systems. Conclusions Development and implementation of Web Services, allowing to make a programmatic access to an exhaustive set of biomedical databases can significantly improve automation of in-silico analysis. SWS supports this activity by making biological databanks that are managed in public SRS sites available through a programmatic interface. PMID:18387203

  14. Towards Thematic Web Services for Generic Data Visualization and Analysis

    NASA Astrophysics Data System (ADS)

    Horanont, T.; Basa, M.; Shibasaki, R.

    2012-07-01

    Spatial analysis packages and thematic mapping are available in a number of traditional desktop GIS. However, visualizing thematic maps through the Internet is still limited to fix contents and restrict changes of the input data. The users with limited GIS knowledge or people who do not own digital map data are normally having difficulties to create output thematic maps from generic data. In this study, we developed thematic mapping services that can be applied to non-spatial data format served through powerful map services solutions. Novice users who have no GIS software experience or have no digital base map can simply input a plain text file with location identifier field such as place name or gazetteer to generate thematic maps online. We implemented a prototype by using web service standards recommended by the Open Geospatial Consortium (OGC) such as Web Map Service (WMS), Web Feature Service (WFS) and Styled Layer Descriptor (SLD) to provide a principle for communication and allow users to visualize spatial information as thematic maps. The system dedicates a great deal of effort to the initial study of geospatial analysis and visualization for novice users including those with no past experience using Geographic Information Systems.

  15. Clinical Predictive Modeling Development and Deployment through FHIR Web Services

    PubMed Central

    Khalilia, Mohammed; Choi, Myung; Henderson, Amelia; Iyengar, Sneha; Braunstein, Mark; Sun, Jimeng

    2015-01-01

    Clinical predictive modeling involves two challenging tasks: model development and model deployment. In this paper we demonstrate a software architecture for developing and deploying clinical predictive models using web services via the Health Level 7 (HL7) Fast Healthcare Interoperability Resources (FHIR) standard. The services enable model development using electronic health records (EHRs) stored in OMOP CDM databases and model deployment for scoring individual patients through FHIR resources. The MIMIC2 ICU dataset and a synthetic outpatient dataset were transformed into OMOP CDM databases for predictive model development. The resulting predictive models are deployed as FHIR resources, which receive requests of patient information, perform prediction against the deployed predictive model and respond with prediction scores. To assess the practicality of this approach we evaluated the response and prediction time of the FHIR modeling web services. We found the system to be reasonably fast with one second total response time per patient prediction. PMID:26958207

  16. Web services in the U.S. geological survey streamstats web application

    USGS Publications Warehouse

    Guthrie, J.D.; Dartiguenave, C.; Ries, Kernell G.

    2009-01-01

    StreamStats is a U.S. Geological Survey Web-based GIS application developed as a tool for waterresources planning and management, engineering design, and other applications. StreamStats' primary functionality allows users to obtain drainage-basin boundaries, basin characteristics, and streamflow statistics for gaged and ungaged sites. Recently, Web services have been developed that provide the capability to remote users and applications to access comprehensive GIS tools that are available in StreamStats, including delineating drainage-basin boundaries, computing basin characteristics, estimating streamflow statistics for user-selected locations, and determining point features that coincide with a National Hydrography Dataset (NHD) reach address. For the state of Kentucky, a web service also has been developed that provides users the ability to estimate daily time series of drainage-basin average values of daily precipitation and temperature. The use of web services allows the user to take full advantage of the datasets and processes behind the Stream Stats application without having to develop and maintain them. ?? 2009 IEEE.

  17. Pragmatic service development and customisation with the CEDA OGC Web Services framework

    NASA Astrophysics Data System (ADS)

    Pascoe, Stephen; Stephens, Ag; Lowe, Dominic

    2010-05-01

    The CEDA OGC Web Services framework (COWS) emphasises rapid service development by providing a lightweight layer of OGC web service logic on top of Pylons, a mature web application framework for the Python language. This approach gives developers a flexible web service development environment without compromising access to the full range of web application tools and patterns: Model-View-Controller paradigm, XML templating, Object-Relational-Mapper integration and authentication/authorization. We have found this approach useful for exploring evolving standards and implementing protocol extensions to meet the requirements of operational deployments. This paper outlines how COWS is being used to implement customised WMS, WCS, WFS and WPS services in a variety of web applications from experimental prototypes to load-balanced cluster deployments serving 10-100 simultaneous users. In particular we will cover 1) The use of Climate Science Modeling Language (CSML) in complex-feature aware WMS, WCS and WFS services, 2) Extending WMS to support applications with features specific to earth system science and 3) A cluster-enabled Web Processing Service (WPS) supporting asynchronous data processing. The COWS WPS underpins all backend services in the UK Climate Projections User Interface where users can extract, plot and further process outputs from a multi-dimensional probabilistic climate model dataset. The COWS WPS supports cluster job execution, result caching, execution time estimation and user management. The COWS WMS and WCS components drive the project-specific NCEO and QESDI portals developed by the British Atmospheric Data Centre. These portals use CSML as a backend description format and implement features such as multiple WMS layer dimensions and climatology axes that are beyond the scope of general purpose GIS tools and yet vital for atmospheric science applications.

  18. The Knowledge Base as an Extension of Distance Learning Reference Service

    ERIC Educational Resources Information Center

    Casey, Anne Marie

    2012-01-01

    This study explores knowledge bases as extension of reference services for distance learners. Through a survey and follow-up interviews with distance learning librarians, this paper discusses their interest in creating and maintaining a knowledge base as a resource for reference services to distance learners. It also investigates their perceptions…

  19. Distance health services under the general agreement on trade in services.

    PubMed

    Yeo, Matthew S

    2002-01-01

    This Article discusses the state of distance health with respect to the General Agreement on Trade in Services (GATS). After examining the various aspects of telemedicine and its place in international trade, the author looks at the structure and functioning of GATS and how telemedicine is regulated under this agreement. The author argues that the potential for telemedicine under this agreement has yet to be fulfilled and suggests a number of ways to realize its potential. Ultimately, however, the author concludes that the single most important international trade objective for the United States healthcare industry should be to get its own house in order with respect to cross-border provision of health services. From an international trade perspective, the problem of non-uniform state licensure requirements within the United States makes it very difficult for the United States to negotiate market access commitments for distance health services with other countries. Therefore, it is not realistic to expect significant progress in the liberalization of distance health services until the United States has in place a reasonably uniform domestic system of licensure and regulation for telemedicine practitioners. PMID:11974523

  20. 75 FR 57086 - Submission for Review: Federal Cyber Service: Scholarship for Service (SFS) Registration Web Site

    Federal Register 2010, 2011, 2012, 2013, 2014

    2010-09-17

    ... April 19, 2010 at 75 FR 20400, allowing for a 60-day public comment period. One comment was received... From the Federal Register Online via the Government Publishing Office OFFICE OF PERSONNEL MANAGEMENT Submission for Review: Federal Cyber Service: Scholarship for Service (SFS) Registration Web...

  1. 75 FR 20400 - Submission for Review: Federal Cyber Service: Scholarship for Service (SFS) Registration Web Site

    Federal Register 2010, 2011, 2012, 2013, 2014

    2010-04-19

    ... From the Federal Register Online via the Government Publishing Office ] OFFICE OF PERSONNEL MANAGEMENT Submission for Review: Federal Cyber Service: Scholarship for Service (SFS) Registration Web Site AGENCY: U.S. Office of Personnel Management. ACTION: 60-Day Notice and request for comments. SUMMARY:...

  2. Augmenting Basic Web Data Services with Middleware Services to Facilitate Usability and Interoperability

    NASA Astrophysics Data System (ADS)

    Werpy, J.; Torbert, C.

    2014-12-01

    Over the past few years many Data Providers have implemented services that allow for web based (HTTP) interfaces to manipulate, organize, modify, and deliver Earth Science Data. This web architecture provides the foundation for streamlining of Earth Science Users utilization of and interaction with the Data. However, critical components are missing and need to be developed in order to increase the capabilities, potential, and reach of these services. Middleware services represent a class of Data Services that are able to communicate their capabilities more clearly and effectively with Science Data Users while also leveraging the more raw web services on the back end. A Middleware layer of a services architecture functions to coordinate the interactions of the users with the core web services. This simplifies execution, parameter selection, data integration, data delivery, and data analysis activities. This presentation will outline how the Land Processes Distributed Active Archive Center (LP DAAC) has utilized core services to provide basic access to data, data manipulation, and processing. Beyond that, the presentation will also detail the enhancement of those efforts through the development and implementation of Middleware to augment capabilities and create workflows necessary to enable Science Users to perform meaningful science activities and analysis faster than before. The Middleware layer acts as the "glue" that allows all these separate services to work together. By moving the algorithms that process and organize data closer to the Data Archive and enabling access to them via web services fronted by Middleware services, the LP DAAC helps Science users to do better, less expensive, and more expansive science much faster than they ever could before.

  3. Secure password-based authenticated key exchange for web services

    SciTech Connect

    Liang, Fang; Meder, Samuel; Chevassut, Olivier; Siebenlist, Frank

    2004-11-22

    This paper discusses an implementation of an authenticated key-exchange method rendered on message primitives defined in the WS-Trust and WS-SecureConversation specifications. This IEEE-specified cryptographic method (AuthA) is proven-secure for password-based authentication and key exchange, while the WS-Trust and WS-Secure Conversation are emerging Web Services Security specifications that extend the WS-Security specification. A prototype of the presented protocol is integrated in the WSRF-compliant Globus Toolkit V4. Further hardening of the implementation is expected to result in a version that will be shipped with future Globus Toolkit releases. This could help to address the current unavailability of decent shared-secret-based authentication options in the Web Services and Grid world. Future work will be to integrate One-Time-Password (OTP) features in the authentication protocol.

  4. The Evolution of a National Distance Guidance Service: Trends and Challenges

    ERIC Educational Resources Information Center

    Watts, A. G.; Dent, Gareth

    2008-01-01

    Three trends in the evolution of the UK Learndirect advice service are identified: the partial migration from telephone to web-based services; the trend within the telephone service from information/advice-oriented interventions to more guidance-oriented interventions; and the move from a mainly learning-oriented service to a more career-oriented…

  5. GWASS: GRASS web application software system based on the GeoBrain web service

    NASA Astrophysics Data System (ADS)

    Qiu, Fang; Ni, Feng; Chastain, Bryan; Huang, Haiting; Zhao, Peisheng; Han, Weiguo; Di, Liping

    2012-10-01

    GRASS is a well-known geographic information system developed more than 30 years ago. As one of the earliest GIS systems, GRASS has currently survived mainly as free, open-source desktop GIS software, with users primarily limited to the research community or among programmers who use it to create customized functions. To allow average GIS end users to continue taking advantage of this widely-used software, we developed a GRASS Web Application Software System (GWASS), a distributed, web-based, multi-tiered Geospatial Information System (GIS) built on top of the GeoBrain web service, a project sponsored by NASA using the latest service oriented architecture (SOA). This SOA enabled system offers an effective and practical alternative to current commercial desktop GIS solutions. With GWASS, all geospatial processing and analyses are conducted by the server, so users are not required to install any software at the client side, which reduces the cost of access for users. The only resource needed to use GWASS is an access to the Internet, and anyone who knows how to use a web browser can operate the system. The SOA framework is revitalizing the GRASS as a new means to bring powerful geospatial analysis and resources to more users with concurrent access.

  6. We Cannot See Them, but They Are There: Marketing Library Services to Distance Learners

    ERIC Educational Resources Information Center

    Dermody, Melinda

    2005-01-01

    Distance learners are a unique target-population for the marketing of library services and resources. Because these patrons do not visit the library often, if at all, it is crucial to actively promote the library resources and services available to them. Marketing strategies for distance learning library services need to take a multifaceted…

  7. MyLibrary@LANL: proximity and semi-metric networks for a collaborative and recommender web service.

    SciTech Connect

    Rocha, L. M.; Simas, T.; Rechtsteiner, A.; DiGiacomo, M.; Luce, R. E.

    2005-01-01

    We describe a network approach to building recommendation systems for a WWW service. We employ two different types of weighted graphs in our analysis and development: Proximity graphs, a type of Fuzzy Graphs based on a co-occurrence probability, and semi-metric distance graphs, which do not observe the triangle inequality of Euclidean distances. Both types of graphs are used to develop intelligent recommendation and collaboration systems for the MyLibrary@LANL web service, a user-centered front-end to the Los Alamos National Laboratory's (LANL) digital library collections and WWW resources.

  8. SAS- Semantic Annotation Service for Geoscience resources on the web

    NASA Astrophysics Data System (ADS)

    Elag, M.; Kumar, P.; Marini, L.; Li, R.; Jiang, P.

    2015-12-01

    There is a growing need for increased integration across the data and model resources that are disseminated on the web to advance their reuse across different earth science applications. Meaningful reuse of resources requires semantic metadata to realize the semantic web vision for allowing pragmatic linkage and integration among resources. Semantic metadata associates standard metadata with resources to turn them into semantically-enabled resources on the web. However, the lack of a common standardized metadata framework as well as the uncoordinated use of metadata fields across different geo-information systems, has led to a situation in which standards and related Standard Names abound. To address this need, we have designed SAS to provide a bridge between the core ontologies required to annotate resources and information systems in order to enable queries and analysis over annotation from a single environment (web). SAS is one of the services that are provided by the Geosematnic framework, which is a decentralized semantic framework to support the integration between models and data and allow semantically heterogeneous to interact with minimum human intervention. Here we present the design of SAS and demonstrate its application for annotating data and models. First we describe how predicates and their attributes are extracted from standards and ingested in the knowledge-base of the Geosemantic framework. Then we illustrate the application of SAS in annotating data managed by SEAD and annotating simulation models that have web interface. SAS is a step in a broader approach to raise the quality of geoscience data and models that are published on the web and allow users to better search, access, and use of the existing resources based on standard vocabularies that are encoded and published using semantic technologies.

  9. Leveling the Playing Field for Users with Web Services

    NASA Astrophysics Data System (ADS)

    Trabant, C. M.; Ahern, T. K.; Karstens, R.; Weertman, B.; Suleiman, Y. Y.

    2013-12-01

    The dawn of digital seismological data recording began approximately 4 decades ago. Since then multiple networks of seismological recording stations have and continue to exist. It is common for each network to operate a data center to store and distribute the collected data. Increasingly there are data centers that archive and distribute data produced by multiple networks and organizations. The modern landscape for seismological data users consists of many data centers spread across the globe offering a variety of data. Luckily most of these centers exchange data in standard formats defined by the International Federation of Digital Seismograph Networks (FDSN). Working with our partners in the FDSN, the IRIS Data Management Center (DMC) developed specifications for 3 standard web service interfaces that are intended to provide an abstraction layer on each center's customized data management system. These services provide access to seismological time series data, related metadata and event (earthquake) parameters. An important part of the interface design is to adhere to web standards and common conventions, which allows use of ubiquitous web client software and toolkits. Another critical design criteria is simple usage, we recognize that our user base is scientific data consumers and not necessarily technologists. The IRIS DMC has implemented each of these 3 service interfaces and made the common software components freely available. Under the NSF's EarthScope program and within the international COOPEUS project, the DMC worked with European partners to help install these standardized interfaces on their own data management systems. One key development was the addition of these web services to the SeisComP3 data handling system, which is common in many seismological data centers, especially in Europe. The combination of standardized data formats and access interfaces removes the need for complex request brokers that translate between centers. Instead, it allows

  10. The Web Experiment List: a Web service for the recruitment of participants and archiving of Internet-based experiments.

    PubMed

    Reips, Ulf-Dietrich; Lengler, Ralph

    2005-05-01

    The Web Experiment List (http://genpsylab-wexlist.unizh.ch/), a free Web-based service for the recruitment of participants in Internet-based experiments, is presented. The Web Experiment List also serves as a searchable archive for the research community. It lists more than 250 links to and descriptions of current and past Web experiments. Searches can be conducted by area of research, language, type of study, date, and status (active vs. archived). Data from log file analyses reveal an increasing use of the Web Experiment List and provide a picture of the distribution of the use of the Web experiment method across disciplines. On a general theoretical note, Web services are discussed as a viable software alternative to the traditional program format.

  11. Stakeholder Expectations of Service Quality in a University Web Portal

    NASA Astrophysics Data System (ADS)

    Tate, Mary; Evermann, Joerg; Hope, Beverley; Barnes, Stuart

    Online service quality is a much-studied concept. There is considerable evidence that user expectations and perceptions of self-service and online service quality differ in different business domains. In addition, the nature of online services is continually changing and universities have been at the forefront of this change, with university websites increasingly acting as a portal for a wide range of online transactions for a wide range of stakeholders. In this qualitative study, we conduct focus groups with a range of stakeholders in a university web portal. Our study offers a number of insights into the changing nature of the relationship between organisations and customers. New technologies are influencing customer expectations. Customers increasingly expect organisations to have integrated information systems, and to utilise new technologies such as SMS and web portals. Organisations can be slow to adopt a customer-centric viewpoint, and persist in providing interfaces that are inconsistent or require inside knowledge of organisational structures and processes. This has a negative effect on customer perceptions.

  12. Semantic Web Service Framework to Intelligent Distributed Manufacturing

    SciTech Connect

    Kulvatunyou, Boonserm

    2005-12-01

    As markets become unexpectedly turbulent with a shortened product life cycle and a power shift towards buyers, the need for methods to develop products, production facilities, and supporting software rapidly and cost-effectively is becoming urgent. The use of a loosely integrated virtual enterprise based framework holds the potential of surviving changing market needs. However, its success requires reliable and large-scale interoperation among trading partners via a semantic web of trading partners services whose properties, capabilities, and interfaces are encoded in an unambiguous as well as computer-understandable form. This paper demonstrates a promising approach to integration and interoperation between a design house and a manufacturer that may or may not have prior relationship by developing semantic web services for business and engineering transactions. To this end, detailed activity and information flow diagrams are developed, in which the two trading partners exchange messages and documents. The properties and capabilities of the manufacturer sites are defined using DARPA Agent Markup Language (DAML) ontology definition language. The prototype development of semantic webs shows that enterprises can interoperate widely in an unambiguous and autonomous manner. This contributes towards the realization of virtual enterprises at a low cost.

  13. Geovisualization in the HydroProg web map service

    NASA Astrophysics Data System (ADS)

    Spallek, Waldemar; Wieczorek, Malgorzata; Szymanowski, Mariusz; Niedzielski, Tomasz; Swierczynska, Malgorzata

    2016-04-01

    The HydroProg system, built at the University of Wroclaw (Poland) in frame of the research project no. 2011/01/D/ST10/04171 financed by the National Science Centre of Poland, has been designed for computing predictions of river stages in real time on a basis of multimodelling. This experimental system works on the upper Nysa Klodzka basin (SW Poland) above the gauge in the town of Bardo, with the catchment area of 1744 square kilometres. The system operates in association with the Local System for Flood Monitoring of Klodzko County (LSOP), and produces hydrograph prognoses as well as inundation predictions. For presenting the up-to-date predictions and their statistics in the online mode, the dedicated real-time web map service has been designed. Geovisualisation in the HydroProg map service concerns: interactive maps of study area, interactive spaghetti hydrograms of water level forecasts along with observed river stages, animated images of inundation. The LSOP network offers a high spatial and temporal resolution of observations, as the length of the sampling interval is equal to 15 minutes. The main environmental elements related to hydrological modelling are shown on the main map. This includes elevation data (hillshading and hypsometric tints), rivers and reservoirs as well as catchment boundaries. Furthermore, we added main towns, roads as well as political and administrative boundaries for better map understanding. The web map was designed as a multi-scale representation, with levels of detail and zooming according to scales: 1:100 000, 1:250 000 and 1:500 000. Observations of water level in LSOP are shown on interactive hydrographs for each gauge. Additionally, predictions and some of their statistical characteristics (like prediction errors and Nash-Sutcliffe efficiency) are shown for selected gauges. Finally, predictions of inundation are presented on animated maps which have been added for four experimental sites. The HydroProg system is a strictly

  14. Satellite Technologies and Services: Implications for International Distance Education.

    ERIC Educational Resources Information Center

    Stahmer, Anna

    1987-01-01

    This examination of international distance education and open university applications of communication satellites at the postsecondary level notes activities in less developed countries (LDCs); presents potential models for cooperation; and describes technical systems for distance education, emphasizing satellite technology and possible problems…

  15. Web based aphasia test using service oriented architecture (SOA)

    NASA Astrophysics Data System (ADS)

    Voos, J. A.; Vigliecca, N. S.; Gonzalez, E. A.

    2007-11-01

    Based on an aphasia test for Spanish speakers which analyze the patient's basic resources of verbal communication, a web-enabled software was developed to automate its execution. A clinical database was designed as a complement, in order to evaluate the antecedents (risk factors, pharmacological and medical backgrounds, neurological or psychiatric symptoms, brain injury -anatomical and physiological characteristics, etc) which are necessary to carry out a multi-factor statistical analysis in different samples of patients. The automated test was developed following service oriented architecture and implemented in a web site which contains a tests suite, which would allow both integrating the aphasia test with other neuropsychological instruments and increasing the available site information for scientific research. The test design, the database and the study of its psychometric properties (validity, reliability and objectivity) were made in conjunction with neuropsychological researchers, who participate actively in the software design, based on the patients or other subjects of investigation feedback.

  16. Electronic personal maternity records: Both web and smartphone services.

    PubMed

    Chang, Chung-Wei; Ma, Tien-Yan; Choi, Mei-San; Hsu, Yu-Yun; Tsai, Yi-Jing; Hou, Ting-Wei

    2015-08-01

    This study develops an antenatal care information system to assist women during pregnancy. We designed and implemented the system as both a web-based service and a multi-platform application for smartphones and tablets. The proposed system has three novel features: (1) web-based maternity records, which contains concise explanations of various antenatal screening and diagnostic tests; (2) self-care journals, which allow pregnant women to keep track of their gestational weight gains, blood pressure, fetal movements, and contractions; and (3) health education, which automatically presents detailed information on antenatal care and other pregnancy-related knowledge according to the women's gestational age. A survey was conducted among pregnant women to evaluate the usability and acceptance of the proposed system. In order to prove that the antenatal care was effective, clinical outcomes should be provided and the results are focused on a usability evaluation.

  17. Electronic personal maternity records: Both web and smartphone services.

    PubMed

    Chang, Chung-Wei; Ma, Tien-Yan; Choi, Mei-San; Hsu, Yu-Yun; Tsai, Yi-Jing; Hou, Ting-Wei

    2015-08-01

    This study develops an antenatal care information system to assist women during pregnancy. We designed and implemented the system as both a web-based service and a multi-platform application for smartphones and tablets. The proposed system has three novel features: (1) web-based maternity records, which contains concise explanations of various antenatal screening and diagnostic tests; (2) self-care journals, which allow pregnant women to keep track of their gestational weight gains, blood pressure, fetal movements, and contractions; and (3) health education, which automatically presents detailed information on antenatal care and other pregnancy-related knowledge according to the women's gestational age. A survey was conducted among pregnant women to evaluate the usability and acceptance of the proposed system. In order to prove that the antenatal care was effective, clinical outcomes should be provided and the results are focused on a usability evaluation. PMID:26004998

  18. The Impact of Web-Logs (Blogs) on Student Perceptions of Isolation and Alienation in a Web-Based Distance-Learning Environment

    ERIC Educational Resources Information Center

    Dickey, Michele D.

    2004-01-01

    In the rush to promote the use of computer-mediated technologies for both traditional and distance learning, relatively little research has been conducted about learner feelings of isolation, alienation and frustration. More recent technologies such as web-logs (blogs) may provide a wider range of tools for bridging learners' feelings of…

  19. The Plate Boundary Observatory: Community Focused Web Services

    NASA Astrophysics Data System (ADS)

    Matykiewicz, J.; Anderson, G.; Lee, E.; Hoyt, B.; Hodgkinson, K.; Persson, E.; Wright, J.; Torrez, D.; Jackson, M.

    2006-12-01

    The Plate Boundary Observatory (PBO), part of the NSF-funded EarthScope project, is designed to study the three-dimensional strain field resulting from deformation across the active boundary zone between the Pacific and North American plates in the western United States. To meet these goals, PBO will install 852 continuous GPS stations, 103 borehole strainmeter stations, 28 tiltmeters, and five laser strainmeters, as well as manage data for 209 previously existing continuous GPS stations. UNAVCO provides access to data products from these stations, as well as general information about the PBO project, via the PBO web site (http://pboweb.unavco.org). GPS and strainmeter data products can be found using a variety of channels, including map searches, text searches, and station specific data retrieval. In addition, the PBO construction status is available via multiple mapping interfaces, including custom web based map widgets and Google Earth. Additional construction details can be accessed from PBO operational pages and station specific home pages. The current state of health for the PBO network is available with the statistical snap-shot, full map interfaces, tabular web based reports, and automatic data mining and alerts. UNAVCO is currently working to enhance the community access to this information by developing a web service framework for the discovery of data products, interfacing with operational engineers, and exposing data services to third party participants. In addition, UNAVCO, through the PBO project, provides advanced data management and monitoring systems for use by the community in operating geodetic networks in the United States and beyond. We will demonstrate these systems during the AGU meeting, and we welcome inquiries from the community at any time.

  20. A Security Architecture for Grid-enabling OGC Web Services

    NASA Astrophysics Data System (ADS)

    Angelini, Valerio; Petronzio, Luca

    2010-05-01

    In the proposed presentation we describe an architectural solution for enabling a secure access to Grids and possibly other large scale on-demand processing infrastructures through OGC (Open Geospatial Consortium) Web Services (OWS). This work has been carried out in the context of the security thread of the G-OWS Working Group. G-OWS (gLite enablement of OGC Web Services) is an international open initiative started in 2008 by the European CYCLOPS , GENESI-DR, and DORII Project Consortia in order to collect/coordinate experiences in the enablement of OWS's on top of the gLite Grid middleware. G-OWS investigates the problem of the development of Spatial Data and Information Infrastructures (SDI and SII) based on the Grid/Cloud capacity in order to enable Earth Science applications and tools. Concerning security issues, the integration of OWS compliant infrastructures and gLite Grids needs to address relevant challenges, due to their respective design principles. In fact OWS's are part of a Web based architecture that demands security aspects to other specifications, whereas the gLite middleware implements the Grid paradigm with a strong security model (the gLite Grid Security Infrastructure: GSI). In our work we propose a Security Architectural Framework allowing the seamless use of Grid-enabled OGC Web Services through the federation of existing security systems (mostly web based) with the gLite GSI. This is made possible mediating between different security realms, whose mutual trust is established in advance during the deployment of the system itself. Our architecture is composed of three different security tiers: the user's security system, a specific G-OWS security system, and the gLite Grid Security Infrastructure. Applying the separation-of-concerns principle, each of these tiers is responsible for controlling the access to a well-defined resource set, respectively: the user's organization resources, the geospatial resources and services, and the Grid

  1. Research on land information web query service for public

    NASA Astrophysics Data System (ADS)

    Liang, Dongdong; Li, Lin; Song, Pingchao; Cheng, Yang; Mei, Song; Min, Yuan

    2009-10-01

    With economics developing fast and internet spreading extensively, the public strongly desire to know about land information. Especially, the policy, Land registration information available to the public inquiry approach, has been executed since March 1st, 2003, which gives the Land Department with guidance to build land information web query service for public. Land information web query service for public requires Land Management Department to provide land registration information which contains attribute and graphics information. When it comes to querying attribute information, precise and fuzzy query methods are commonly used in realistic applications. To improve the speed and accuracy of fuzzy query, Chinese word segmentation method is currently used. Especially, there is no previous example by this method used in cadastre information inquiry. Meanwhile, as for querying lands' spatial information, it is necessary to query attribute information before retrieving the actual graphics information. Then turning to the map service, eagle eye can show which part of whole cadastre map the specified cadastre land located in. But it is obvious the display speed of eagle eye is not as fast as that of cadastre map. Hence, we try to implement the multi-level query with frame selection on cadastre map and identify the different cadastre land with different colors, as eagle eye's display and panning speed are also accelerated. The accomplishments of our research have been applied to Land information query system of Ningbo. It is hoped that the solutions in this system will help to develop and study analogous issues.

  2. Diy Geospatial Web Service Chains: Geochaining Make it Easy

    NASA Astrophysics Data System (ADS)

    Wu, H.; You, L.; Gui, Z.

    2011-08-01

    It is a great challenge for beginners to create, deploy and utilize a Geospatial Web Service Chain (GWSC). People in Computer Science are usually not familiar with geospatial domain knowledge. Geospatial practitioners may lack the knowledge about web services and service chains. The end users may lack both. However, integrated visual editing interfaces, validation tools, and oneclick deployment wizards may help to lower the learning curve and improve modelling skills so beginners will have a better experience. GeoChaining is a GWSC modelling tool designed and developed based on these ideas. GeoChaining integrates visual editing, validation, deployment, execution etc. into a unified platform. By employing a Virtual Globe, users can intuitively visualize raw data and results produced by GeoChaining. All of these features allow users to easily start using GWSC, regardless of their professional background and computer skills. Further, GeoChaining supports GWSC model reuse, meaning that an entire GWSC model created or even a specific part can be directly reused in a new model. This greatly improves the efficiency of creating a new GWSC, and also contributes to the sharing and interoperability of GWSC.

  3. Free Factories: Unified Infrastructure for Data Intensive Web Services.

    PubMed

    Zaranek, Alexander Wait; Clegg, Tom; Vandewege, Ward; Church, George M

    2008-05-01

    We introduce the Free Factory, a platform for deploying data-intensive web services using small clusters of commodity hardware and free software. Independently administered virtual machines called Freegols give application developers the flexibility of a general purpose web server, along with access to distributed batch processing, cache and storage services. Each cluster exploits idle RAM and disk space for cache, and reserves disks in each node for high bandwidth storage. The batch processing service uses a variation of the MapReduce model. Virtualization allows every CPU in the cluster to participate in batch jobs. Each 48-node cluster can achieve 4-8 gigabytes per second of disk I/O. Our intent is to use multiple clusters to process hundreds of simultaneous requests on multi-hundred terabyte data sets. Currently, our applications achieve 1 gigabyte per second of I/O with 123 disks by scheduling batch jobs on two clusters, one of which is located in a remote data center. PMID:20514356

  4. Web services interface for Space Weather: NeQuick 2 web and experimental TEC Calibration

    NASA Astrophysics Data System (ADS)

    Migoya Orue, Yenca O.; Nava, Bruno; Radicella, Sandro M.; Alazo Cuartas, Katy; Luigi, Ciraolo

    2013-04-01

    A web front-end has been recently developed and released to allow retrieving and plotting ionospheric parameters computed by the latest version of the model, NeQuick 2. NeQuick is a quick-run ionospheric electron density model particularly designed for trans-ionospheric propagation applications. It has been developed at the Aeronomy and Radiopropagation Laboratory (now T/ICT4D Laboratory) of the Abdus Salam International Centre for Theoretical Physics (ICTP) - Trieste, Italy with the collaboration of the Institute for Geophysics, Astrophysics and Meteorology (IGAM) of the University of Graz, Austria. To describe the electron density of the ionosphere up to the peak of the F2 layer, NeQuick uses a profile formulation which includes five semi-Epstein layers with modelled thickness parameters. Through a simple web interface users can exploit all the model features including the possibility of computing the electron density and visualizing the corresponding Total Electron Content (TEC) along any ground-to-satellite straight line ray-path. Indeed, the TEC is the ionospheric parameter retrieved from the GPS measurements. It complements the experimental data obtained with diverse kinds of sensors and can be considered a major source of ionospheric information. Since the TEC is not a direct measurement, a "de-biasing" procedure or calibration has to be applied to obtain the relevant values from the raw GPS observables. Using the observation and navigation RINEX files corresponding to a single receiver as input data, the web application allows the user to compute the slant and/or vertical TEC following the concept of the "arc-by-arc" offsets estimation. The combined use of both tools, freely available from the T/ICT4D Web site, will allow the comparison of experimentally derived slant and vertical TEC with modelled values. An online demonstration of the capabilities of the mentioned web services will be illustrated.

  5. Chapter 18: Web-based Tools - NED VO Services

    NASA Astrophysics Data System (ADS)

    Mazzarella, J. M.; NED Team

    The NASA/IPAC Extragalactic Database (NED) is a thematic, web-based research facility in widespread use by scientists, educators, space missions, and observatory operations for observation planning, data analysis, discovery, and publication of research about objects beyond our Milky Way galaxy. NED is a portal into a systematic fusion of data from hundreds of sky surveys and tens of thousands of research publications. The contents and services span the entire electromagnetic spectrum from gamma rays through radio frequencies, and are continuously updated to reflect the current literature and releases of large-scale sky survey catalogs. NED has been on the Internet since 1990, growing in content, automation and services with the evolution of information technology. NED is the world's largest database of crossidentified extragalactic objects. As of December 2006, the system contains approximately 10 million objects and 15 million multi-wavelength cross-IDs. Over 4 thousand catalogs and published lists covering the entire electromagnetic spectrum have had their objects cross-identified or associated, with fundamental data parameters federated for convenient queries and retrieval. This chapter describes the interoperability of NED services with other components of the Virtual Observatory (VO). Section 1 is a brief overview of the primary NED web services. Section 2 provides a tutorial for using NED services currently available through the NVO Registry. The "name resolver" provides VO portals and related internet services with celestial coordinates for objects specified by catalog identifier (name); any alias can be queried because this service is based on the source cross-IDs established by NED. All major services have been updated to provide output in VOTable (XML) format that can be accessed directly from the NED web interface or using the NVO registry. These include access to images via SIAP, Cone- Search queries, and services providing fundamental, multi

  6. Heterogeneous Database Integration Middleware Based on Web Services

    NASA Astrophysics Data System (ADS)

    Wei, Xianmin

    With the company further accelerating the process of information, the heterogeneity problem of each business system database is more and more salient. How to effectively use these historical data, integration of legacy systems, become a hot spot of integration field. In this paper, how to effectively address heterogeneity of data distribution, to achieve customer access, while maintaining the autonomy of the local system for the objectives of heterogeneous database integration middleware based on Web Services solutions.Applications show that middleware solution has certain feasibility and practicality.

  7. WEB-IS2: Next Generation Web Services Using Amira Visualization Package

    NASA Astrophysics Data System (ADS)

    Yang, X.; Wang, Y.; Bollig, E. F.; Kadlec, B. J.; Garbow, Z. A.; Yuen, D. A.; Erlebacher, G.

    2003-12-01

    Amira (www.amiravis.com) is a powerful 3-D visualization package and has been employed recently by the science and engineering communities to gain insight into their data. We present a new web-based interface to Amira, packaged in a Java applet. We have developed a module called WEB-IS/Amira (WEB-IS2), which provides web-based access to Amira. This tool allows earth scientists to manipulate Amira controls remotely and to analyze, render and view large datasets over the internet, without regard for time or location. This could have important ramifications for GRID computing. The design of our implementation will soon allow multiple users to visually collaborate by manipulating a single dataset through a variety of client devices. These clients will only require a browser capable of displaying Java applets. As the deluge of data continues, innovative solutions that maximize ease of use without sacrificing efficiency or flexibility will continue to gain in importance, particularly in the Earth sciences. Major initiatives, such as Earthscope (http://www.earthscope.org), which will generate at least a terabyte of data daily, stand to profit enormously by a system such as WEB-IS/Amira (WEB-IS2). We discuss our use of SOAP (Livingston, D., Advanced SOAP for Web development, Prentice Hall, 2002), a novel 2-way communication protocol, as a means of providing remote commands, and efficient point-to-point transfer of binary image data. We will present our initial experiences with the use of Naradabrokering (www.naradabrokering.org) as a means to decouple clients and servers. Information is submitted to the system as a published item, while it is retrieved through a subscription mechanisms, via what is known as "topics". These topic headers, their contents, and the list of subscribers are automatically tracked by Naradabrokering. This novel approach promises a high degree of fault tolerance, flexibility with respect to client diversity, and language independence for the

  8. Distance Learning: Opportunities Exist for DOD to Capitalize on Services' Efforts.

    ERIC Educational Resources Information Center

    General Accounting Office, Washington, DC. National Security and International Affairs Div.

    The General Accounting Office reviewed the Department of Defense's distance education (DE) initiatives. Data were gathered through meetings of service officials responsible for the military services' training plans and strategies and a review of services' training plans and related documents. It was discovered that the services have begun…

  9. Reliable Execution Based on CPN and Skyline Optimization for Web Service Composition

    PubMed Central

    Ha, Weitao; Zhang, Guojun

    2013-01-01

    With development of SOA, the complex problem can be solved by combining available individual services and ordering them to best suit user's requirements. Web services composition is widely used in business environment. With the features of inherent autonomy and heterogeneity for component web services, it is difficult to predict the behavior of the overall composite service. Therefore, transactional properties and nonfunctional quality of service (QoS) properties are crucial for selecting the web services to take part in the composition. Transactional properties ensure reliability of composite Web service, and QoS properties can identify the best candidate web services from a set of functionally equivalent services. In this paper we define a Colored Petri Net (CPN) model which involves transactional properties of web services in the composition process. To ensure reliable and correct execution, unfolding processes of the CPN are followed. The execution of transactional composition Web service (TCWS) is formalized by CPN properties. To identify the best services of QoS properties from candidate service sets formed in the TCSW-CPN, we use skyline computation to retrieve dominant Web service. It can overcome that the reduction of individual scores to an overall similarity leads to significant information loss. We evaluate our approach experimentally using both real and synthetically generated datasets. PMID:23935431

  10. Reliable execution based on CPN and skyline optimization for Web service composition.

    PubMed

    Chen, Liping; Ha, Weitao; Zhang, Guojun

    2013-01-01

    With development of SOA, the complex problem can be solved by combining available individual services and ordering them to best suit user's requirements. Web services composition is widely used in business environment. With the features of inherent autonomy and heterogeneity for component web services, it is difficult to predict the behavior of the overall composite service. Therefore, transactional properties and nonfunctional quality of service (QoS) properties are crucial for selecting the web services to take part in the composition. Transactional properties ensure reliability of composite Web service, and QoS properties can identify the best candidate web services from a set of functionally equivalent services. In this paper we define a Colored Petri Net (CPN) model which involves transactional properties of web services in the composition process. To ensure reliable and correct execution, unfolding processes of the CPN are followed. The execution of transactional composition Web service (TCWS) is formalized by CPN properties. To identify the best services of QoS properties from candidate service sets formed in the TCSW-CPN, we use skyline computation to retrieve dominant Web service. It can overcome that the reduction of individual scores to an overall similarity leads to significant information loss. We evaluate our approach experimentally using both real and synthetically generated datasets. PMID:23935431

  11. Transitioning the DSAP infrastructure to a web service environment

    NASA Astrophysics Data System (ADS)

    McGraw, Robert M.; Lammers, Craig

    2006-05-01

    RAM Laboratories and AFRL are developing a software infrastructure to provide a Dynamic Situation Assessment and Prediction (DSAP) capability through the use of an embedded simulation infrastructure that can be linked to real-time Command, Control, Communications, and Computers, Intelligence, Surveillance, and Reconnaissance (C4ISR) sensors and systems and Command and Control (C2) activities. The resulting capabilities will allow Commanders to evaluate and analyze Courses of Action and potential alternatives through real-time and faster-than-real-time simulation via executing multiple plans simultaneously across a computing grid. In order to support users in a distributed C2 operational capacity, the DSAP infrastructure is being web-enabled to support net-centric services and common data formats and specifications that will allow it to support users on the Global Information Grid. This paper reviews DSAP and its underlying Multiple Replication Framework architecture and discusses steps that must be taken to play in a Service-Oriented Architecture.

  12. WS/PIDS: standard interoperable PIDS in web services environments.

    PubMed

    Vasilescu, E; Dorobanţu, M; Govoni, S; Padh, S; Mun, S K

    2008-01-01

    An electronic health record depends on the consistent handling of people's identities within and outside healthcare organizations. Currently, the Person Identification Service (PIDS), a CORBA specification, is the only well-researched standard that meets these needs. In this paper, we introduce WS/PIDS, a PIDS specification for Web Services (WS) that closely matches the original PIDS and improves on it by providing explicit support for medical multimedia attributes. WS/PIDS is currently supported by a test implementation, layered on top of a PIDS back-end, with Java- and NET-based, and Web clients. WS/PIDS is interoperable among platforms; it preserves PIDS semantics to a large extent, and it is intended to be fully compliant with established and emerging WS standards. The specification is open source and immediately usable in dynamic clinical systems participating in grid environments. WS/PIDS has been tested successfully with a comprehensive set of use cases, and it is being used in a clinical research setting.

  13. WBSA: web service for bisulfite sequencing data analysis.

    PubMed

    Liang, Fang; Tang, Bixia; Wang, Yanqing; Wang, Jianfeng; Yu, Caixia; Chen, Xu; Zhu, Junwei; Yan, Jiangwei; Zhao, Wenming; Li, Rujiao

    2014-01-01

    Whole-Genome Bisulfite Sequencing (WGBS) and genome-wide Reduced Representation Bisulfite Sequencing (RRBS) are widely used to study DNA methylation. However, data analysis is complicated, lengthy, and hampered by a lack of seamless analytical pipelines. To address these issues, we developed a convenient, stable, and efficient web service called Web Service for Bisulfite Sequencing Data Analysis (WBSA) to analyze bisulfate sequencing data. WBSA focuses on not only CpG methylation, which is the most common biochemical modification in eukaryotic DNA, but also non-CG methylation, which have been observed in plants, iPS cells, oocytes, neurons and stem cells of human. WBSA comprises three main modules as follows: WGBS data analysis, RRBS data analysis, and differentially methylated region (DMR) identification. The WGBS and RRBS modules execute read mapping, methylation site identification, annotation, and advanced analysis, whereas the DMR module identifies actual DMRs and annotates their correlations to genes. WBSA can be accessed and used without charge either online or local version. WBSA also includes the executables of the Portable Batch System (PBS) and standalone versions that can be downloaded from the website together with the installation instructions. WBSA is available at no charge for academic users at http://wbsa.big.ac.cn.

  14. WBSA: Web Service for Bisulfite Sequencing Data Analysis

    PubMed Central

    Wang, Yanqing; Wang, Jianfeng; Yu, Caixia; Chen, Xu; Zhu, Junwei; Yan, Jiangwei; Zhao, Wenming; Li, Rujiao

    2014-01-01

    Whole-Genome Bisulfite Sequencing (WGBS) and genome-wide Reduced Representation Bisulfite Sequencing (RRBS) are widely used to study DNA methylation. However, data analysis is complicated, lengthy, and hampered by a lack of seamless analytical pipelines. To address these issues, we developed a convenient, stable, and efficient web service called Web Service for Bisulfite Sequencing Data Analysis (WBSA) to analyze bisulfate sequencing data. WBSA focuses on not only CpG methylation, which is the most common biochemical modification in eukaryotic DNA, but also non-CG methylation, which have been observed in plants, iPS cells, oocytes, neurons and stem cells of human. WBSA comprises three main modules as follows: WGBS data analysis, RRBS data analysis, and differentially methylated region (DMR) identification. The WGBS and RRBS modules execute read mapping, methylation site identification, annotation, and advanced analysis, whereas the DMR module identifies actual DMRs and annotates their correlations to genes. WBSA can be accessed and used without charge either online or local version. WBSA also includes the executables of the Portable Batch System (PBS) and standalone versions that can be downloaded from the website together with the installation instructions. WBSA is available at no charge for academic users at http://wbsa.big.ac.cn. PMID:24497972

  15. Design and implementation of CUAHSI WaterML and WaterOneFlow Web Services

    NASA Astrophysics Data System (ADS)

    Valentine, D. W.; Zaslavsky, I.; Whitenack, T.; Maidment, D.

    2007-12-01

    WaterOneFlow is a term for a group of web services created by and for the Consortium of Universities for the Advancement of Hydrologic Science, Inc. (CUAHSI) community. CUAHSI web services facilitate the retrieval of hydrologic observations information from online data sources using the SOAP protocol. CUAHSI Water Markup Language (below referred to as WaterML) is an XML schema defining the format of messages returned by the WaterOneFlow web services. \

  16. ClicO FS: an interactive web-based service of Circos

    PubMed Central

    Cheong, Wei-Hien; Tan, Yung-Chie; Yap, Soon-Joo; Ng, Kee-Peng

    2015-01-01

    Summary: We present ClicO Free Service, an online web-service based on Circos, which provides a user-friendly, interactive web-based interface with configurable features to generate Circos circular plots. Availability and implementation: Online web-service is freely available at http://clicofs.codoncloud.com Contact: soonjoo.yap@codongenomics.com Supplementary information: Supplementary data are available at Bioinformatics online. PMID:26227146

  17. International cooperation in veterinary public health curricula using web-based distance interactive education.

    PubMed

    Lipman, Len J; Barnier, Valérie M; de Balogh, Katalin K

    2003-01-01

    The expanding field of Veterinary Public Health places new demands on the knowledge and skills of veterinarians. Veterinary curricula must therefore adapt to this new profile. Through the introduction of case studies dealing with up-to-date issues, students are being trained to solve (real-life) problems and come up with realistic solutions. At the Department of Public Health and Food Safety of the Veterinary Faculty at the University of Utrecht in the Netherlands, positive experiences have resulted from the new opportunities offered by the use of information and communication technology (ICT) in education. The possibility of creating a virtual classroom on the Internet through the use of WebCT software has enabled teachers and students to tackle emerging issues by working together with students in other countries and across disciplines. This article presents some of these experiences, through which international exchange of ideas and realities were stimulated, in addition to consolidating relations between universities in different countries. Long-distance education methodologies provide an important tool to achieve the increasing need for international cooperation in Veterinary Public Health curricula. PMID:14976622

  18. Web Service Infrastructure for Correcting InSAR Imaging

    NASA Astrophysics Data System (ADS)

    von Allmen, P. A.; Fielding, E. J.; Xing, Z.; Pan, L.; Fishbein, E.

    2011-12-01

    InSAR images can be obtained from satellite radar data by combining signals acquired at two different times along the spacecraft's orbit, at geospatial locations nearly identical. Changes in the propagation of the radar signal from the first acquisition to the second, caused for example by changes in the tropospheric water vapor content, can lead to a deterioration of the quality of the interferometric data analysis. Other extraneous effects such as ocean tidal loading can also lead to errors that reduce the potential science return of InSAR missions. Data from Global Positioning Systems and infrared radiometers are current used on an ad hoc basis for the tropospheric corrections when available, and operational weather forecast was demonstrated to be able to fill in the remaining spatial and temporal gaps. We have developed a set of web services named OSCAR (Online Services for Correcting Atmosphere in Radar) that transparently to the user retrieves remote sensing and weather forecast data and delivers atmospheric radar delays on a latitude longitude grid that can be directly integrated with Interferometric Synthetic Aperture Radar data processing software. We will describe the common web service architecture, relying on RESTful, that we developed to streamline the development of OSCAR's capabilities. We will also discuss the Bayesian averaging process that we use for merging the radiometric data with numerical weather forecast results. Correcting for biases and estimating the error model will be discussed in detail and validation results will be presented. The success of the correction procedure will be demonstrated by using MODIS data and ECMWF model output. We will also outline the extension of our online correction system to include GPS data to automatically correct for biases in the radiometric data, and a model of ocean tidal loading to correct for long wavelength errors near coastal regions.

  19. Customer Decision Making in Web Services with an Integrated P6 Model

    NASA Astrophysics Data System (ADS)

    Sun, Zhaohao; Sun, Junqing; Meredith, Grant

    Customer decision making (CDM) is an indispensable factor for web services. This article examines CDM in web services with a novel P6 model, which consists of the 6 Ps: privacy, perception, propensity, preference, personalization and promised experience. This model integrates the existing 6 P elements of marketing mix as the system environment of CDM in web services. The new integrated P6 model deals with the inner world of the customer and incorporates what the customer think during the DM process. The proposed approach will facilitate the research and development of web services and decision support systems.

  20. Establishing Transportation Framework Services Using the Open Geospatial Consortium Web Feature Service Specification

    NASA Astrophysics Data System (ADS)

    Yang, C.; Wong, D. W.; Phillips, T.; Wright, R. A.; Lindsey, S.; Kafatos, M.

    2005-12-01

    As a teamed partnership of the Center for Earth Observing and Space Research (CEOSR) at George Mason University (GMU), Virginia Department of Transportation (VDOT), Bureau of Transportation Statistics at the Department of Transportation (BTS/DOT), and Intergraph, we established Transportation Framework Data Services using Open Geospatial Consortium (OGC)'s Web Feature Service (WFS) Specification to enable the sharing of transportation data among the federal level with data from BTS/DOT, the state level through VDOT, the industries through Intergraph. CEOSR develops WFS solutions using Intergraph software. Relevant technical documents are also developed and disseminated through the partners. The WFS is integrated with operational geospatial systems at CEOSR and VDOT. CEOSR works with Intergraph on developing WFS solutions and technical documents. GeoMedia WebMap WFS toolkit is used with software and technical support from Intergraph. ESRI ArcIMS WFS connector is used with GMU's campus license of ESRI products. Tested solutions are integrated with framework data service operational systems, including 1) CEOSR's interoperable geospatial information services, FGDC clearinghouse Node, Geospatial One Stop (GOS) portal, and WMS services, 2) VDOT's state transportation data and GIS infrastructure, and 3)BTS/DOT's national transportation data. The project presents: 1) develop and deploy an operational OGC WFS 1.1 interfaces at CEOSR for registering with FGDC/GOS Portal and responding to Web ``POST'' requests for transportation Framework data as listed in Table 1; 2) build the WFS service that can return the data that conform to the drafted ANSI/INCITS L1 Standard (when available) for each identified theme in the format given by OGC Geography Markup Language (GML) Version 3.0 or higher; 3) integrate the OGC WFS with CEOSR's clearinghouse nodes, 4) establish a formal partnership to develop and share WFS-based geospatial interoperability technology among GMU, VDOT, BTS

  1. Towards Semantic Web Services on Large, Multi-Dimensional Coverages

    NASA Astrophysics Data System (ADS)

    Baumann, P.

    2009-04-01

    Observed and simulated data in the Earth Sciences often come as coverages, the general term for space-time varying phenomena as set forth by standardization bodies like the Open GeoSpatial Consortium (OGC) and ISO. Among such data are 1-d time series, 2-D surface data, 3-D surface data time series as well as x/y/z geophysical and oceanographic data, and 4-D metocean simulation results. With increasing dimensionality the data sizes grow exponentially, up to Petabyte object sizes. Open standards for exploiting coverage archives over the Web are available to a varying extent. The OGC Web Coverage Service (WCS) standard defines basic extraction operations: spatio-temporal and band subsetting, scaling, reprojection, and data format encoding of the result - a simple interoperable interface for coverage access. More processing functionality is available with products like Matlab, Grid-type interfaces, and the OGC Web Processing Service (WPS). However, these often lack properties known as advantageous from databases: declarativeness (describe results rather than the algorithms), safe in evaluation (no request can keep a server busy infinitely), and optimizable (enable the server to rearrange the request so as to produce the same result faster). WPS defines a geo-enabled SOAP interface for remote procedure calls. This allows to webify any program, but does not allow for semantic interoperability: a function is identified only by its function name and parameters while the semantics is encoded in the (only human readable) title and abstract. Hence, another desirable property is missing, namely an explicit semantics which allows for machine-machine communication and reasoning a la Semantic Web. The OGC Web Coverage Processing Service (WCPS) language, which has been adopted as an international standard by OGC in December 2008, defines a flexible interface for the navigation, extraction, and ad-hoc analysis of large, multi-dimensional raster coverages. It is abstract in that it

  2. Towards Semantic Web Services on Large, Multi-Dimensional Coverages

    NASA Astrophysics Data System (ADS)

    Baumann, P.

    2009-04-01

    Observed and simulated data in the Earth Sciences often come as coverages, the general term for space-time varying phenomena as set forth by standardization bodies like the Open GeoSpatial Consortium (OGC) and ISO. Among such data are 1-d time series, 2-D surface data, 3-D surface data time series as well as x/y/z geophysical and oceanographic data, and 4-D metocean simulation results. With increasing dimensionality the data sizes grow exponentially, up to Petabyte object sizes. Open standards for exploiting coverage archives over the Web are available to a varying extent. The OGC Web Coverage Service (WCS) standard defines basic extraction operations: spatio-temporal and band subsetting, scaling, reprojection, and data format encoding of the result - a simple interoperable interface for coverage access. More processing functionality is available with products like Matlab, Grid-type interfaces, and the OGC Web Processing Service (WPS). However, these often lack properties known as advantageous from databases: declarativeness (describe results rather than the algorithms), safe in evaluation (no request can keep a server busy infinitely), and optimizable (enable the server to rearrange the request so as to produce the same result faster). WPS defines a geo-enabled SOAP interface for remote procedure calls. This allows to webify any program, but does not allow for semantic interoperability: a function is identified only by its function name and parameters while the semantics is encoded in the (only human readable) title and abstract. Hence, another desirable property is missing, namely an explicit semantics which allows for machine-machine communication and reasoning a la Semantic Web. The OGC Web Coverage Processing Service (WCPS) language, which has been adopted as an international standard by OGC in December 2008, defines a flexible interface for the navigation, extraction, and ad-hoc analysis of large, multi-dimensional raster coverages. It is abstract in that it

  3. A "Virtual Fieldtrip": Service Learning in Distance Education Technical Writing Courses

    ERIC Educational Resources Information Center

    Soria, Krista M.; Weiner, Brad

    2013-01-01

    This mixed-methods experimental study examined the effect of service learning in a distance education technical writing course. Quantitative analysis of data found evidence for a positive relationship between participation in service learning and technical writing learning outcomes. Additionally, qualitative analysis suggests that service learning…

  4. INCLUSive: a web portal and service registry for microarray and regulatory sequence analysis

    PubMed Central

    Coessens, Bert; Thijs, Gert; Aerts, Stein; Marchal, Kathleen; De Smet, Frank; Engelen, Kristof; Glenisson, Patrick; Moreau, Yves; Mathys, Janick; De Moor, Bart

    2003-01-01

    INCLUSive is a suite of algorithms and tools for the analysis of gene expression data and the discovery of cis-regulatory sequence elements. The tools allow normalization, filtering and clustering of microarray data, functional scoring of gene clusters, sequence retrieval, and detection of known and unknown regulatory elements using probabilistic sequence models and Gibbs sampling. All tools are available via different web pages and as web services. The web pages are connected and integrated to reflect a methodology and facilitate complex analysis using different tools. The web services can be invoked using standard SOAP messaging. Example clients are available for download to invoke the services from a remote computer or to be integrated with other applications. All services are catalogued and described in a web service registry. The INCLUSive web portal is available for academic purposes at http://www.esat.kuleuven.ac.be/inclusive. PMID:12824346

  5. The use of geospatial web services for exchanging utilities data

    NASA Astrophysics Data System (ADS)

    Kuczyńska, Joanna

    2013-04-01

    Geographic information technologies and related geo-information systems currently play an important role in the management of public administration in Poland. One of these tasks is to maintain and update Geodetic Evidence of Public Utilities (GESUT), part of the National Geodetic and Cartographic Resource, which contains an important for many institutions information of technical infrastructure. It requires an active exchange of data between the Geodesy and Cartography Documentation Centers and institutions, which administrate transmission lines. The administrator of public utilities, is legally obliged to provide information about utilities to GESUT. The aim of the research work was to develop a universal data exchange methodology, which can be implemented on a variety of hardware and software platforms. This methodology use Unified Modeling Language (UML), eXtensible Markup Language (XML), and Geography Markup Language (GML). The proposed methodology is based on the two different strategies: Model Driven Architecture (MDA) and Service Oriented Architecture (SOA). Used solutions are consistent with the INSPIRE Directive and ISO 19100 series standards for geographic information. On the basis of analysis of the input data structures, conceptual models were built for both databases. Models were written in the universal modeling language: UML. Combined model that defines a common data structure was also built. This model was transformed into developed for the exchange of geographic information GML standard. The structure of the document describing the data that may be exchanged is defined in the .xsd file. Network services were selected and implemented in the system designed for data exchange based on open source tools. Methodology was implemented and tested. Data in the agreed data structure and metadata were set up on the server. Data access was provided by geospatial network services: data searching possibilities by Catalog Service for the Web (CSW), data

  6. Footprint Database and web services for the Herschel space observatory

    NASA Astrophysics Data System (ADS)

    Verebélyi, Erika; Dobos, László; Kiss, Csaba

    2015-08-01

    Using all telemetry and observational meta-data, we created a searchable database of Herschel observation footprints. Data from the Herschel space observatory is freely available for everyone but no uniformly processed catalog of all observations has been published yet. As a first step, we unified the data model for all three Herschel instruments in all observation modes and compiled a database of sky coverage information. As opposed to methods using a pixellation of the sphere, in our database, sky coverage is stored in exact geometric form allowing for precise area calculations. Indexing of the footprints allows for very fast search among observations based on pointing, time, sky coverage overlap and meta-data. This enables us, for example, to find moving objects easily in Herschel fields. The database is accessible via a web site and also as a set of REST web service functions which makes it usable from program clients like Python or IDL scripts. Data is available in various formats including Virtual Observatory standards.

  7. RESTFul based heterogeneous Geoprocessing workflow interoperation for Sensor Web Service

    NASA Astrophysics Data System (ADS)

    Yang, Chao; Chen, Nengcheng; Di, Liping

    2012-10-01

    Advanced sensors on board satellites offer detailed Earth observations. A workflow is one approach for designing, implementing and constructing a flexible and live link between these sensors' resources and users. It can coordinate, organize and aggregate the distributed sensor Web services to meet the requirement of a complex Earth observation scenario. A RESTFul based workflow interoperation method is proposed to integrate heterogeneous workflows into an interoperable unit. The Atom protocols are applied to describe and manage workflow resources. The XML Process Definition Language (XPDL) and Business Process Execution Language (BPEL) workflow standards are applied to structure a workflow that accesses sensor information and one that processes it separately. Then, a scenario for nitrogen dioxide (NO2) from a volcanic eruption is used to investigate the feasibility of the proposed method. The RESTFul based workflows interoperation system can describe, publish, discover, access and coordinate heterogeneous Geoprocessing workflows.

  8. Bridging the Distance: Service Learning in International Perspective

    ERIC Educational Resources Information Center

    Florman, Jean C.; Just, Craig; Naka, Tomomi; Peterson, Jim; Seaba, Hazel H.

    2009-01-01

    In this article, the authors describe how an existing partnership between two communities, one in eastern Iowa and one in Mexico, was turned into a cross-disciplinary and international service learning course for students in the University of Iowa Colleges of Engineering, Pharmacy, and Liberal Arts and Sciences. The projects that students worked…

  9. QoS prediction for Web service in Mobile Internet environment

    NASA Astrophysics Data System (ADS)

    Sun, Qibo; Wang, Lubao; Wang, Shangguang; Ma, You; Hsu, Ching-Hsien

    2016-07-01

    Quality of Services (QoS) prediction plays an important role in Web service recommendation. Many existing Web service QoS prediction approaches are highly accurate and useful in Internet environments. However, the QoS data of Web service in Mobile Internet are notably more volatile, which makes these approaches fail in making accurate QoS predictions of Web services. In this paper, by weakening the volatility of QoS data, we propose an accurate Web service QoS prediction approach based on the collaborative filtering algorithm. This approach contains three processes, that is, QoS preprocessing, user similarity computing and QoS predicting. We have implemented our proposed approach with an experiment based on real-world and synthetic datasets. The results demonstrate that our approach outperforms other approaches in Mobile Internet.

  10. Protecting Database Centric Web Services against SQL/XPath Injection Attacks

    NASA Astrophysics Data System (ADS)

    Laranjeiro, Nuno; Vieira, Marco; Madeira, Henrique

    Web services represent a powerful interface for back-end database systems and are increasingly being used in business critical applications. However, field studies show that a large number of web services are deployed with security flaws (e.g., having SQL Injection vulnerabilities). Although several techniques for the identification of security vulnerabilities have been proposed, developing non-vulnerable web services is still a difficult task. In fact, security-related concerns are hard to apply as they involve adding complexity to already complex code. This paper proposes an approach to secure web services against SQL and XPath Injection attacks, by transparently detecting and aborting service invocations that try to take advantage of potential vulnerabilities. Our mechanism was applied to secure several web services specified by the TPC-App benchmark, showing to be 100% effective in stopping attacks, non-intrusive and very easy to use.

  11. A verification strategy for web services composition using enhanced stacked automata model.

    PubMed

    Nagamouttou, Danapaquiame; Egambaram, Ilavarasan; Krishnan, Muthumanickam; Narasingam, Poonkuzhali

    2015-01-01

    Currently, Service-Oriented Architecture (SOA) is becoming the most popular software architecture of contemporary enterprise applications, and one crucial technique of its implementation is web services. Individual service offered by some service providers may symbolize limited business functionality; however, by composing individual services from different service providers, a composite service describing the intact business process of an enterprise can be made. Many new standards have been defined to decipher web service composition problem namely Business Process Execution Language (BPEL). BPEL provides an initial work for forming an Extended Markup Language (XML) specification language for defining and implementing business practice workflows for web services. The problems with most realistic approaches to service composition are the verification of composed web services. It has to depend on formal verification method to ensure the correctness of composed services. A few research works has been carried out in the literature survey for verification of web services for deterministic system. Moreover the existing models did not address the verification properties like dead transition, deadlock, reachability and safetyness. In this paper, a new model to verify the composed web services using Enhanced Stacked Automata Model (ESAM) has been proposed. The correctness properties of the non-deterministic system have been evaluated based on the properties like dead transition, deadlock, safetyness, liveness and reachability. Initially web services are composed using Business Process Execution Language for Web Service (BPEL4WS) and it is converted into ESAM (combination of Muller Automata (MA) and Push Down Automata (PDA)) and it is transformed into Promela language, an input language for Simple ProMeLa Interpreter (SPIN) tool. The model is verified using SPIN tool and the results revealed better recital in terms of finding dead transition and deadlock in contrast to the

  12. Pre-Service Teachers' Views on Web-Based Classroom Management

    ERIC Educational Resources Information Center

    Boyaci, Adnan

    2010-01-01

    With the invention of World Wide Web in 1992, delivery of distance education via internet and emergency of web-based classrooms have rapidly gained acceptance as an alternative and supplement to traditional face to face classroom instruction (Alavi, Yoo & Vogel, 1997; Rahm & Reed, 1997), which represents a paradigm shift challenging all…

  13. Determinants of Corporate Web Services Adoption: A Survey of Companies in Korea

    ERIC Educational Resources Information Center

    Kim, Daekil

    2010-01-01

    Despite the growing interest and attention from Information Technology researchers and practitioners, empirical research on factors that influence an organization's likelihood of adoption of Web Services has been limited. This study identified the factors influencing Web Services adoption from the perspective of 151 South Korean firms. The…

  14. Sharing geoscience algorithms in a Web service-oriented environment (GRASS GIS example)

    NASA Astrophysics Data System (ADS)

    Li, Xiaoyan; Di, Liping; Han, Weiguo; Zhao, Peisheng; Dadi, Upendra

    2010-08-01

    Effective use of the large amounts of geospatial data available for geospatial research and applications is needed. In this paper, the emerging SOAP-based Web service technologies have been used to develop a large number of standard compliant, chainable geospatial Web services, using existing geospatial modules in software systems or specific geoscientific algorithms. A prototype for wrapping legacy software modules or geoscientific algorithms into loosely coupled Web services is proposed from an implementation viewpoint. Module development for Web services adheres to the Open GIS Consortium (OGC) geospatial implementation and the World Wide Web consortium (W3C) standards. The Web service interfaces are designed using Web Services Description Language (WSDL) documents. This paper presents how the granularity of an individual existing geospatial service module used by other geoscientific workflows is decided. A treatment of concurrence processes and clustered deployment of Web services is used to overcome multi-user access and network speed limit problems. This endeavor should allow extensive use of geoscientific algorithms and geospatial data.

  15. Promoting and tracking the use of hospital library web services by outside entities.

    PubMed

    Leman, Hope

    2010-04-01

    This column describes a process that can be used to market a hospital library Web service for use by external entities and monitor its use by the worldwide audience (e.g., medical, academic and public libraries, offices of research administration). Included are concrete suggestions to help hospital librarians in their efforts to encourage adoption of their Web service by other institutions.

  16. SSE Announcement - New GIS Web Mapping Applications and Services

    Atmospheric Science Data Center

    2016-06-30

    ... If you haven’t already noticed the link to the new SSE-GIS web application on the SSE homepage entitled “GIS Web Mapping Applications and Services”, we invite you to visit the site. The Surface meteorology and Solar Energy (SSE) v1.0.3 Web Mapping Application ...

  17. The Knowledge Web: Learning and Collaborating on the Net. Open and Distance Learning Series.

    ERIC Educational Resources Information Center

    Eisenstadt, Marc, Ed.; Vincent, Tom, Ed.

    This book contains a collection of examples of new and effective uses of the World Wide Web in education from the Knowledge Media Institute (KMi) at the Open University (Great Britain). The publication is organized in three main sections--"Learning Media,""Collaboration and Presence," and "Knowledge Systems on the Web"--and contains the following…

  18. Web-based distance continuing education: a new way of thinking for students and instructors.

    PubMed

    Garrison, J A; Schardt, C; Kochi, J K

    2000-07-01

    As people have more difficulty taking time away from work to attend conferences and workshops, the idea of offering courses via the Web has become more desirable. Addressing a need voiced by Medical Library Association membership, the authors developed a Web-based continuing-education course on the subject of the librarian's role in evidence-based medicine. The aim of the course was to provide medical librarians with a well-constructed, content-rich learning experience available to them at their convenience via the Web. This paper includes a discussion of the considerations that need to be taken into account when developing Web-based courses, the issues that arise when the information delivery changes from face-to-face to online, the changing role of the instructor, and the pros and cons of offering Web-based versus traditional courses. The results of the beta test and future plans for the course are also discussed. PMID:10928706

  19. Understanding Transactional Distance in Web-Based Learning Environments: An Empirical Study

    ERIC Educational Resources Information Center

    Huang, Xiaoxia; Chandra, Aruna; DePaolo, Concetta A.; Simmons, Lakisha L.

    2016-01-01

    Transactional distance is an important pedagogical theory in distance education that calls for more empirical support. The purpose of this study was to verify the theory by operationalizing and examining the relationship of (1) dialogue, structure and learner autonomy to transactional distance, and (2) environmental factors and learner demographic…

  20. Technology News; Distance Education Project: Extending Extension Programming via Telecommunications Technology; [and] Fax for Library Services.

    ERIC Educational Resources Information Center

    Coyle, Larry; Spitzer, Kathleen L.

    1992-01-01

    Three articles discuss (1) the numbers of microcomputers installed in elementary and secondary schools; (2) a distance education project in the Minnesota Extension Service that used a satellite delivery system and integrated it with a computer information network; and (3) the use of facsimile machines for library services. (LRW)

  1. Improving access to space weather data via workflows and web services

    NASA Astrophysics Data System (ADS)

    Sundaravel, Anu Swapna

    The Space Physics Interactive Data Resource (SPIDR) is a web-based interactive tool developed by NOAA's National Geophysical Data Center to provide access to historical space physics datasets. These data sets are widely used by physicists for space weather modeling and predictions. Built on a distributed network of databases and application servers, SPIDR offers services in two ways: via a web page interface and via a web service interface. SPIDR exposes several SOAP-based web services that client applications implement to connect to a number of data sources for data download and processing. At present, the usage of the web services has been difficult, adding unnecessary complexity to client applications and inconvenience to the scientists who want to use these datasets. The purpose of this study focuses on improving SPIDR's web interface to better support data access, integration and display. This is accomplished in two ways: (1) examining the needs of scientists to better understand what web services they require to better access and process these datasets and (2) developing a client application to support SPIDR's SOAP-based services using the Kepler scientific workflow system. To this end, we identified, designed and developed several web services for filtering the existing datasets and created several Kepler workflows to automate routine tasks associated with these datasets. These workflows are a part of the custom NGDC build of the Kepler tool. Scientists are already familiar with Kepler due to its extensive use in this domain. As a result, this approach provides them with tools that are less daunting than raw web services and ultimately more useful and customizable. We evaluated our work by interviewing various scientists who make use of SPIDR and having them use the developed Kepler workflows while recording their feedback and suggestions. Our work has improved SPIDR such that new web services are now available and scientists have access to a desktop

  2. A Privacy Access Control Framework for Web Services Collaboration with Role Mechanisms

    NASA Astrophysics Data System (ADS)

    Liu, Linyuan; Huang, Zhiqiu; Zhu, Haibin

    With the popularity of Internet technology, web services are becoming the most promising paradigm for distributed computing. This increased use of web services has meant that more and more personal information of consumers is being shared with web service providers, leading to the need to guarantee the privacy of consumers. This paper proposes a role-based privacy access control framework for Web services collaboration, it utilizes roles to specify the privacy privileges of services, and considers the impact on the reputation degree of the historic experience of services in playing roles. Comparing to the traditional privacy access control approaches, this framework can make the fine-grained authorization decision, thus efficiently protecting consumers' privacy.

  3. Web Based Rapid Mapping of Disaster Areas using Satellite Images, Web Processing Service, Web Mapping Service, Frequency Based Change Detection Algorithm and J-iView

    NASA Astrophysics Data System (ADS)

    Bandibas, J. C.; Takarada, S.

    2013-12-01

    Timely identification of areas affected by natural disasters is very important for a successful rescue and effective emergency relief efforts. This research focuses on the development of a cost effective and efficient system of identifying areas affected by natural disasters, and the efficient distribution of the information. The developed system is composed of 3 modules which are the Web Processing Service (WPS), Web Map Service (WMS) and the user interface provided by J-iView (fig. 1). WPS is an online system that provides computation, storage and data access services. In this study, the WPS module provides online access of the software implementing the developed frequency based change detection algorithm for the identification of areas affected by natural disasters. It also sends requests to WMS servers to get the remotely sensed data to be used in the computation. WMS is a standard protocol that provides a simple HTTP interface for requesting geo-registered map images from one or more geospatial databases. In this research, the WMS component provides remote access of the satellite images which are used as inputs for land cover change detection. The user interface in this system is provided by J-iView, which is an online mapping system developed at the Geological Survey of Japan (GSJ). The 3 modules are seamlessly integrated into a single package using J-iView, which could rapidly generate a map of disaster areas that is instantaneously viewable online. The developed system was tested using ASTER images covering the areas damaged by the March 11, 2011 tsunami in northeastern Japan. The developed system efficiently generated a map showing areas devastated by the tsunami. Based on the initial results of the study, the developed system proved to be a useful tool for emergency workers to quickly identify areas affected by natural disasters.

  4. Job submission and management through web services: the experience with the CREAM service

    NASA Astrophysics Data System (ADS)

    Aiftimiei, C.; Andreetto, P.; Bertocco, S.; Fina, S. D.; Ronco, S. D.; Dorigo, A.; Gianelle, A.; Marzolla, M.; Mazzucato, M.; Sgaravatto, M.; Verlato, M.; Zangrando, L.; Corvo, M.; Miccio, V.; Sciaba, A.; Cesini, D.; Dongiovanni, D.; Grandi, C.

    2008-07-01

    Modern Grid middleware is built around components providing basic functionality, such as data storage, authentication, security, job management, resource monitoring and reservation. In this paper we describe the Computing Resource Execution and Management (CREAM) service. CREAM provides a Web service-based job execution and management capability for Grid systems; in particular, it is being used within the gLite middleware. CREAM exposes a Web service interface allowing conforming clients to submit and manage computational jobs to a Local Resource Management System. We developed a special component, called ICE (Interface to CREAM Environment) to integrate CREAM in gLite. ICE transfers job submissions and cancellations from the Workload Management System, allowing users to manage CREAM jobs from the gLite User Interface. This paper describes some recent studies aimed at assessing the performance and reliability of CREAM and ICE; those tests have been performed as part of the acceptance tests for integration of CREAM and ICE in gLite. We also discuss recent work towards enhancing CREAM with a BES and JSDL compliant interface.

  5. The Semantic Automated Discovery and Integration (SADI) Web service Design-Pattern, API and Reference Implementation

    PubMed Central

    2011-01-01

    Background The complexity and inter-related nature of biological data poses a difficult challenge for data and tool integration. There has been a proliferation of interoperability standards and projects over the past decade, none of which has been widely adopted by the bioinformatics community. Recent attempts have focused on the use of semantics to assist integration, and Semantic Web technologies are being welcomed by this community. Description SADI - Semantic Automated Discovery and Integration - is a lightweight set of fully standards-compliant Semantic Web service design patterns that simplify the publication of services of the type commonly found in bioinformatics and other scientific domains. Using Semantic Web technologies at every level of the Web services "stack", SADI services consume and produce instances of OWL Classes following a small number of very straightforward best-practices. In addition, we provide codebases that support these best-practices, and plug-in tools to popular developer and client software that dramatically simplify deployment of services by providers, and the discovery and utilization of those services by their consumers. Conclusions SADI Services are fully compliant with, and utilize only foundational Web standards; are simple to create and maintain for service providers; and can be discovered and utilized in a very intuitive way by biologist end-users. In addition, the SADI design patterns significantly improve the ability of software to automatically discover appropriate services based on user-needs, and automatically chain these into complex analytical workflows. We show that, when resources are exposed through SADI, data compliant with a given ontological model can be automatically gathered, or generated, from these distributed, non-coordinating resources - a behaviour we have not observed in any other Semantic system. Finally, we show that, using SADI, data dynamically generated from Web services can be explored in a manner

  6. SOAP based web services and their future role in VO projects

    NASA Astrophysics Data System (ADS)

    Topf, F.; Jacquey, C.; Génot, V.; Cecconi, B.; André, N.; Zhang, T. L.; Kallio, E.; Lammer, H.; Facsko, G.; Stöckler, R.; Khodachenko, M.

    2011-10-01

    Modern state-of-the-art web services are from crucial importance for the interoperability of different VO tools existing in the planetary community. SOAP based web services assure the interconnectability between different data sources and tools by providing a common protocol for communication. This paper will point out a best practice approach with the Automated Multi-Dataset Analysis Tool (AMDA) developed by CDPP, Toulouse and the provision of VEX/MAG data from a remote database located at IWF, Graz. Furthermore a new FP7 project IMPEx will be introduced with a potential usage example of AMDA web services in conjunction with simulation models.

  7. Thinking on sharing 3D GIS data by web service following OGC standard

    NASA Astrophysics Data System (ADS)

    Wu, Lvewei

    2016-03-01

    Sharing based on web service is a well-understood method in GIS. This paper mainly study on how to share 3D GIS data in this way. Based on the analysis of current situation, this paper concludes the existing problem and put forward the idea of sharing 3D by web service following OGC standard. Besides, feasibility and practicality of the idea are verified by relevant experimental evidence. In the end, a framework conforming to the idea above is present, providing reference to the sharing 3D GIS based on web service.

  8. AFAL: a web service for profiling amino acids surrounding ligands in proteins.

    PubMed

    Arenas-Salinas, Mauricio; Ortega-Salazar, Samuel; Gonzales-Nilo, Fernando; Pohl, Ehmke; Holmes, David S; Quatrini, Raquel

    2014-11-01

    With advancements in crystallographic technology and the increasing wealth of information populating structural databases, there is an increasing need for prediction tools based on spatial information that will support the characterization of proteins and protein-ligand interactions. Herein, a new web service is presented termed amino acid frequency around ligand (AFAL) for determining amino acids type and frequencies surrounding ligands within proteins deposited in the Protein Data Bank and for assessing the atoms and atom-ligand distances involved in each interaction (availability: http://structuralbio.utalca.cl/AFAL/index.html ). AFAL allows the user to define a wide variety of filtering criteria (protein family, source organism, resolution, sequence redundancy and distance) in order to uncover trends and evolutionary differences in amino acid preferences that define interactions with particular ligands. Results obtained from AFAL provide valuable statistical information about amino acids that may be responsible for establishing particular ligand-protein interactions. The analysis will enable investigators to compare ligand-binding sites of different proteins and to uncover general as well as specific interaction patterns from existing data. Such patterns can be used subsequently to predict ligand binding in proteins that currently have no structural information and to refine the interpretation of existing protein models. The application of AFAL is illustrated by the analysis of proteins interacting with adenosine-5'-triphosphate.

  9. AFAL: a web service for profiling amino acids surrounding ligands in proteins.

    PubMed

    Arenas-Salinas, Mauricio; Ortega-Salazar, Samuel; Gonzales-Nilo, Fernando; Pohl, Ehmke; Holmes, David S; Quatrini, Raquel

    2014-11-01

    With advancements in crystallographic technology and the increasing wealth of information populating structural databases, there is an increasing need for prediction tools based on spatial information that will support the characterization of proteins and protein-ligand interactions. Herein, a new web service is presented termed amino acid frequency around ligand (AFAL) for determining amino acids type and frequencies surrounding ligands within proteins deposited in the Protein Data Bank and for assessing the atoms and atom-ligand distances involved in each interaction (availability: http://structuralbio.utalca.cl/AFAL/index.html ). AFAL allows the user to define a wide variety of filtering criteria (protein family, source organism, resolution, sequence redundancy and distance) in order to uncover trends and evolutionary differences in amino acid preferences that define interactions with particular ligands. Results obtained from AFAL provide valuable statistical information about amino acids that may be responsible for establishing particular ligand-protein interactions. The analysis will enable investigators to compare ligand-binding sites of different proteins and to uncover general as well as specific interaction patterns from existing data. Such patterns can be used subsequently to predict ligand binding in proteins that currently have no structural information and to refine the interpretation of existing protein models. The application of AFAL is illustrated by the analysis of proteins interacting with adenosine-5'-triphosphate. PMID:25085083

  10. Video Conferencing and Knowledge Management in In-Service Teacher Distance Lifelong Training and Development

    NASA Astrophysics Data System (ADS)

    Kamakari, Anna; Drigas, Athanasios

    In Greece, the training system and education can be adapted to distance lifelong e-learning and the teachers, the society and the economy can be benefited via the introduction and training of in-service teachers in a) the Video Conferencing (VC) technology and mode of distant training, granted that VC is put into a learning framework, and b) Knowledge Management (KM) methodology, through lectures and seminars. Incidentally, VC technology and KM methodology can be integrated into in-service teacher distance lifelong training and development, and to this end, an example, a hybrid model, of the manner VC technology and KM methodology can be integrated into in-service teacher distance lifelong training and development, that is, a microteaching session via VC and KM, which is advisable for Greek teachers of all specialisations in Greece or abroad, is given.

  11. Enrollment in Distance Education Courses, by State: Fall 2012. Web Tables. NCES 2014-023

    ERIC Educational Resources Information Center

    National Center for Education Statistics, 2014

    2014-01-01

    Postsecondary enrollment in distance education courses, particularly those offered online, has rapidly increased in recent years (Allen and Seaman 2013). Traditionally, distance education offerings and enrollment levels have varied across different types of institutions. For example, researchers have found that undergraduate enrollment in at least…

  12. Study Procrastination, Achievement, and Academic Motivation in Web-Based and Blended Distance Learning

    ERIC Educational Resources Information Center

    Romano, Jeanine; Wallace, Tary L.; Helmick, Ina J.; Carey, Lou M.; Adkins, Lisa

    2005-01-01

    Growth in distance education is increasing the need to examine students' learning strategies in distance and blended learning environments. Students' cramming or spaced-review behaviors were measured and compared across delivery formats as well as examined related to course achievement and attitudes across a term. Although theory would predict…

  13. Measuring Transactional Distance in Web-Based Learning Environments: An Initial Instrument Development

    ERIC Educational Resources Information Center

    Huang, Xiaoxia; Chandra, Aruna; DePaolo, Concetta; Cribbs, Jennifer; Simmons, Lakisha

    2015-01-01

    This study was an initial attempt to operationalise Moore's transactional distance theory by developing and validating an instrument measuring the related constructs: dialogue, structure, learner autonomy and transactional distance. Data were collected from 227 online students and analysed through an exploratory factor analysis. Results suggest…

  14. The EarthServer Geology Service: web coverage services for geosciences

    NASA Astrophysics Data System (ADS)

    Laxton, John; Sen, Marcus; Passmore, James

    2014-05-01

    The EarthServer FP7 project is implementing web coverage services using the OGC WCS and WCPS standards for a range of earth science domains: cryospheric; atmospheric; oceanographic; planetary; and geological. BGS is providing the geological service (http://earthserver.bgs.ac.uk/). Geoscience has used remote sensed data from satellites and planes for some considerable time, but other areas of geosciences are less familiar with the use of coverage data. This is rapidly changing with the development of new sensor networks and the move from geological maps to geological spatial models. The BGS geology service is designed initially to address two coverage data use cases and three levels of data access restriction. Databases of remote sensed data are typically very large and commonly held offline, making it time-consuming for users to assess and then download data. The service is designed to allow the spatial selection, editing and display of Landsat and aerial photographic imagery, including band selection and contrast stretching. This enables users to rapidly view data, assess is usefulness for their purposes, and then enhance and download it if it is suitable. At present the service contains six band Landsat 7 (Blue, Green, Red, NIR 1, NIR 2, MIR) and three band false colour aerial photography (NIR, green, blue), totalling around 1Tb. Increasingly 3D spatial models are being produced in place of traditional geological maps. Models make explicit spatial information implicit on maps and thus are seen as a better way of delivering geosciences information to non-geoscientists. However web delivery of models, including the provision of suitable visualisation clients, has proved more challenging than delivering maps. The EarthServer geology service is delivering 35 surfaces as coverages, comprising the modelled superficial deposits of the Glasgow area. These can be viewed using a 3D web client developed in the EarthServer project by Fraunhofer. As well as remote sensed

  15. Market Research: The World Wide Web Meets the Online Services.

    ERIC Educational Resources Information Center

    Bing, Michelle

    1996-01-01

    The World Wide Web can provide direct market research data inexpensively or can target the appropriate professional online database and narrow the search. This article discusses the Web presence of research and investment firms, financial pages, trade associations, and electronic publications containing market research data. It lists Uniform…

  16. Enabling Web-Based Analysis of CUAHSI HIS Hydrologic Data Using R and Web Processing Services

    NASA Astrophysics Data System (ADS)

    Ames, D. P.; Kadlec, J.; Bayles, M.; Seul, M.; Hooper, R. P.; Cummings, B.

    2015-12-01

    The CUAHSI Hydrologic Information System (CUAHSI HIS) provides open access to a large number of hydrological time series observation and modeled data from many parts of the world. Several software tools have been designed to simplify searching and access to the CUAHSI HIS datasets. These software tools include: Desktop client software (HydroDesktop, HydroExcel), developer libraries (WaterML R Package, OWSLib, ulmo), and the new interactive search website, http://data.cuahsi.org. An issue with using the time series data from CUAHSI HIS for further analysis by hydrologists (for example for verification of hydrological and snowpack models) is the large heterogeneity of the time series data. The time series may be regular or irregular, contain missing data, have different time support, and be recorded in different units. R is a widely used computational environment for statistical analysis of time series and spatio-temporal data that can be used to assess fitness and perform scientific analyses on observation data. R includes the ability to record a data analysis in the form of a reusable script. The R script together with the input time series dataset can be shared with other users, making the analysis more reproducible. The major goal of this study is to examine the use of R as a Web Processing Service for transforming time series data from the CUAHSI HIS and sharing the results on the Internet within HydroShare. HydroShare is an online data repository and social network for sharing large hydrological data sets such as time series, raster datasets, and multi-dimensional data. It can be used as a permanent cloud storage space for saving the time series analysis results. We examine the issues associated with running R scripts online: including code validation, saving of outputs, reporting progress, and provenance management. An explicit goal is that the script which is run locally should produce exactly the same results as the script run on the Internet. Our design can

  17. A Base Solution for Exposing IMS Telecommunication Services to Web 2.0 Enabled Applications

    NASA Astrophysics Data System (ADS)

    Deinert, Florian; Murarasu, Alin; Bachmann, Andreas; Magedanz, Thomas

    The convergence of telecommunication and Web 2.0 services is leading to new opportunities for the telecommunications market. Companies are looking for ways to include their services in Web 2.0 applications. Predictions suggest that future telecommunication networks will be based on the IP Multimedia Subsystem (IMS), an all IP telecommunication core network. This paper describes an approach to combining Web 2.0 enabled applications, namely widgets, with telecommunication features using IMS. Widgets are small applications based on Web technologies that run on the client device. A new abstraction layer with interfaces for the different telecommunication features will be introduced. In addition a widget engine that makes these telecommunication interfaces available to its widgets will be presented. This will allow the rapid development of IMS applications for external developers and the combination of other Web 2.0 services with IMS features.

  18. Web Services and Handle Infrastructure - WDCC's Contributions to International Projects

    NASA Astrophysics Data System (ADS)

    Föll, G.; Weigelt, T.; Kindermann, S.; Lautenschlager, M.; Toussaint, F.

    2012-04-01

    Climate science demands on data management are growing rapidly as climate models grow in the precision with which they depict spatial structures and in the completeness with which they describe a vast range of physical processes. The ExArch project is exploring the challenges of developing a software management infrastructure which will scale to the multi-exabyte archives of climate data which are likely to be crucial to major policy decisions in by the end of the decade. The ExArch approach to future integration of exascale climate archives is based on one hand on a distributed web service architecture providing data analysis and quality control functionality across archvies. On the other hand a consistent persistent identifier infrastructure is deployed to support distributed data management and data replication. Distributed data analysis functionality is based on the CDO climate data operators' package. The CDO-Tool is used for processing of the archived data and metadata. CDO is a collection of command line Operators to manipulate and analyse Climate and forecast model Data. A range of formats is supported and over 500 operators are provided. CDO presently is designed to work in a scripting environment with local files. ExArch will extend the tool to support efficient usage in an exascale archive with distributed data and computational resources by providing flexible scheduling capabilities. Quality control will become increasingly important in an exascale computing context. Researchers will be dealing with millions of data files from multiple sources and will need to know whether the files satisfy a range of basic quality criterea. Hence ExArch will provide a flexible and extensible quality control system. The data will be held at more than 30 computing centres and data archives around the world, but for users it will appear as a single archive due to a standardized ExArch Web Processing Service. Data infrastructures such as the one built by ExArch can greatly

  19. 47 CFR 54.625 - Support for services beyond the maximum supported distance for rural health care providers.

    Code of Federal Regulations, 2012 CFR

    2012-10-01

    ... supported distance for rural health care providers. 54.625 Section 54.625 Telecommunication FEDERAL... Support for Health Care Providers § 54.625 Support for services beyond the maximum supported distance for rural health care providers. (a) The maximum support distance is the distance from the health...

  20. 47 CFR 54.625 - Support for services beyond the maximum supported distance for rural health care providers.

    Code of Federal Regulations, 2010 CFR

    2010-10-01

    ... supported distance for rural health care providers. 54.625 Section 54.625 Telecommunication FEDERAL... Support for Health Care Providers § 54.625 Support for services beyond the maximum supported distance for rural health care providers. (a) The maximum support distance is the distance from the health...

  1. 47 CFR 54.625 - Support for services beyond the maximum supported distance for rural health care providers.

    Code of Federal Regulations, 2011 CFR

    2011-10-01

    ... supported distance for rural health care providers. 54.625 Section 54.625 Telecommunication FEDERAL... Support for Health Care Providers § 54.625 Support for services beyond the maximum supported distance for rural health care providers. (a) The maximum support distance is the distance from the health...

  2. 78 FR 60303 - Agency Information Collection Activities: Online Survey of Web Services Employers; New...

    Federal Register 2010, 2011, 2012, 2013, 2014

    2013-10-01

    ... previously published in the Federal Register on July 16, 2013, at 78 FR 42537, allowing for a 60-day public... of Web Services Employers; New Information Collection ACTION: 30-Day Notice. SUMMARY: The Department... via the Federal eRulemaking Portal Web site at http://www.Regulations.gov under e-Docket ID...

  3. Using the Internet as a Tool for Public Service: Creating a Community History Web Site

    ERIC Educational Resources Information Center

    Henson, Darold Leigh

    2005-01-01

    Creating a community history Web site is a way for technical communication practitioners, students, and teachers to improve their expertise while performing a valuable public service. Developers of this kind of Web site combine personal interest in the history and culture of their chosen communities with professional interest in a wide range of…

  4. Motivating Pre-Service Teachers in Technology Integration of Web 2.0 for Teaching Internships

    ERIC Educational Resources Information Center

    Kim, Hye Jeong; Jang, Hwan Young

    2015-01-01

    The aim of this study was to examine the predictors of pre-service teachers' use of Web 2.0 tools during a teaching internship, after a course that emphasized the use of the tools for instructional activities. Results revealed that integrating Web 2.0 tools during their teaching internship was strongly predicted by participants' perceived…

  5. A Web-Based Model for Developing Assessment Literacy of Secondary In-Service Teachers

    ERIC Educational Resources Information Center

    Fan, Ya-Ching; Wang, Tzu-Hua; Wang, Kuo-Hua

    2011-01-01

    This research investigates the effect of a web-based model, named "Practicing, Reflecting, and Revising with Web-based Assessment and Test Analysis system (P2R-WATA) Assessment Literacy Development Model," on enhancing assessment knowledge and perspectives of secondary in-service teachers, and adopts a single group experimental research design.…

  6. Crossing the Threshold: Practical Foundations for Government Services on the World Wide Web.

    ERIC Educational Resources Information Center

    Dawes, Sharon S.; Pardo, Theresa A.; DiCaterino, Ann

    1999-01-01

    Two Internet testbeds conducted at the New York State Center for Technology in Government (CTG) identified five threshold factors that shape a government agency's ability to initiate Web-based services: technical infrastructure, user capabilities, information content, cost, and security. Six kinds of benefits of the Web to deliver government…

  7. Services for Graduate Students: A Review of Academic Library Web Sites

    ERIC Educational Resources Information Center

    Rempel, Hannah Gascho

    2010-01-01

    A library's Web site is well recognized as the gateway to the library for the vast majority of users. Choosing the most user-friendly Web architecture to reflect the many services libraries offer is a complex process, and librarians are still experimenting to find what works best for their users. As part of a redesign of the Oregon State…

  8. OneGeology Web Services and Portal as a global geological SDI - latest standards and technology

    NASA Astrophysics Data System (ADS)

    Duffy, Tim; Tellez-Arenas, Agnes

    2014-05-01

    The global coverage of OneGeology Web Services (www.onegeology.org and portal.onegeology.org) achieved since 2007 from the 120 participating geological surveys will be reviewed and issues arising discussed. Recent enhancements to the OneGeology Web Services capabilities will be covered including new up to 5 star service accreditation scheme utilising the ISO/OGC Web Mapping Service standard version 1.3, core ISO 19115 metadata additions and Version 2.0 Web Feature Services (WFS) serving the new IUGS-CGI GeoSciML V3.2 geological web data exchange language standard (http://www.geosciml.org/) with its associated 30+ IUGS-CGI available vocabularies (http://resource.geosciml.org/ and http://srvgeosciml.brgm.fr/eXist2010/brgm/client.html). Use of the CGI simpelithology and timescale dictionaries now allow those who wish to do so to offer data harmonisation to query their GeoSciML 3.2 based Web Feature Services and their GeoSciML_Portrayal V2.0.1 (http://www.geosciml.org/) Web Map Services in the OneGeology portal (http://portal.onegeology.org). Contributing to OneGeology involves offering to serve ideally 1:1000,000 scale geological data (in practice any scale now is warmly welcomed) as an OGC (Open Geospatial Consortium) standard based WMS (Web Mapping Service) service from an available WWW server. This may either be hosted within the Geological Survey or a neighbouring, regional or elsewhere institution that offers to serve that data for them i.e. offers to help technically by providing the web serving IT infrastructure as a 'buddy'. OneGeology is a standards focussed Spatial Data Infrastructure (SDI) and works to ensure that these standards work together and it is now possible for European Geological Surveys to register their INSPIRE web services within the OneGeology SDI (e.g. see http://www.geosciml.org/geosciml/3.2/documentation/cookbook/INSPIRE_GeoSciML_Cookbook%20_1.0.pdf). The Onegeology portal (http://portal.onegeology.org) is the first port of call for anyone

  9. A Survey of the Usability of Digital Reference Services on Academic Health Science Library Web Sites

    ERIC Educational Resources Information Center

    Dee, Cheryl; Allen, Maryellen

    2006-01-01

    Reference interactions with patrons in a digital library environment using digital reference services (DRS) has become widespread. However, such services in many libraries appear to be underutilized. A study surveying the ease and convenience of such services for patrons in over 100 academic health science library Web sites suggests that…

  10. QoS Measurement of Workflow-Based Web Service Compositions Using Colored Petri Net

    PubMed Central

    Nematzadeh, Hossein; Motameni, Homayun; Nematzadeh, Zahra

    2014-01-01

    Workflow-based web service compositions (WB-WSCs) is one of the main composition categories in service oriented architecture (SOA). Eflow, polymorphic process model (PPM), and business process execution language (BPEL) are the main techniques of the category of WB-WSCs. Due to maturity of web services, measuring the quality of composite web services being developed by different techniques becomes one of the most important challenges in today's web environments. Business should try to provide good quality regarding the customers' requirements to a composed web service. Thus, quality of service (QoS) which refers to nonfunctional parameters is important to be measured since the quality degree of a certain web service composition could be achieved. This paper tried to find a deterministic analytical method for dependability and performance measurement using Colored Petri net (CPN) with explicit routing constructs and application of theory of probability. A computer tool called WSET was also developed for modeling and supporting QoS measurement through simulation. PMID:25110748

  11. The value of Web-based library services at Cedars-Sinai Health System.

    PubMed Central

    Halub, L P

    1999-01-01

    Cedars-Sinai Medical Library/Information Center has maintained Web-based services since 1995 on the Cedars-Sinai Health System network. In that time, the librarians have found the provision of Web-based services to be a very worthwhile endeavor. Library users value the services that they access from their desktops because the services save time. They also appreciate being able to access services at their convenience, without restriction by the library's hours of operation. The library values its Web site because it brings increased visibility within the health system, and it enables library staff to expand services when budget restrictions have forced reduced hours of operation. In creating and maintaining the information center Web site, the librarians have learned the following lessons: consider the design carefully; offer what services you can, but weigh the advantages of providing the services against the time required to maintain them; make the content as accessible as possible; promote your Web site; and make friends in other departments, especially information services. PMID:10427423

  12. Using Allaire's ColdFusion to Deliver Web-Based Information to Distance Learning Students.

    ERIC Educational Resources Information Center

    Arashiro, Peter; Milton, Kirby

    1999-01-01

    Discusses Lansing Community College's (Michigan) Virtual College, initiated in 1997 and now offering 47 courses for online delivery. Reviews ColdFusion 3.1, a database-driven Web application server, and describes how the college has successfully used it to support students and faculty. (EMH)

  13. Web-Based Lecture Technologies: Blurring the Boundaries between Face-to-Face and Distance Learning

    ERIC Educational Resources Information Center

    Woo, Karen; Gosper, Maree; McNeill, Margot; Preston, Greg; Green, David; Phillips, Rob

    2008-01-01

    Web-based lecture technologies (WBLT) have gained popularity amongst universities in Australia as a tool for delivering lecture recordings to students in close to real time. This paper reports on a selection of results from a larger research project investigating the impact of WBLT on teaching and learning. Results show that while staff see the…

  14. Design Dimensions and Attributes for Web-Based Distance Learning Modules

    ERIC Educational Resources Information Center

    Pomales-Garcia, Cristina; Lopez, Angel D.; Liu, Yili

    2010-01-01

    This study investigates and compares the relative importance of Web module design dimensions and their attributes as perceived by student participants in a research study and those defined by a selected group of researchers in the literature. We aim to understand whether the dimensions of clarity, organization, structure, visual/aesthetical…

  15. Learner Characteristics, Behavior, and Achievement within Web-Based Distance Education: Learner-Centered Models.

    ERIC Educational Resources Information Center

    Reese, Debbie Denise

    Second year Instructional Technology masters (ITMA) students (N=49) in a 3-year Web-based Instructional Technology program at Virginia Tech chose to communicate with faculty, support staff, and other students via a virtual forum rather than privately if the students had poor coping skills. Specifically, these students posted more messages to the…

  16. Design, Development and Testing of Web Services for Multi-Sensor Snow Cover Mapping

    NASA Astrophysics Data System (ADS)

    Kadlec, Jiri

    This dissertation presents the design, development and validation of new data integration methods for mapping the extent of snow cover based on open access ground station measurements, remote sensing images, volunteer observer snow reports, and cross country ski track recordings from location-enabled mobile devices. The first step of the data integration procedure includes data discovery, data retrieval, and data quality control of snow observations at ground stations. The WaterML R package developed in this work enables hydrologists to retrieve and analyze data from multiple organizations that are listed in the Consortium of Universities for the Advancement of Hydrologic Sciences Inc (CUAHSI) Water Data Center catalog directly within the R statistical software environment. Using the WaterML R package is demonstrated by running an energy balance snowpack model in R with data inputs from CUAHSI, and by automating uploads of real time sensor observations to CUAHSI HydroServer. The second step of the procedure requires efficient access to multi-temporal remote sensing snow images. The Snow Inspector web application developed in this research enables the users to retrieve a time series of fractional snow cover from the Moderate Resolution Imaging Spectroradiometer (MODIS) for any point on Earth. The time series retrieval method is based on automated data extraction from tile images provided by a Web Map Tile Service (WMTS). The average required time for retrieving 100 days of data using this technique is 5.4 seconds, which is significantly faster than other methods that require the download of large satellite image files. The presented data extraction technique and space-time visualization user interface can be used as a model for working with other multi-temporal hydrologic or climate data WMTS services. The third, final step of the data integration procedure is generating continuous daily snow cover maps. A custom inverse distance weighting method has been developed

  17. Using JavaScript and the FDSN web service to create an interactive earthquake information system

    NASA Astrophysics Data System (ADS)

    Fischer, Kasper D.

    2015-04-01

    The FDSN web service provides a web interface to access earthquake meta-data (e. g. event or station information) and waveform date over the internet. Requests are send to a server as URLs and the output is either XML or miniSEED. This makes it hard to read by humans but easy to process with different software. Different data centers are already supporting the FDSN web service, e. g. USGS, IRIS, ORFEUS. The FDSN web service is also part of the Seiscomp3 (http://www.seiscomp3.org) software. The Seismological Observatory of the Ruhr-University switched to Seiscomp3 as the standard software for the analysis of mining induced earthquakes at the beginning of 2014. This made it necessary to create a new web-based earthquake information service for the publication of results to the general public. This has be done by processing the output of a FDSN web service query by javascript running in a standard browser. The result is an interactive map presenting the observed events and further information of events and stations on a single web page as a table and on a map. In addition the user can download event information, waveform data and station data in different formats like miniSEED, quakeML or FDSNxml. The developed code and all used libraries are open source and freely available.

  18. 47 CFR 54.625 - Support for telecommunications services beyond the maximum supported distance for rural health...

    Code of Federal Regulations, 2014 CFR

    2014-10-01

    ... the maximum supported distance for rural health care providers. 54.625 Section 54.625... SERVICE Universal Service Support for Health Care Providers Telecommunications Program § 54.625 Support for telecommunications services beyond the maximum supported distance for rural health care...

  19. 47 CFR 54.625 - Support for telecommunications services beyond the maximum supported distance for rural health...

    Code of Federal Regulations, 2013 CFR

    2013-10-01

    ... the maximum supported distance for rural health care providers. 54.625 Section 54.625... SERVICE Universal Service Support for Health Care Providers Telecommunications Program § 54.625 Support for telecommunications services beyond the maximum supported distance for rural health care...

  20. Adopting and adapting a commercial view of web services for the Navy

    NASA Astrophysics Data System (ADS)

    Warner, Elizabeth; Ladner, Roy; Katikaneni, Uday; Petry, Fred

    2005-05-01

    Web Services are being adopted as the enabling technology to provide net-centric capabilities for many Department of Defense operations. The Navy Enterprise Portal, for example, is Web Services-based, and the Department of the Navy is promulgating guidance for developing Web Services. Web Services, however, only constitute a baseline specification that provides the foundation on which users, under current approaches, write specialized applications in order to retrieve data over the Internet. Application development may increase dramatically as the number of different available Web Services increases. Reasons for specialized application development include XML schema versioning differences, adoption/use of diverse business rules, security access issues, and time/parameter naming constraints, among others. We are currently developing for the US Navy a system which will improve delivery of timely and relevant meteorological and oceanographic (MetOc) data to the warfighter. Our objective is to develop an Advanced MetOc Broker (AMB) that leverages Web Services technology to identify, retrieve and integrate relevant MetOc data in an automated manner. The AMB will utilize a Mediator, which will be developed by applying ontological research and schema matching techniques to MetOc forms of data. The AMB, using the Mediator, will support a new, advanced approach to the use of Web Services; namely, the automated identification, retrieval and integration of MetOc data. Systems based on this approach will then not require extensive end-user application development for each Web Service from which data can be retrieved. Users anywhere on the globe will be able to receive timely environmental data that fits their particular needs.

  1. Building from Where We Are: Web Services and Java-Based Clients to Enable Virtual Observatories

    NASA Astrophysics Data System (ADS)

    Candey, R. M.; Chimiak, R. A.; Han, D. B.; Harris, B. T.; Johnson, R. C.; Klipsch, C. A.; Kovalick, T. J.; Leckner, H. A.; Liu, M. H.; McGuire, R. E.

    2005-12-01

    The Space Physics Data Facility at NASA Goddard has developed a strong foundation in space science mission services and data for enhancing the scientific return of space physics research and enabling integration of these services into the emerging Virtual discipline Observatory (VxO) paradigm. We are deploying a critical set of foundation components, leveraging our data format expertise and our existing and very popular science and orbit data web-based services, such as Coordinated Data Analysis Web [CDAWeb] and Satellite Situation Center Web [SSCweb]. We have developed web services APIs for orbit location, data finding across FTP sites and in CDAWeb, data file format translation, and data visualizations that tie together existing data holdings, standardize and simplify their use, and enable much enhanced interoperability and data analysis. We describe the technologies we've developed, our experiences and lessons-learned in implementing them, why we chose some technologies over others (web services vs. CORBA or simple CGI, Java vs. JavaScript or Flash), what we might do differently now, and our future direction. We discuss the difficulties in maintaining compatibility and inter-operability through various versions of web services and in merging various client projects, while adding extended functionality such as sonification interfaces in a modular fashion.

  2. Mobile Device Intervention for Student Support Services in Distance Education Context--FRAME Model Perspective

    ERIC Educational Resources Information Center

    Kumar, Lalita S.; Jamatia, Biplab; Aggarwal, A. K.; Kannan, S.

    2011-01-01

    This paper reports the findings of a study conducted to analyse the effect of mobile device intervention for student support services and to gauge its use for enhancing teaching--learning process as a future study in the context of offer of Distance Education programmes. The study was conducted with the learners of the coveted Post Graduate…

  3. Libraries' Services at Distance: A Survey of Allama Iqbal Open University Tutors in Pakistan

    ERIC Educational Resources Information Center

    Arif, Muhammad; Mahmood, Khalid

    2009-01-01

    A survey was conducted to determine the satisfaction level of distance education tutors with the location and physical setup, collection, resources and services being offered at thirty-four regional campuses and centers' libraries network of Allama Iqbal Open University (AIOU) in Pakistan including Azad Jummu and Kashmir territory. A semi…

  4. Gazing into the Crystal Ball: Using Scenarios for Future Visioning of a Distance Learning Library Service

    ERIC Educational Resources Information Center

    Casey, Anne Marie; Cawthorne, Jon E.; Citro, Kathleen

    2014-01-01

    This article describes the use of scenarios as a tool to assist a large distance learning library service in its strategic planning. Through a description of the scenario process from beginning to end, the authors detail the steps that the library director and the consultant took initially; their missteps; and the successful conclusion. This study…

  5. A Meta-Ethnographic Synthesis of Support Services in Distance Learning Programs

    ERIC Educational Resources Information Center

    Tuquero, Jean Marie

    2011-01-01

    This qualitative study utilized a meta-ethnographic approach to synthesize findings of five dissertations that focused on distance learning support services. The history of ethnographical studies stemmed from an anthropological background. Researchers have applied ethnographical approaches to examine and document various phenomena. This…

  6. A Meta-Ethnographic Synthesis of Support Services for Adult Learners in Distance Learning Programs

    ERIC Educational Resources Information Center

    Tuquero, Jean M.

    2010-01-01

    This qualitative research study utilized Noblit and Hare's (1988) meta-ethnographic approach to synthesize findings of five dissertations that focused on distance learning support services for adult learners. Noblit and Hare's (1988) meta-ethnographic approach consists of seven phases. Each meta-ethnographic phase guided the identification process…

  7. Experimenting with semantic web services to understand the role of NLP technologies in healthcare.

    PubMed

    Jagannathan, V

    2006-01-01

    NLP technologies can play a significant role in healthcare where a predominant segment of the clinical documentation is in text form. In a graduate course focused on understanding semantic web services at West Virginia University, a class project was designed with the purpose of exploring potential use for NLP-based abstraction of clinical documentation. The role of NLP-technology was simulated using human abstractors and various workflows were investigated using public domain workflow and semantic web service technologies. This poster explores the potential use of NLP and the role of workflow and semantic web technologies in developing healthcare IT environments.

  8. A Smart Modeling Framework for Integrating BMI-enabled Models as Web Services

    NASA Astrophysics Data System (ADS)

    Jiang, P.; Elag, M.; Kumar, P.; Peckham, S. D.; Liu, R.; Marini, L.; Hsu, L.

    2015-12-01

    Serviced-oriented computing provides an opportunity to couple web service models using semantic web technology. Through this approach, models that are exposed as web services can be conserved in their own local environment, thus making it easy for modelers to maintain and update the models. In integrated modeling, the serviced-oriented loose-coupling approach requires (1) a set of models as web services, (2) the model metadata describing the external features of a model (e.g., variable name, unit, computational grid, etc.) and (3) a model integration framework. We present the architecture of coupling web service models that are self-describing by utilizing a smart modeling framework. We expose models that are encapsulated with CSDMS (Community Surface Dynamics Modeling System) Basic Model Interfaces (BMI) as web services. The BMI-enabled models are self-describing by uncovering models' metadata through BMI functions. After a BMI-enabled model is serviced, a client can initialize, execute and retrieve the meta-information of the model by calling its BMI functions over the web. Furthermore, a revised version of EMELI (Peckham, 2015), an Experimental Modeling Environment for Linking and Interoperability, is chosen as the framework for coupling BMI-enabled web service models. EMELI allows users to combine a set of component models into a complex model by standardizing model interface using BMI as well as providing a set of utilities smoothing the integration process (e.g., temporal interpolation). We modify the original EMELI so that the revised modeling framework is able to initialize, execute and find the dependencies of the BMI-enabled web service models. By using the revised EMELI, an example will be presented on integrating a set of topoflow model components that are BMI-enabled and exposed as web services. Reference: Peckham, S.D. (2014) EMELI 1.0: An experimental smart modeling framework for automatic coupling of self-describing models, Proceedings of HIC 2014

  9. 76 FR 28439 - Submission for OMB Review; Comment Request; NCI Cancer Genetics Services Directory Web-Based...

    Federal Register 2010, 2011, 2012, 2013, 2014

    2011-05-17

    ... Genetics Services Directory Web-Based Application Form and Update Mailer Summary: Under the provisions of... Collection: Title: NCI Cancer Genetics Services Directory Web-based Application Form and Update Mailer. Type... Information Collection: The purpose of the online application form and the Web-based update mailer is...

  10. A study on heterogeneous distributed spatial information platform based on semantic Web services

    NASA Astrophysics Data System (ADS)

    Peng, Shuang-yun; Yang, Kun; Xu, Quan-li; Huang, Bang-mei

    2008-10-01

    With the development of Semantic Web technology, the spatial information service based on ontology is an effective way for sharing and interoperation of heterogeneous information resources in the distributed network environment. This paper discusses spatial information sharing and interoperability in the Semantic Web Services architecture. Through using Ontology record spatial information in sharing knowledge system, explicit and formalization expresses the default and the concealment semantic information. It provides the prerequisite for spatial information sharing and interoperability; Through Semantic Web Services technology parses Ontology and intelligent buildings services under network environment, form a network of services. In order to realize the practical applications of spatial information sharing and interoperation in different brunches of CDC system, a prototype system for HIV/AIDS information sharing based on geo-ontology has also been developed by using the methods described above.

  11. Temporal logic and operation relations based knowledge representation for land cover change web services

    NASA Astrophysics Data System (ADS)

    Chen, Jun; Wu, Hao; Li, Songnian; Liao, Anping; He, Chaoying; Peng, Shu

    2013-09-01

    Providing land cover spatio-temporal information and geo-computing through web service is a new challenge for supporting global change research, earth system simulation and many other societal benefit areas. This requires an integrated knowledge representation and web implementation of static land cover and change information, as well as the related operations for geo-computing. The temporal logic relations among land cover snapshots and increments were examined with a matrix-based three-step analysis. Twelve temporal logic relations were identified and five basic spatial operations were formalized with set operators, which were all used to develop algorithms for deriving implicit change information. A knowledge representation for land cover change information was then developed based on these temporal logic and operation relations. A prototype web-service system was further implemented based on OWL-DL. Both online access and conversion of land cover spatio-temporal information can be facilitated with such a web service system.

  12. A Web-Based Model for Online Collaboration between Distance Learning and Campus Students.

    ERIC Educational Resources Information Center

    Mouza, Chrystalla; Kaplan, Danielle; Espinet, Ivana

    This paper presents a hybrid course framework at the Teachers College at Columbia University (New York) that seamlessly integrates a traditional course plan designed for on-campus students with an online course plan designed for distance learning students. Based on innovative teaching and learning principles, the course encourages active…

  13. Real Time with the Librarian: Using Web Conferencing Software to Connect to Distance Students

    ERIC Educational Resources Information Center

    Riedel, Tom; Betty, Paul

    2013-01-01

    A pilot program to provide real-time library webcasts to Regis University distance students using Adobe Connect software was initiated in fall of 2011. Previously, most interaction between librarians and online students had been accomplished by asynchronous discussion threads in the Learning Management System. Library webcasts were offered in…

  14. The Evolution of Distance Education: Implications for Instructional Design on the Potential of the Web

    ERIC Educational Resources Information Center

    Moller, Leslie; Foshay, Wellesley R.; Huett, Jason

    2008-01-01

    In this article, the authors offer the first of a three-part series on distance education. The authors discuss key trends in training and development which have five profound impacts on the field of Instruction Design (ID). These five effects concern: (1) quality; (2) needs assessment, ROI and measurement of outcomes; (3) the influence and fusion…

  15. A RESTful Web service interface to the ATLAS COOL database

    NASA Astrophysics Data System (ADS)

    Roe, S. A.; ATLAS Collaboration

    2010-04-01

    The COOL database in ATLAS is primarily used for storing detector conditions data, but also status flags which are uploaded summaries of information to indicate the detector reliability during a run. This paper introduces the use of CherryPy, a Python application server which acts as an intermediate layer between a web interface and the database, providing a simple means of storing to and retrieving from the COOL database which has found use in many web applications. The software layer is designed to be RESTful, implementing the common CRUD (Create, Read, Update, Delete) database methods by means of interpreting the HTTP method (POST, GET, PUT, DELETE) on the server along with a URL identifying the database resource to be operated on. The format of the data (text, xml etc) is also determined by the HTTP protocol. The details of this layer are described along with a popular application demonstrating its use, the ATLAS run list web page.

  16. AdaFF: Adaptive Failure-Handling Framework for Composite Web Services

    NASA Astrophysics Data System (ADS)

    Kim, Yuna; Lee, Wan Yeon; Kim, Kyong Hoon; Kim, Jong

    In this paper, we propose a novel Web service composition framework which dynamically accommodates various failure recovery requirements. In the proposed framework called Adaptive Failure-handling Framework (AdaFF), failure-handling submodules are prepared during the design of a composite service, and some of them are systematically selected and automatically combined with the composite Web service at service instantiation in accordance with the requirement of individual users. In contrast, existing frameworks cannot adapt the failure-handling behaviors to user's requirements. AdaFF rapidly delivers a composite service supporting the requirement-matched failure handling without manual development, and contributes to a flexible composite Web service design in that service architects never care about failure handling or variable requirements of users. For proof of concept, we implement a prototype system of the AdaFF, which automatically generates a composite service instance with Web Services Business Process Execution Language (WS-BPEL) according to the users' requirement specified in XML format and executes the generated instance on the ActiveBPEL engine.

  17. Web Services Implementations at Land Process and Goddard Earth Sciences Distributed Active Archive Centers

    NASA Astrophysics Data System (ADS)

    Cole, M.; Bambacus, M.; Lynnes, C.; Sauer, B.; Falke, S.; Yang, W.

    2007-12-01

    NASA's vast array of scientific data within its Distributed Active Archive Centers (DAACs) is especially valuable to both traditional research scientists as well as the emerging market of Earth Science Information Partners. For example, the air quality science and management communities are increasingly using satellite derived observations in their analyses and decision making. The Air Quality Cluster in the Federation of Earth Science Information Partners (ESIP) uses web infrastructures of interoperability, or Service Oriented Architecture (SOA), to extend data exploration, use, and analysis and provides a user environment for DAAC products. In an effort to continually offer these NASA data to the broadest research community audience, and reusing emerging technologies, both NASA's Goddard Earth Science (GES) and Land Process (LP) DAACs have engaged in a web services pilot project. Through these projects both GES and LP have exposed data through the Open Geospatial Consortiums (OGC) Web Services standards. Reusing several different existing applications and implementation techniques, GES and LP successfully exposed a variety data, through distributed systems to be ingested into multiple end-user systems. The results of this project will enable researchers world wide to access some of NASA's GES & LP DAAC data through OGC protocols. This functionality encourages inter-disciplinary research while increasing data use through advanced technologies. This paper will concentrate on the implementation and use of OGC Web Services, specifically Web Map and Web Coverage Services (WMS, WCS) at GES and LP DAACs, and the value of these services within scientific applications, including integration with the DataFed air quality web infrastructure and in the development of data analysis web applications.

  18. Investigating metrics of geospatial web services: The case of a CEOS federated catalog service for earth observation data

    NASA Astrophysics Data System (ADS)

    Han, Weiguo; Di, Liping; Yu, Genong; Shao, Yuanzheng; Kang, Lingjun

    2016-07-01

    Geospatial Web Services (GWS) make geospatial information and computing resources discoverable and accessible over the Web. Among them, Open Geospatial Consortium (OGC) standards-compliant data, catalog and processing services are most popular, and have been widely adopted and leveraged in geospatial research and applications. The GWS metrics, such as visit count, average processing time, and user distribution, are important to evaluate their overall performance and impacts. However, these metrics, especially of federated catalog service, have not been systematically evaluated and reported to relevant stakeholders from the point of view of service providers. Taking an integrated catalog service for earth observation data as an example, this paper describes metrics information retrieval, organization, and representation of a catalog service federation. An extensible and efficient log file analyzer is implemented to retrieve a variety of service metrics from the log file and store analysis results in an easily programmable format. An Ajax powered Web portal is built to provide stakeholders, sponsors, developers, partners, and other types of users with specific and relevant insights into metrics information in an interactive and informative form. The deployed system has provided useful information for periodical reports, service delivery, and decision support. The proposed measurement strategy and analytics framework can be a guidance to help GWS providers evaluate their services.

  19. Weaving Silos--A Leadership Challenge: A Cross-Functional Team Approach to Supporting Web-Based Student Services

    ERIC Educational Resources Information Center

    Kleemann, Gary L.

    2005-01-01

    The author reviews the evolution of Web services--from information sharing to transactional to relationship building--and the progression from first-generation to fourth-generation Web sites. (Contains 3 figures.)

  20. An Automated End-To Multi-Agent Qos Based Architecture for Selection of Geospatial Web Services

    NASA Astrophysics Data System (ADS)

    Shah, M.; Verma, Y.; Nandakumar, R.

    2012-07-01

    Over the past decade, Service-Oriented Architecture (SOA) and Web services have gained wide popularity and acceptance from researchers and industries all over the world. SOA makes it easy to build business applications with common services, and it provides like: reduced integration expense, better asset reuse, higher business agility, and reduction of business risk. Building of framework for acquiring useful geospatial information for potential users is a crucial problem faced by the GIS domain. Geospatial Web services solve this problem. With the help of web service technology, geospatial web services can provide useful geospatial information to potential users in a better way than traditional geographic information system (GIS). A geospatial Web service is a modular application designed to enable the discovery, access, and chaining of geospatial information and services across the web that are often both computation and data-intensive that involve diverse sources of data and complex processing functions. With the proliferation of web services published over the internet, multiple web services may provide similar functionality, but with different non-functional properties. Thus, Quality of Service (QoS) offers a metric to differentiate the services and their service providers. In a quality-driven selection of web services, it is important to consider non-functional properties of the web service so as to satisfy the constraints or requirements of the end users. The main intent of this paper is to build an automated end-to-end multi-agent based solution to provide the best-fit web service to service requester based on QoS.

  1. Leveraging Web Technologies in Student Support Self-Services

    ERIC Educational Resources Information Center

    Herndon, M. Craig

    2011-01-01

    The use of Web technologies to connect with and disperse information to prospective and current students can be effective for students as well as efficient for colleges. Early results of the use of such technologies in a statewide system point to high rates of satisfaction among students when information is delivered, provide clues on how various…

  2. IAServ: an intelligent home care web services platform in a cloud for aging-in-place.

    PubMed

    Su, Chuan-Jun; Chiang, Chang-Yu

    2013-11-01

    As the elderly population has been rapidly expanding and the core tax-paying population has been shrinking, the need for adequate elderly health and housing services continues to grow while the resources to provide such services are becoming increasingly scarce. Thus, increasing the efficiency of the delivery of healthcare services through the use of modern technology is a pressing issue. The seamless integration of such enabling technologies as ontology, intelligent agents, web services, and cloud computing is transforming healthcare from hospital-based treatments to home-based self-care and preventive care. A ubiquitous healthcare platform based on this technological integration, which synergizes service providers with patients' needs to be developed to provide personalized healthcare services at the right time, in the right place, and the right manner. This paper presents the development and overall architecture of IAServ (the Intelligent Aging-in-place Home care Web Services Platform) to provide personalized healthcare service ubiquitously in a cloud computing setting to support the most desirable and cost-efficient method of care for the aged-aging in place. The IAServ is expected to offer intelligent, pervasive, accurate and contextually-aware personal care services. Architecturally the implemented IAServ leverages web services and cloud computing to provide economic, scalable, and robust healthcare services over the Internet. PMID:24225647

  3. Pilot Evaluation of a Web-Based Intervention Targeting Sexual Health Service Access

    ERIC Educational Resources Information Center

    Brown, K. E.; Newby, K.; Caley, M.; Danahay, A.; Kehal, I.

    2016-01-01

    Sexual health service access is fundamental to good sexual health, yet interventions designed to address this have rarely been implemented or evaluated. In this article, pilot evaluation findings for a targeted public health behavior change intervention, delivered via a website and web-app, aiming to increase uptake of sexual health services among…

  4. Building to Scale: An Analysis of Web-Based Services in CIC (Big Ten) Libraries.

    ERIC Educational Resources Information Center

    Dewey, Barbara I.

    Advancing library services in large universities requires creative approaches for "building to scale." This is the case for CIC, Committee on Institutional Cooperation (Big Ten), libraries whose home institutions serve thousands of students, faculty, staff, and others. Developing virtual Web-based services is an increasingly viable avenue to bring…

  5. Use and Evaluation of Web-Based Professional Development Services across Participant Levels of Support

    ERIC Educational Resources Information Center

    Whitaker, Steve; Kinzie, Mable; Kraft-Sayre, Marcia E.; Mashburn, Andrew; Pianta, Robert C.

    2007-01-01

    When participating in a large-scale, web-based professional development program, to what degree do teachers participate? How useful do they find the program? To what degree do they feel supported in their efforts? What are the associations between participation, evaluation of services, and the level of service teachers receive? MyTeachingPartner…

  6. Pre-Service Teacher Reflections, Video-Conference and WebCT: An Exploratory Case Study

    ERIC Educational Resources Information Center

    Melville, Wayne; Bowen, G. Michael; Passmore, Graham

    2011-01-01

    Introduction: The development of video-conference and WebCT technology offers new possibilities for the education of pre-service teachers; opportunities that are only just beginning to be touched on. In this exploratory article, we investigate the opportunities for reflection that technology afforded three pre-service teachers in Canada as they a…

  7. Final Report for DOE Project: Portal Web Services: Support of DOE SciDAC Collaboratories

    SciTech Connect

    Mary Thomas, PI; Geoffrey Fox, Co-PI; Gannon, D; Pierce, M; Moore, R; Schissel, D; Boisseau, J

    2007-10-01

    Grid portals provide the scientific community with familiar and simplified interfaces to the Grid and Grid services, and it is important to deploy grid portals onto the SciDAC grids and collaboratories. The goal of this project is the research, development and deployment of interoperable portal and web services that can be used on SciDAC National Collaboratory grids. This project has four primary task areas: development of portal systems; management of data collections; DOE science application integration; and development of web and grid services in support of the above activities.

  8. The Activity Chain Safety and Liveness Specification of Composite Web Services

    NASA Astrophysics Data System (ADS)

    Chen, Bo; Huang, Xiaomei

    Web service composition is most impressing method for development and deployment of e-business. Description and modeling the behavior requirements of composite Web services for users and verifying composite Web service compliance to specific requirements is an important key in design of services. But most work does not address the issue of how to model the requirements that the BPEL4WS processes are supposed to satisfy. The specifications in verification works are general temporal relation based on activity or scenario in essence. Distinguish with these work, we propose a novel concept of behavior specification based on activity chain in which granularity is between activity and scenario. Chain existence mode, chain absence mode are designed to express such behavioral requirements based on activity chain that is similar with safety or liveness specification based on activity respectively. Encode them on Labeled Transition System LTS and then give them exact operation semantics. Finally, an example is illustrated.

  9. Persistent identifiers for web service requests relying on a provenance ontology design pattern

    NASA Astrophysics Data System (ADS)

    Car, Nicholas; Wang, Jingbo; Wyborn, Lesley; Si, Wei

    2016-04-01

    Delivering provenance information for datasets produced from static inputs is relatively straightforward: we represent the processing actions and data flow using provenance ontologies and link to stored copies of the inputs stored in repositories. If appropriate detail is given, the provenance information can then describe what actions have occurred (transparency) and enable reproducibility. When web service-generated data is used by a process to create a dataset instead of a static inputs, we need to use sophisticated provenance representations of the web service request as we can no longer just link to data stored in a repository. A graph-based provenance representation, such as the W3C's PROV standard, can be used to model the web service request as a single conceptual dataset and also as a small workflow with a number of components within the same provenance report. This dual representation does more than just allow simplified or detailed views of a dataset's production to be used where appropriate. It also allow persistent identifiers to be assigned to instances of a web service requests, thus enabling one form of dynamic data citation, and for those identifiers to resolve to whatever level of detail implementers think appropriate in order for that web service request to be reproduced. In this presentation we detail our reasoning in representing web service requests as small workflows. In outline, this stems from the idea that web service requests are perdurant things and in order to most easily persist knowledge of them for provenance, we should represent them as a nexus of relationships between endurant things, such as datasets and knowledge of particular system types, as these endurant things are far easier to persist. We also describe the ontology design pattern that we use to represent workflows in general and how we apply it to different types of web service requests. We give examples of specific web service requests instances that were made by systems

  10. AMBIT RESTful web services: an implementation of the OpenTox application programming interface

    PubMed Central

    2011-01-01

    The AMBIT web services package is one of the several existing independent implementations of the OpenTox Application Programming Interface and is built according to the principles of the Representational State Transfer (REST) architecture. The Open Source Predictive Toxicology Framework, developed by the partners in the EC FP7 OpenTox project, aims at providing a unified access to toxicity data and predictive models, as well as validation procedures. This is achieved by i) an information model, based on a common OWL-DL ontology ii) links to related ontologies; iii) data and algorithms, available through a standardized REST web services interface, where every compound, data set or predictive method has a unique web address, used to retrieve its Resource Description Framework (RDF) representation, or initiate the associated calculations. The AMBIT web services package has been developed as an extension of AMBIT modules, adding the ability to create (Quantitative) Structure-Activity Relationship (QSAR) models and providing an OpenTox API compliant interface. The representation of data and processing resources in W3C Resource Description Framework facilitates integrating the resources as Linked Data. By uploading datasets with chemical structures and arbitrary set of properties, they become automatically available online in several formats. The services provide unified interfaces to several descriptor calculation, machine learning and similarity searching algorithms, as well as to applicability domain and toxicity prediction models. All Toxtree modules for predicting the toxicological hazard of chemical compounds are also integrated within this package. The complexity and diversity of the processing is reduced to the simple paradigm "read data from a web address, perform processing, write to a web address". The online service allows to easily run predictions, without installing any software, as well to share online datasets and models. The downloadable web application

  11. Exploring Distance Learning Experiences of In-Service Music Teachers from Puerto Rico in a Master's Program

    ERIC Educational Resources Information Center

    Vega-Martinez, Juan Carlos

    2013-01-01

    The purpose of this study was to explore the experiences of in-service music teachers who chose to pursue a master's degree in music education through distance learning. In this study, I examined the motivations of in-service music teachers for choosing to pursue a master's degree in music education through distance learning; the benefits teachers…

  12. Satellite-Based Distance Courses for In-Service Training: The Case of HeadsUp! Reading

    ERIC Educational Resources Information Center

    Morrison, Johnetta Wade; Raya-Carlton, Pamela; Henk, Jennifer K.; Thornburg, Kathy R.

    2007-01-01

    This article discusses the use of distance courses as an in-service training mechanism for early childhood personnel. The authors evaluated the efficacy of the in-service, satellite based distance course HeadsUp! Reading (HU!R). The analysis of HU!R data revealed that there were no initial differences in the Language and Literacy Early Childhood:…

  13. Sensor Webs with a Service-Oriented Architecture for On-demand Science Products

    NASA Technical Reports Server (NTRS)

    Mandl, Daniel; Ungar, Stephen; Ames, Troy; Justice, Chris; Frye, Stuart; Chien, Steve; Tran, Daniel; Cappelaere, Patrice; Derezinsfi, Linda; Paules, Granville; Di, Liping; Kolitz, Stephan

    2007-01-01

    This paper describes the work being managed by the NASA Goddard Space Flight Center (GSFC) Information System Division (ISD) under a NASA Earth Science Technology Ofice (ESTO) Advanced Information System Technology (AIST) grant to develop a modular sensor web architecture which enables discovery of sensors and workflows that can create customized science via a high-level service-oriented architecture based on Open Geospatial Consortium (OGC) Sensor Web Enablement (SWE) web service standards. These capabilities serve as a prototype to a user-centric architecture for Global Earth Observing System of Systems (GEOSS). This work builds and extends previous sensor web efforts conducted at NASA/GSFC using the Earth Observing 1 (EO-1) satellite and other low-earth orbiting satellites.

  14. DataFed: Web Services-Based Mediator of Distributed Data Flow and Processing

    NASA Astrophysics Data System (ADS)

    Husar, R. B.; Falke, S. R.; Hojarvi, K.

    2005-12-01

    DataFed is a distributed web-services-based computing environment for accessing, processing and rendering environmental data in support of air quality management and science. The flexible, adoptive environment facilitates the access and flow of atmospheric data from provider to users by enabling the creation of user-driven data processing value chains. The approach of DataFed is mediation between users and data providers. DataFed assumes the autonomous and spontaneous emergence of distributed data sources. It non-intrusively wraps datasets for access by web services. The mediator software, composed of web services, provides homogeneous data views (e.g. geospatial, time views) using a global multi-dimensional data model. Application software written using web services consists of data browsers and analysis tools for distributed AQ data. Currently DataFed serves as data gateway for user programs; web pages, GIS, science tools as well as for loosely coupled Service Oriented Architecture Applications. Its federated data pool consists of over 50 datasets and the tools have been applied in several air pollution projects.

  15. Technical note: Harmonizing met-ocean model data via standard web services within small research groups

    NASA Astrophysics Data System (ADS)

    Signell, R. P.; Camossi, E.

    2015-11-01

    Work over the last decade has resulted in standardized web-services and tools that can significantly improve the efficiency and effectiveness of working with meteorological and ocean model data. While many operational modelling centres have enabled query and access to data via common web services, most small research groups have not. The penetration of this approach into the research community, where IT resources are limited, can be dramatically improved by: (1) making it simple for providers to enable web service access to existing output files; (2) using technology that is free, and that is easy to deploy and configure; and (3) providing tools to communicate with web services that work in existing research environments. We present a simple, local brokering approach that lets modelers continue producing custom data, but virtually aggregates and standardizes the data using NetCDF Markup Language. The THREDDS Data Server is used for data delivery, pycsw for data search, NCTOOLBOX (Matlab®1) and Iris (Python) for data access, and Ocean Geospatial Consortium Web Map Service for data preview. We illustrate the effectiveness of this approach with two use cases involving small research modelling groups at NATO and USGS.1 Mention of trade names or commercial products does not constitute endorsement or recommendation for use by the US Government.

  16. Chaining of web services and its application in geographic information integration and visualization

    NASA Astrophysics Data System (ADS)

    Chen, Chuanbin; Wu, Qunyong; Chen, Chongcheng; Chen, Han

    2005-11-01

    The rapid development of web services technology ushers Geographical Information Systems (GISs) into the era of geographic information web services (GIWS), which requires a scalable and extensible GISs model to deliver distributed geographic information and GISs functions integrated as independently-provided, interoperable services in a distributed computing environment. Several distributed services can be dynamically chained as a new service to accomplish a specific task. Such a model of service chaining is one of the most important research topics of next generation GISs. The paper highlights the issues of service chaining, the process of combining several distributed, interoperable GIWS dynamically to construct customized applications, and analyses characters of each pattern. Then, based on pattern of client-coordinated chaining, we design a service chaining which is developed in a J2EE development environment using web services technology, and construct a web services-oriented Geo-spatial data integration and visualization platform in order to integrate multi-sources and heterogeneous Geo-spatial data using Geography Markup Language (GML) and Geo-spatial data integration technology, and to visualize geographic information using Scalable Vector Graphics (SVG) and JavaScript technology. During the process of design, several GIWS are defined, and functions, interfaces and related methods of these services are discussed in detail. The paper focuses on the method for chaining distributed GIWS, the mechanism for geographic information dissemination and error handling. Finally, forest Geo-spatial Data which have two typical types of data E00 and Shapefile (SHP) was used to test the platform. The result indicates that using service chaining for multi-sources and heterogeneous Geo-spatial data integration and visualization can efficiently meet customized needs, but further research is needed for better application.

  17. Automated web service composition supporting conditional branch structures

    NASA Astrophysics Data System (ADS)

    Wang, Pengwei; Ding, Zhijun; Jiang, Changjun; Zhou, Mengchu

    2014-01-01

    The creation of value-added services by automatic composition of existing ones is gaining a significant momentum as the potential silver bullet in service-oriented architecture. However, service composition faces two aspects of difficulties. First, users' needs present such characteristics as diversity, uncertainty and personalisation; second, the existing services run in a real-world environment that is highly complex and dynamically changing. These difficulties may cause the emergence of nondeterministic choices in the process of service composition, which has gone beyond what the existing automated service composition techniques can handle. According to most of the existing methods, the process model of composite service includes sequence constructs only. This article presents a method to introduce conditional branch structures into the process model of composite service when needed, in order to satisfy users' diverse and personalised needs and adapt to the dynamic changes of real-world environment. UML activity diagrams are used to represent dependencies in composite service. Two types of user preferences are considered in this article, which have been ignored by the previous work and a simple programming language style expression is adopted to describe them. Two different algorithms are presented to deal with different situations. A real-life case is provided to illustrate the proposed concepts and methods.

  18. Integrating semantic web technologies and geospatial catalog services for geospatial information discovery and processing in cyberinfrastructure

    SciTech Connect

    Yue, Peng; Gong, Jianya; Di, Liping; He, Lianlian; Wei, Yaxing

    2011-04-01

    Abstract A geospatial catalogue service provides a network-based meta-information repository and interface for advertising and discovering shared geospatial data and services. Descriptive information (i.e., metadata) for geospatial data and services is structured and organized in catalogue services. The approaches currently available for searching and using that information are often inadequate. Semantic Web technologies show promise for better discovery methods by exploiting the underlying semantics. Such development needs special attention from the Cyberinfrastructure perspective, so that the traditional focus on discovery of and access to geospatial data can be expanded to support the increased demand for processing of geospatial information and discovery of knowledge. Semantic descriptions for geospatial data, services, and geoprocessing service chains are structured, organized, and registered through extending elements in the ebXML Registry Information Model (ebRIM) of a geospatial catalogue service, which follows the interface specifications of the Open Geospatial Consortium (OGC) Catalogue Services for the Web (CSW). The process models for geoprocessing service chains, as a type of geospatial knowledge, are captured, registered, and discoverable. Semantics-enhanced discovery for geospatial data, services/service chains, and process models is described. Semantic search middleware that can support virtual data product materialization is developed for the geospatial catalogue service. The creation of such a semantics-enhanced geospatial catalogue service is important in meeting the demands for geospatial information discovery and analysis in Cyberinfrastructure.

  19. An Approach for Web Service Selection Based on Confidence Level of Decision Maker

    PubMed Central

    Khezrian, Mojtaba; Jahan, Ali; Wan Kadir, Wan Mohd Nasir; Ibrahim, Suhaimi

    2014-01-01

    Web services today are among the most widely used groups for Service Oriented Architecture (SOA). Service selection is one of the most significant current discussions in SOA, which evaluates discovered services and chooses the best candidate from them. Although a majority of service selection techniques apply Quality of Service (QoS), the behaviour of QoS-based service selection leads to service selection problems in Multi-Criteria Decision Making (MCDM). In the existing works, the confidence level of decision makers is neglected and does not consider their expertise in assessing Web services. In this paper, we employ the VIKOR (VIšekriterijumskoKOmpromisnoRangiranje) method, which is absent in the literature for service selection, but is well-known in other research. We propose a QoS-based approach that deals with service selection by applying VIKOR with improvement of features. This research determines the weights of criteria based on user preference and accounts for the confidence level of decision makers. The proposed approach is illustrated by an example in order to demonstrate and validate the model. The results of this research may facilitate service consumers to attain a more efficient decision when selecting the appropriate service. PMID:24897426

  20. Prototype of Partial Cutting Tool of Geological Map Images Distributed by Geological Web Map Service

    NASA Astrophysics Data System (ADS)

    Nonogaki, S.; Nemoto, T.

    2014-12-01

    Geological maps and topographical maps play an important role in disaster assessment, resource management, and environmental preservation. These map information have been distributed in accordance with Web services standards such as Web Map Service (WMS) and Web Map Tile Service (WMTS) recently. In this study, a partial cutting tool of geological map images distributed by geological WMTS was implemented with Free and Open Source Software. The tool mainly consists of two functions: display function and cutting function. The former function was implemented using OpenLayers. The latter function was implemented using Geospatial Data Abstraction Library (GDAL). All other small functions were implemented by PHP and Python. As a result, this tool allows not only displaying WMTS layer on web browser but also generating a geological map image of intended area and zoom level. At this moment, available WTMS layers are limited to the ones distributed by WMTS for the Seamless Digital Geological Map of Japan. The geological map image can be saved as GeoTIFF format and WebGL format. GeoTIFF is one of the georeferenced raster formats that is available in many kinds of Geographical Information System. WebGL is useful for confirming a relationship between geology and geography in 3D. In conclusion, the partial cutting tool developed in this study would contribute to create better conditions for promoting utilization of geological information. Future work is to increase the number of available WMTS layers and the types of output file format.

  1. Distance Education Survey, 2007. A Report on Course Structure and Educational Services in Distance Education and Training Council Member Institutions

    ERIC Educational Resources Information Center

    Distance Education and Training Council, 2007

    2007-01-01

    In April 2007 the Distance Education and Training Council (DETC) surveyed 67 of its accredited institutions to determine current aspects of the distance study educational practice. (Military and international institutions were omitted.) This report is a collection and summary of the data received. This survey contained questions in the following…

  2. High-performance web services for querying gene and variant annotation.

    PubMed

    Xin, Jiwen; Mark, Adam; Afrasiabi, Cyrus; Tsueng, Ginger; Juchler, Moritz; Gopal, Nikhil; Stupp, Gregory S; Putman, Timothy E; Ainscough, Benjamin J; Griffith, Obi L; Torkamani, Ali; Whetzel, Patricia L; Mungall, Christopher J; Mooney, Sean D; Su, Andrew I; Wu, Chunlei

    2016-01-01

    Efficient tools for data management and integration are essential for many aspects of high-throughput biology. In particular, annotations of genes and human genetic variants are commonly used but highly fragmented across many resources. Here, we describe MyGene.info and MyVariant.info, high-performance web services for querying gene and variant annotation information. These web services are currently accessed more than three million times permonth. They also demonstrate a generalizable cloud-based model for organizing and querying biological annotation information. MyGene.info and MyVariant.info are provided as high-performance web services, accessible at http://mygene.info and http://myvariant.info . Both are offered free of charge to the research community. PMID:27154141

  3. Web service activities at the IRIS DMC to support federated and multidisciplinary access

    NASA Astrophysics Data System (ADS)

    Trabant, Chad; Ahern, Timothy K.

    2013-04-01

    At the IRIS Data Management Center (DMC) we have developed a suite of web service interfaces to access our large archive of, primarily seismological, time series data and related metadata. The goals of these web services include providing: a) next-generation and easily used access interfaces for our current users, b) access to data holdings in a form usable for non-seismologists, c) programmatic access to facilitate integration into data processing workflows and d) a foundation for participation in federated data discovery and access systems. To support our current users, our services provide access to the raw time series data and metadata or conversions of the raw data to commonly used formats. Our services also support simple, on-the-fly signal processing options that are common first steps in many workflows. Additionally, high-level data products derived from raw data are available via service interfaces. To support data access by researchers unfamiliar with seismic data we offer conversion of the data to broadly usable formats (e.g. ASCII text) and data processing to convert the data to Earth units. By their very nature, web services are programmatic interfaces. Combined with ubiquitous support for web technologies in programming & scripting languages and support in many computing environments, web services are very well suited for integrating data access into data processing workflows. As programmatic interfaces that can return data in both discipline-specific and broadly usable formats, our services are also well suited for participation in federated and brokered systems either specific to seismology or multidisciplinary. Working within the International Federation of Digital Seismograph Networks, the DMC collaborated on the specification of standardized web service interfaces for use at any seismological data center. These data access interfaces, when supported by multiple data centers, will form a foundation on which to build discovery and access mechanisms

  4. Programmatic access to data and information at the IRIS DMC via web services

    NASA Astrophysics Data System (ADS)

    Weertman, B. R.; Trabant, C.; Karstens, R.; Suleiman, Y. Y.; Ahern, T. K.; Casey, R.; Benson, R. B.

    2011-12-01

    The IRIS Data Management Center (DMC) has developed a suite of web services that provide access to the DMC's time series holdings, their related metadata and earthquake catalogs. In addition, services are available to perform simple, on-demand time series processing at the DMC prior to being shipped to the user. The primary goal is to provide programmatic access to data and processing services in a manner usable by and useful to the research community. The web services are relatively simple to understand and use and will form the foundation on which future DMC access tools will be built. Based on standard Web technologies they can be accessed programmatically with a wide range of programming languages (e.g. Perl, Python, Java), command line utilities such as wget and curl or with any web browser. We anticipate these services being used for everything from simple command line access, used in shell scripts and higher programming languages to being integrated within complex data processing software. In addition to improving access to our data by the seismological community the web services will also make our data more accessible to other disciplines. The web services available from the DMC include ws-bulkdataselect for the retrieval of large volumes of miniSEED data, ws-timeseries for the retrieval of individual segments of time series data in a variety of formats (miniSEED, SAC, ASCII, audio WAVE, and PNG plots) with optional signal processing, ws-station for station metadata in StationXML format, ws-resp for the retrieval of instrument response in RESP format, ws-sacpz for the retrieval of sensor response in the SAC poles and zeros convention and ws-event for the retrieval of earthquake catalogs. To make the services even easier to use, the DMC is developing a library that allows Java programmers to seamlessly retrieve and integrate DMC information into their own programs. The library will handle all aspects of dealing with the services and will parse the returned

  5. Wysiwyg Geoprocessing: Coupling Sensor Web and Geoprocessing Services in Virtual Globes

    NASA Astrophysics Data System (ADS)

    Zhai, X.; Gong, J.; Yue, P.; Sun, Z.; Lu, X.

    2011-08-01

    We propose to advance the scientific understanding and applications of geospatial data by coupling Sensor Web and Geoprocessing Services in Virtual Globes for higher-education teaching and research. The vision is the concept of "What You See is What You Get" geoprocessing, shortly known as WYSIWYG geoprocessing. Virtual Globes offer tremendous opportunities, such as providing a learning tool to help educational users and researchers digest global-scale geospatial information about the world, and acting as WYSIWYG platforms, where domain experts can see what their fingertips act in an interactive three-dimensional virtual environment. In the meantime, Sensor Web and Web Service technologies make a large amount of Earth observing sensors and geoprocessing functionalities easily accessible to educational users and researchers like their local resources. Coupling Sensor Web and geoprocessing Services in Virtual Globes will bring a virtual learning and research environment to the desktops of students and professors, empowering them with WYSIWYG geoprocessing capabilities. The implementation combines the visualization and communication power of Virtual Globes with the on-demand data collection and analysis functionalities of Sensor Web and geoprocessing services, to help students and researchers investigate various scientific problems in an environment with natural and intuitive user experiences. The work will contribute to the scientific and educational activities of geoinformatic communities in that they will have a platform that are easily accessible and help themselves perceive world space and perform live geoscientific processes.

  6. Web services at the IRIS DMC to support integration of data sets

    NASA Astrophysics Data System (ADS)

    Ahern, T.; Trabant, C.; Weertman, B.; Karstens, R.; Suleiman, Y.

    2012-04-01

    At the IRIS Data Management Center (DMC) we have developed web service interfaces to almost all of the data types we manage. These service interfaces provide access to raw time series data and associated metadata. In addition, we offer services that apply processing to the data before it is sent to the user. Whenever possible we chose open and recognized standards. The services themselves use only a simple subset of widespread web service technologies and methods, ensuring the widest possible support in programming languages and operating systems. To promote easier access to seismological and other geophysical data we are coordinating our web service developments with national and international parters. We are working within the Federation of Digital Seismograph Stations (FDSN) to define web service standards for international seismological data centers. Additionally we are working with key partners in Europe to complete the initial implementation of these international standards. Within the United States we have started coordinating with other data centers for related geophysical data types (e.g. geodetic) with the intention of facilitating cross-domain data discovery and use. The overall goal of these efforts is to make data more open, easily found and used by research scientists regardless of the repository in which any given data resides. To foster integration across geophysical domains the IRIS DMC is also making information available in the form of higher level products. These products often are more easily understood by a wider audience than the more esoteric seismographic time series. The support of products will support better integration of data sets. We will present the status of our web service developments and the related coordination efforts in addition to the next steps and vision for the future.

  7. Applying Semantic Web Services and Wireless Sensor Networks for System Integration

    NASA Astrophysics Data System (ADS)

    Berkenbrock, Gian Ricardo; Hirata, Celso Massaki; de Oliveira Júnior, Frederico Guilherme Álvares; de Oliveira, José Maria Parente

    In environments like factories, buildings, and homes automation services tend to often change during their lifetime. Changes are concerned to business rules, process optimization, cost reduction, and so on. It is important to provide a smooth and straightforward way to deal with these changes so that could be handled in a faster and low cost manner. Some prominent solutions use the flexibility of Wireless Sensor Networks and the meaningful description of Semantic Web Services to provide service integration. In this work, we give an overview of current solutions for machinery integration that combine both technologies as well as a discussion about some perspectives and open issues when applying Wireless Sensor Networks and Semantic Web Services for automation services integration.

  8. Description and testing of the Geo Data Portal: Data integration framework and Web processing services for environmental science collaboration

    USGS Publications Warehouse

    Blodgett, David L.; Booth, Nathaniel L.; Kunicki, Thomas C.; Walker, Jordan I.; Viger, Roland J.

    2011-01-01

    Interest in sharing interdisciplinary environmental modeling results and related data is increasing among scientists. The U.S. Geological Survey Geo Data Portal project enables data sharing by assembling open-standard Web services into an integrated data retrieval and analysis Web application design methodology that streamlines time-consuming and resource-intensive data management tasks. Data-serving Web services allow Web-based processing services to access Internet-available data sources. The Web processing services developed for the project create commonly needed derivatives of data in numerous formats. Coordinate reference system manipulation and spatial statistics calculation components implemented for the Web processing services were confirmed using ArcGIS 9.3.1, a geographic information science software package. Outcomes of the Geo Data Portal project support the rapid development of user interfaces for accessing and manipulating environmental data.

  9. Integrated web system of geospatial data services for climate research

    NASA Astrophysics Data System (ADS)

    Okladnikov, Igor; Gordov, Evgeny; Titov, Alexander

    2016-04-01

    Georeferenced datasets are currently actively used for modeling, interpretation and forecasting of climatic and ecosystem changes on different spatial and temporal scales. Due to inherent heterogeneity of environmental datasets as well as their huge size (up to tens terabytes for a single dataset) a special software supporting studies in the climate and environmental change areas is required. An approach for integrated analysis of georefernced climatological data sets based on combination of web and GIS technologies in the framework of spatial data infrastructure paradigm is presented. According to this approach a dedicated data-processing web system for integrated analysis of heterogeneous georeferenced climatological and meteorological data is being developed. It is based on Open Geospatial Consortium (OGC) standards and involves many modern solutions such as object-oriented programming model, modular composition, and JavaScript libraries based on GeoExt library, ExtJS Framework and OpenLayers software. This work is supported by the Ministry of Education and Science of the Russian Federation, Agreement #14.613.21.0037.

  10. Implementing a Web-based clinical information system using EMR middle layer services.

    PubMed Central

    Kittredge, R. L.; Estey, G.; Pappas, J. J.; Barnett, G. O.

    1996-01-01

    The Clinical Summary is a Web-based application for accessing the clinical database at the Massachusetts General Hospital. The application has been developed to give physicians in our health care community access to clinical information for patients they refer to our hospital. "Middle layer" services, written previously for the hospital's clinical workstation, supply much of the application's functionality. Employment of reusable services together with a Web-based front end has afforded a rapid and inexpensive means for developing a new clinical information system. This paper discusses the system's design, function, and methods of implementation. PMID:8947742

  11. Performance Issues Related to Web Service Usage for Remote Data Access

    SciTech Connect

    Pais, V. F.; Stancalie, V.; Mihailescu, F. A.; Totolici, M. C.

    2008-04-07

    Web services are starting to be widely used in applications for remotely accessing data. This is of special interest for research based on small and medium scale fusion devices, since scientists participating remotely to experiments are accessing large amounts of data over the Internet. Recent tests were conducted to see how the new network traffic, generated by the use of web services, can be integrated in the existing infrastructure and what would be the impact over existing applications, especially those used in a remote participation scenario.

  12. A Methodology for the Development of RESTful Semantic Web Services for Gene Expression Analysis

    PubMed Central

    Guardia, Gabriela D. A.; Pires, Luís Ferreira; Vêncio, Ricardo Z. N.; Malmegrim, Kelen C. R.; de Farias, Cléver R. G.

    2015-01-01

    Gene expression studies are generally performed through multi-step analysis processes, which require the integrated use of a number of analysis tools. In order to facilitate tool/data integration, an increasing number of analysis tools have been developed as or adapted to semantic web services. In recent years, some approaches have been defined for the development and semantic annotation of web services created from legacy software tools, but these approaches still present many limitations. In addition, to the best of our knowledge, no suitable approach has been defined for the functional genomics domain. Therefore, this paper aims at defining an integrated methodology for the implementation of RESTful semantic web services created from gene expression analysis tools and the semantic annotation of such services. We have applied our methodology to the development of a number of services to support the analysis of different types of gene expression data, including microarray and RNASeq. All developed services are publicly available in the Gene Expression Analysis Services (GEAS) Repository at http://dcm.ffclrp.usp.br/lssb/geas. Additionally, we have used a number of the developed services to create different integrated analysis scenarios to reproduce parts of two gene expression studies documented in the literature. The first study involves the analysis of one-color microarray data obtained from multiple sclerosis patients and healthy donors. The second study comprises the analysis of RNA-Seq data obtained from melanoma cells to investigate the role of the remodeller BRG1 in the proliferation and morphology of these cells. Our methodology provides concrete guidelines and technical details in order to facilitate the systematic development of semantic web services. Moreover, it encourages the development and reuse of these services for the creation of semantically integrated solutions for gene expression analysis. PMID:26207740

  13. A Methodology for the Development of RESTful Semantic Web Services for Gene Expression Analysis.

    PubMed

    Guardia, Gabriela D A; Pires, Luís Ferreira; Vêncio, Ricardo Z N; Malmegrim, Kelen C R; de Farias, Cléver R G

    2015-01-01

    Gene expression studies are generally performed through multi-step analysis processes, which require the integrated use of a number of analysis tools. In order to facilitate tool/data integration, an increasing number of analysis tools have been developed as or adapted to semantic web services. In recent years, some approaches have been defined for the development and semantic annotation of web services created from legacy software tools, but these approaches still present many limitations. In addition, to the best of our knowledge, no suitable approach has been defined for the functional genomics domain. Therefore, this paper aims at defining an integrated methodology for the implementation of RESTful semantic web services created from gene expression analysis tools and the semantic annotation of such services. We have applied our methodology to the development of a number of services to support the analysis of different types of gene expression data, including microarray and RNASeq. All developed services are publicly available in the Gene Expression Analysis Services (GEAS) Repository at http://dcm.ffclrp.usp.br/lssb/geas. Additionally, we have used a number of the developed services to create different integrated analysis scenarios to reproduce parts of two gene expression studies documented in the literature. The first study involves the analysis of one-color microarray data obtained from multiple sclerosis patients and healthy donors. The second study comprises the analysis of RNA-Seq data obtained from melanoma cells to investigate the role of the remodeller BRG1 in the proliferation and morphology of these cells. Our methodology provides concrete guidelines and technical details in order to facilitate the systematic development of semantic web services. Moreover, it encourages the development and reuse of these services for the creation of semantically integrated solutions for gene expression analysis.

  14. A Methodology for the Development of RESTful Semantic Web Services for Gene Expression Analysis.

    PubMed

    Guardia, Gabriela D A; Pires, Luís Ferreira; Vêncio, Ricardo Z N; Malmegrim, Kelen C R; de Farias, Cléver R G

    2015-01-01

    Gene expression studies are generally performed through multi-step analysis processes, which require the integrated use of a number of analysis tools. In order to facilitate tool/data integration, an increasing number of analysis tools have been developed as or adapted to semantic web services. In recent years, some approaches have been defined for the development and semantic annotation of web services created from legacy software tools, but these approaches still present many limitations. In addition, to the best of our knowledge, no suitable approach has been defined for the functional genomics domain. Therefore, this paper aims at defining an integrated methodology for the implementation of RESTful semantic web services created from gene expression analysis tools and the semantic annotation of such services. We have applied our methodology to the development of a number of services to support the analysis of different types of gene expression data, including microarray and RNASeq. All developed services are publicly available in the Gene Expression Analysis Services (GEAS) Repository at http://dcm.ffclrp.usp.br/lssb/geas. Additionally, we have used a number of the developed services to create different integrated analysis scenarios to reproduce parts of two gene expression studies documented in the literature. The first study involves the analysis of one-color microarray data obtained from multiple sclerosis patients and healthy donors. The second study comprises the analysis of RNA-Seq data obtained from melanoma cells to investigate the role of the remodeller BRG1 in the proliferation and morphology of these cells. Our methodology provides concrete guidelines and technical details in order to facilitate the systematic development of semantic web services. Moreover, it encourages the development and reuse of these services for the creation of semantically integrated solutions for gene expression analysis. PMID:26207740

  15. Interactive distance education service (IDES) using the Centralized Access Node System (CANS)

    NASA Astrophysics Data System (ADS)

    Moon, Phil J.; Choi, Mun K.; Jun, Mun S.; Lee, Chul H.

    1995-02-01

    The successes of B-ISDN project depend on attractive application services. Thus, it is very important to analyze application services which users desire, and to study the customer access network for applications. Especially, because the cost of the constructions for the customer access network need high, it is very important to configure this network economically and effectively. We recognize the importance how to provide B-ISDN application services and to configure the customer access network. This paper describes the Interactive Distance Education Service (IDES) requirements in the user and network aspects. The requirements include the service definition, reference configuration, procedure and Quality Of Service. And also this paper proposes the configuration of customer access network for providing IDES using the Centralized Access Node system which is a kind of the B-NT system. There are many applications area in IDES, we describe the three scenarios--for rural area, between the major and minor campus and for joint education including the university, research institute and enterprise.

  16. Technical Note: Harmonizing met-ocean model data via standard web services within small research groups

    USGS Publications Warehouse

    Signell, Richard; Camossi, E.

    2016-01-01

    Work over the last decade has resulted in standardised web services and tools that can significantly improve the efficiency and effectiveness of working with meteorological and ocean model data. While many operational modelling centres have enabled query and access to data via common web services, most small research groups have not. The penetration of this approach into the research community, where IT resources are limited, can be dramatically improved by (1) making it simple for providers to enable web service access to existing output files; (2) using free technologies that are easy to deploy and configure; and (3) providing standardised, service-based tools that work in existing research environments. We present a simple, local brokering approach that lets modellers continue to use their existing files and tools, while serving virtual data sets that can be used with standardised tools. The goal of this paper is to convince modellers that a standardised framework is not only useful but can be implemented with modest effort using free software components. We use NetCDF Markup language for data aggregation and standardisation, the THREDDS Data Server for data delivery, pycsw for data search, NCTOOLBOX (MATLAB®) and Iris (Python) for data access, and Open Geospatial Consortium Web Map Service for data preview. We illustrate the effectiveness of this approach with two use cases involving small research modelling groups at NATO and USGS.

  17. Technical note: Harmonising metocean model data via standard web services within small research groups

    NASA Astrophysics Data System (ADS)

    Signell, Richard P.; Camossi, Elena

    2016-05-01

    Work over the last decade has resulted in standardised web services and tools that can significantly improve the efficiency and effectiveness of working with meteorological and ocean model data. While many operational modelling centres have enabled query and access to data via common web services, most small research groups have not. The penetration of this approach into the research community, where IT resources are limited, can be dramatically improved by (1) making it simple for providers to enable web service access to existing output files; (2) using free technologies that are easy to deploy and configure; and (3) providing standardised, service-based tools that work in existing research environments. We present a simple, local brokering approach that lets modellers continue to use their existing files and tools, while serving virtual data sets that can be used with standardised tools. The goal of this paper is to convince modellers that a standardised framework is not only useful but can be implemented with modest effort using free software components. We use NetCDF Markup language for data aggregation and standardisation, the THREDDS Data Server for data delivery, pycsw for data search, NCTOOLBOX (MATLAB®) and Iris (Python) for data access, and Open Geospatial Consortium Web Map Service for data preview. We illustrate the effectiveness of this approach with two use cases involving small research modelling groups at NATO and USGS.

  18. The Scope of WebCounseling: A Survey of Services and Compliance with NBCC "Standards for the Ethical Practice of WebCounseling."

    ERIC Educational Resources Information Center

    Heinlen, Kathleen T.; Welfel, Elizabeth Reynolds; Richmond, Elizabeth N.; Rak, Carl F.

    2003-01-01

    A survey of 136 Web sites offering counseling through computer chat rooms and e-mail revealed a wide range of credentials of providers, fees for service, and low levels of compliance with the ethical standards for WebCounseling published by the National Board for Certified Counselors. Implications for professional practice and public confidence in…

  19. Clearing your Desk! Software and Data Services for Collaborative Web Based GIS Analysis

    NASA Astrophysics Data System (ADS)

    Tarboton, D. G.; Idaszak, R.; Horsburgh, J. S.; Ames, D. P.; Goodall, J. L.; Band, L. E.; Merwade, V.; Couch, A.; Hooper, R. P.; Maidment, D. R.; Dash, P. K.; Stealey, M.; Yi, H.; Gan, T.; Gichamo, T.; Yildirim, A. A.; Liu, Y.

    2015-12-01

    Can your desktop computer crunch the large GIS datasets that are becoming increasingly common across the geosciences? Do you have access to or the know-how to take advantage of advanced high performance computing (HPC) capability? Web based cyberinfrastructure takes work off your desk or laptop computer and onto infrastructure or "cloud" based data and processing servers. This talk will describe the HydroShare collaborative environment and web based services being developed to support the sharing and processing of hydrologic data and models. HydroShare supports the upload, storage, and sharing of a broad class of hydrologic data including time series, geographic features and raster datasets, multidimensional space-time data, and other structured collections of data. Web service tools and a Python client library provide researchers with access to HPC resources without requiring them to become HPC experts. This reduces the time and effort spent in finding and organizing the data required to prepare the inputs for hydrologic models and facilitates the management of online data and execution of models on HPC systems. This presentation will illustrate the use of web based data and computation services from both the browser and desktop client software. These web-based services implement the Terrain Analysis Using Digital Elevation Model (TauDEM) tools for watershed delineation, generation of hydrology-based terrain information, and preparation of hydrologic model inputs. They allow users to develop scripts on their desktop computer that call analytical functions that are executed completely in the cloud, on HPC resources using input datasets stored in the cloud, without installing specialized software, learning how to use HPC, or transferring large datasets back to the user's desktop. These cases serve as examples for how this approach can be extended to other models to enhance the use of web and data services in the geosciences.

  20. Eight ways to go "e" without launching a Web service.

    PubMed

    Toth, C L; Goldstein, D

    2000-01-01

    With more than 100 million Americans on the Internet, the physician-patient relationship is evolving. The Net brings with it unlimited potential for communication and information dissemination, but these benefits do come with some side effects. No everything on the Net is reliable or credible, but many patients don't understand that. For this reason and others, physicians must involve themselves with online communication and e-commerce. There are many realistic and practical ways for physicians to do this, and not all of them involve building a Web site. Sending patients clinical information in a cost-effective manner, improving receivables management, and decreasing the volume of phone calls are just some of the many ways the Net can enhance your practice. Think of the Net as just another business tool--not a replacement for physician-patient relationships. Like the telephone before it, the Net affords improved communication, patient education, and relationship-building with patients. PMID:14608768

  1. The BCube Crawler: Web Scale Data and Service Discovery for EarthCube.

    NASA Astrophysics Data System (ADS)

    Lopez, L. A.; Khalsa, S. J. S.; Duerr, R.; Tayachow, A.; Mingo, E.

    2014-12-01

    Web-crawling, a core component of the NSF-funded BCube project, is researching and applying the use of big data technologies to find and characterize different types of web services, catalog interfaces, and data feeds such as the ESIP OpenSearch, OGC W*S, THREDDS, and OAI-PMH that describe or provide access to scientific datasets. Given the scale of the Internet, which challenges even large search providers such as Google, the BCube plan for discovering these web accessible services is to subdivide the problem into three smaller, more tractable issues. The first, to be able to discover likely sites where relevant data and data services might be found, the second, to be able to deeply crawl the sites discovered to find any data and services which might be present. Lastly, to leverage the use of semantic technologies to characterize the services and data found, and to filter out everything but those relevant to the geosciences. To address the first two challenges BCube uses an adapted version of Apache Nutch (which originated Hadoop), a web scale crawler, and Amazon's ElasticMapReduce service for flexibility and cost effectiveness. For characterization of the services found, BCube is examining existing web service ontologies for their applicability to our needs and will re-use and/or extend these in order to query for services with specific well-defined characteristics in scientific datasets such as the use of geospatial namespaces. The original proposal for the crawler won a grant from Amazon's academic program, which allowed us to become operational; we successfully tested the Bcube Crawler at web scale obtaining a significant corpus, sizeable enough to enable work on characterization of the services and data found. There is still plenty of work to be done, doing "smart crawls" by managing the frontier, developing and enhancing our scoring algorithms and fully implementing the semantic characterization technologies. We describe the current status of the project

  2. Do Altmetrics Work? Twitter and Ten Other Social Web Services

    PubMed Central

    Thelwall, Mike; Haustein, Stefanie; Larivière, Vincent; Sugimoto, Cassidy R.

    2013-01-01

    Altmetric measurements derived from the social web are increasingly advocated and used as early indicators of article impact and usefulness. Nevertheless, there is a lack of systematic scientific evidence that altmetrics are valid proxies of either impact or utility although a few case studies have reported medium correlations between specific altmetrics and citation rates for individual journals or fields. To fill this gap, this study compares 11 altmetrics with Web of Science citations for 76 to 208,739 PubMed articles with at least one altmetric mention in each case and up to 1,891 journals per metric. It also introduces a simple sign test to overcome biases caused by different citation and usage windows. Statistically significant associations were found between higher metric scores and higher citations for articles with positive altmetric scores in all cases with sufficient evidence (Twitter, Facebook wall posts, research highlights, blogs, mainstream media and forums) except perhaps for Google+ posts. Evidence was insufficient for LinkedIn, Pinterest, question and answer sites, and Reddit, and no conclusions should be drawn about articles with zero altmetric scores or the strength of any correlation between altmetrics and citations. Nevertheless, comparisons between citations and metric values for articles published at different times, even within the same year, can remove or reverse this association and so publishers and scientometricians should consider the effect of time when using altmetrics to rank articles. Finally, the coverage of all the altmetrics except for Twitter seems to be low and so it is not clear if they are prevalent enough to be useful in practice. PMID:23724101

  3. Applying Web Service on the Simulation and Demonstration for Groundwater Resources

    NASA Astrophysics Data System (ADS)

    liu, H.; Wang, H.; Chang, L.

    2013-12-01

    This study uses web service technology to reach following goals: (1) unified data format, (2) cross-platform data input/output, (3) linkage to pre-developed groundwater simulation model with database, and (4) friendly cross-platform user interface. With this web service technology, users can remotely (1) read, extract, manage, and display groundwater related data and (2) run the selected groundwater simulation model. In this study, three web formats are provided for different needs: (1)website, (2) the format for mobile devices, and (3) the interface for researchers. For the first two formats, the users can visualize the selected data or model results. The web format allows users to browse it using commonly available browsers. The Android system is used for the second format, which is for mobile devices, simplifying the data demonstration process. Investigators can use mobile devices to show space information, groundwater level, and hydrogeological parameters. The third format provides professional researchers to use the provided API to access massive groundwater related data and parameters. The developed web service is applied to Pingtong groundwater area to simulate multiple groundwater flow scenarios and show the simulation results.

  4. Load Index Metrics for an Optimized Management of Web Services: A Systematic Evaluation

    PubMed Central

    Souza, Paulo S. L.; Santana, Regina H. C.; Santana, Marcos J.; Zaluska, Ed; Faical, Bruno S.; Estrella, Julio C.

    2013-01-01

    The lack of precision to predict service performance through load indices may lead to wrong decisions regarding the use of web services, compromising service performance and raising platform cost unnecessarily. This paper presents experimental studies to qualify the behaviour of load indices in the web service context. The experiments consider three services that generate controlled and significant server demands, four levels of workload for each service and six distinct execution scenarios. The evaluation considers three relevant perspectives: the capability for representing recent workloads, the capability for predicting near-future performance and finally stability. Eight different load indices were analysed, including the JMX Average Time index (proposed in this paper) specifically designed to address the limitations of the other indices. A systematic approach is applied to evaluate the different load indices, considering a multiple linear regression model based on the stepwise-AIC method. The results show that the load indices studied represent the workload to some extent; however, in contrast to expectations, most of them do not exhibit a coherent correlation with service performance and this can result in stability problems. The JMX Average Time index is an exception, showing a stable behaviour which is tightly-coupled to the service runtime for all executions. Load indices are used to predict the service runtime and therefore their inappropriate use can lead to decisions that will impact negatively on both service performance and execution cost. PMID:23874776

  5. A Study of Older Adult Students' Satisfaction with Web-Based Distance Learning at the National Open University of Taiwan

    ERIC Educational Resources Information Center

    Chen, Ho-Yuan

    2010-01-01

    The purpose of this study was to investigate the relationships between older learners' demographic characteristics and their satisfaction with distance learning in the Web-based environment at National Open University in Taiwan (NOUT). Increases in the older adult population have had many impacts throughout societies. The major purpose of…

  6. The Use of Video-Taped Lectures and Web-Based Communications in Teaching: A Distance-Teaching and Cross-Atlantic Collaboration Experiment.

    ERIC Educational Resources Information Center

    Herder, P. M.; Subrahmanian, E.; Talukdar, S.; Turk, A. L.; Westerberg, A. W.

    2002-01-01

    Explains distance education approach applied to the 'Engineering Design Problem Formulation' course simultaneously at the Delft University of Technology (the Netherlands) and at Carnegie Mellon University (CMU, Pittsburgh, USA). Uses video taped lessons, video conferencing, electronic mails and web-accessible document management system LIRE in the…

  7. ChEMBL web services: streamlining access to drug discovery data and utilities.

    PubMed

    Davies, Mark; Nowotka, Michał; Papadatos, George; Dedman, Nathan; Gaulton, Anna; Atkinson, Francis; Bellis, Louisa; Overington, John P

    2015-07-01

    ChEMBL is now a well-established resource in the fields of drug discovery and medicinal chemistry research. The ChEMBL database curates and stores standardized bioactivity, molecule, target and drug data extracted from multiple sources, including the primary medicinal chemistry literature. Programmatic access to ChEMBL data has been improved by a recent update to the ChEMBL web services (version 2.0.x, https://www.ebi.ac.uk/chembl/api/data/docs), which exposes significantly more data from the underlying database and introduces new functionality. To complement the data-focused services, a utility service (version 1.0.x, https://www.ebi.ac.uk/chembl/api/utils/docs), which provides RESTful access to commonly used cheminformatics methods, has also been concurrently developed. The ChEMBL web services can be used together or independently to build applications and data processing workflows relevant to drug discovery and chemical biology.

  8. Lynx web services for annotations and systems analysis of multi-gene disorders.

    PubMed

    Sulakhe, Dinanath; Taylor, Andrew; Balasubramanian, Sandhya; Feng, Bo; Xie, Bingqing; Börnigen, Daniela; Dave, Utpal J; Foster, Ian T; Gilliam, T Conrad; Maltsev, Natalia

    2014-07-01

    Lynx is a web-based integrated systems biology platform that supports annotation and analysis of experimental data and generation of weighted hypotheses on molecular mechanisms contributing to human phenotypes and disorders of interest. Lynx has integrated multiple classes of biomedical data (genomic, proteomic, pathways, phenotypic, toxicogenomic, contextual and others) from various public databases as well as manually curated data from our group and collaborators (LynxKB). Lynx provides tools for gene list enrichment analysis using multiple functional annotations and network-based gene prioritization. Lynx provides access to the integrated database and the analytical tools via REST based Web Services (http://lynx.ci.uchicago.edu/webservices.html). This comprises data retrieval services for specific functional annotations, services to search across the complete LynxKB (powered by Lucene), and services to access the analytical tools built within the Lynx platform.

  9. Semantic-based web service discovery and chaining for building an Arctic spatial data infrastructure

    NASA Astrophysics Data System (ADS)

    Li, W.; Yang, C.; Nebert, D.; Raskin, R.; Houser, P.; Wu, H.; Li, Z.

    2011-11-01

    Increasing interests in a global environment and climate change have led to studies focused on the changes in the multinational Arctic region. To facilitate Arctic research, a spatial data infrastructure (SDI), where Arctic data, information, and services are shared and integrated in a seamless manner, particularly in light of today's climate change scenarios, is urgently needed. In this paper, we utilize the knowledge-based approach and the spatial web portal technology to prototype an Arctic SDI (ASDI) by proposing (1) a hybrid approach for efficient service discovery from distributed web catalogs and the dynamic Internet; (2) a domain knowledge base to model the latent semantic relationships among scientific data and services; and (3) an intelligent logic reasoning mechanism for (semi-)automatic service selection and chaining. A study of the influence of solid water dynamics to the bio-habitat of the Arctic region is used as an example to demonstrate the prototype.

  10. Web-Based Academic Support Services: Guidelines for Extensibility

    ERIC Educational Resources Information Center

    McCracken, Holly

    2005-01-01

    Using the experience of the University of Illinois at Springfield's College of Liberal Arts and Sciences at the as a foundation for discussion, this paper addresses the provision of student support services to distant students within the context of development and expansion. Specific issues for consideration include: integrating student support…

  11. Context-Adaptive Learning Designs by Using Semantic Web Services

    ERIC Educational Resources Information Center

    Dietze, Stefan; Gugliotta, Alessio; Domingue, John

    2007-01-01

    IMS Learning Design (IMS-LD) is a promising technology aimed at supporting learning processes. IMS-LD packages contain the learning process metadata as well as the learning resources. However, the allocation of resources--whether data or services--within the learning design is done manually at design-time on the basis of the subjective appraisals…

  12. New Web-Monitoring Service Worries Some Legal Experts

    ERIC Educational Resources Information Center

    Sander, Libby

    2008-01-01

    A software program that searches for offensive content on college athletes' social-networking sites has drawn skeptical reactions from legal experts, who say it could threaten students' constitutional rights. Billed as a "social-network monitoring service" and marketed exclusively to college athletics departments, YouDiligence was on display at…

  13. Tampa Bay Ecosystem Services Demonstration Pilot Phase 2 web site

    EPA Science Inventory

    The value of nature's benefits is difficult to consider in environmental decision-making since ecosystem goods and services are usually not well measured or quantified in economic terms. The Tampa Bay Estuary Program, Tampa Bay Regional Planning Council, the U.S. Environmental Pr...

  14. Enabling Interoperability and Servicing Multiple User Segments Through Web Services, Standards, and Data Tools

    NASA Astrophysics Data System (ADS)

    Palanisamy, Giriprakash; Wilson, Bruce E.; Cook, Robert B.; Lenhardt, Chris W.; Santhana Vannan, Suresh; Pan, Jerry; McMurry, Ben F.; Devarakonda, Ranjeet

    2010-12-01

    The Oak Ridge National Laboratory Distributed Active Archive Center (ORNL DAAC) is one of the science-oriented data centers in EOSDIS, aligned primarily with terrestrial ecology. The ORNL DAAC archives and serves data from NASA-funded field campaigns (such as BOREAS, FIFE, and LBA), regional and global data sets relevant to biogeochemical cycles, land validation studies for remote sensing, and source code for some terrestrial ecology models. Users of the ORNL DAAC include field ecologists, remote sensing scientists, modelers at various scales, synthesis scientific groups, a range of educational users (particularly baccalaureate and graduate instruction), and decision support analysts. It is clear that the wide range of users served by the ORNL DAAC have differing needs and differing capabilities for accessing and using data. It is also not possible for the ORNL DAAC, or the other data centers in EDSS to develop all of the tools and interfaces to support even most of the potential uses of data directly. As is typical of Information Technology to support a research enterprise, the user needs will continue to evolve rapidly over time and users themselves cannot predict future needs, as those needs depend on the results of current investigation. The ORNL DAAC is addressing these needs by targeted implementation of web services and tools which can be consumed by other applications, so that a modeler can retrieve data in netCDF format with the Climate Forecasting convention and a field ecologist can retrieve subsets of that same data in a comma separated value format, suitable for use in Excel or R. Tools such as our MODIS Subsetting capability, the Spatial Data Access Tool (SDAT; based on OGC web services), and OPeNDAP-compliant servers such as THREDDS particularly enable such diverse means of access. We also seek interoperability of metadata, recognizing that terrestrial ecology is a field where there are a very large number of relevant data repositories. ORNL DAAC

  15. Web Services as Building Blocks for an Open Coastal Observing System

    NASA Astrophysics Data System (ADS)

    Breitbach, G.; Krasemann, H.

    2012-04-01

    In coastal observing systems it is needed to integrate different observing methods like remote sensing, in-situ measurements, and models into a synoptic view of the state of the observed region. This integration can be based solely on web services combining data and metadata. Such an approach is pursued for COSYNA (Coastal Observing System for Northern and Artic seas). Data from satellite and radar remote sensing, measurements of buoys, stations and Ferryboxes are the observation part of COSYNA. These data are assimilated into models to create pre-operational forecasts. For discovering data an OGC Web Feature Service (WFS) is used by the COSYNA data portal. This Web Feature Service knows the necessary metadata not only for finding data, but in addition the URLs of web services to view and download the data. To make the data from different resources comparable a common vocabulary is needed. For COSYNA the standard names from CF-conventions are stored within the metadata whenever possible. For the metadata an INSPIRE and ISO19115 compatible data format is used. The WFS is fed from the metadata-system using database-views. Actual data are stored in two different formats, in NetCDF-files for gridded data and in an RDBMS for time-series-like data. The web service URLs are mostly standard based the standards are mainly OGC standards. Maps were created from netcdf files with the help of the ncWMS tool whereas a self-developed java servlet is used for maps of moving measurement platforms. In this case download of data is offered via OGC SOS. For NetCDF-files OPeNDAP is used for the data download. The OGC CSW is used for accessing extended metadata. The concept of data management in COSYNA will be presented which is independent of the special services used in COSYNA. This concept is parameter and data centric and might be useful for other observing systems.

  16. Design, Implementation and Applications of 3d Web-Services in DB4GEO

    NASA Astrophysics Data System (ADS)

    Breunig, M.; Kuper, P. V.; Dittrich, A.; Wild, P.; Butwilowski, E.; Al-Doori, M.

    2013-09-01

    The object-oriented database architecture DB4GeO was originally designed to support sub-surface applications in the geo-sciences. This is reflected in DB4GeO's geometric data model as well as in its import and export functions. Initially, these functions were designed for communication with 3D geological modeling and visualization tools such as GOCAD or MeshLab. However, it soon became clear that DB4GeO was suitable for a much wider range of applications. Therefore it is natural to move away from a standalone solution and to open the access to DB4GeO data by standardized OGC web-services. Though REST and OGC services seem incompatible at first sight, the implementation in DB4GeO shows that OGC-based implementation of web-services may use parts of the DB4GeO-REST implementation. Starting with initial solutions in the history of DB4GeO, this paper will introduce the design, adaptation (i.e. model transformation), and first steps in the implementation of OGC Web Feature (WFS) and Web Processing Services (WPS), as new interfaces to DB4GeO data and operations. Among its capabilities, DB4GeO can provide data in different data formats like GML, GOCAD, or DB3D XML through a WFS, as well as its ability to run operations like a 3D-to-2D service, or mesh-simplification (Progressive Meshes) through a WPS. We then demonstrate, an Android-based mobile 3D augmented reality viewer for DB4GeO that uses the Web Feature Service to visualize 3D geo-database query results. Finally, we explore future research work considering DB4GeO in the framework of the research group "Computer-Aided Collaborative Subway Track Planning in Multi-Scale 3D City and Building Models".

  17. Cloud-based Web Services for Near-Real-Time Web access to NPP Satellite Imagery and other Data

    NASA Astrophysics Data System (ADS)

    Evans, J. D.; Valente, E. G.

    2010-12-01

    We are building a scalable, cloud computing-based infrastructure for Web access to near-real-time data products synthesized from the U.S. National Polar-Orbiting Environmental Satellite System (NPOESS) Preparatory Project (NPP) and other geospatial and meteorological data. Given recent and ongoing changes in the the NPP and NPOESS programs (now Joint Polar Satellite System), the need for timely delivery of NPP data is urgent. We propose an alternative to a traditional, centralized ground segment, using distributed Direct Broadcast facilities linked to industry-standard Web services by a streamlined processing chain running in a scalable cloud computing environment. Our processing chain, currently implemented on Amazon.com's Elastic Compute Cloud (EC2), retrieves raw data from NASA's Moderate Resolution Imaging Spectroradiometer (MODIS) and synthesizes data products such as Sea-Surface Temperature, Vegetation Indices, etc. The cloud computing approach lets us grow and shrink computing resources to meet large and rapid fluctuations (twice daily) in both end-user demand and data availability from polar-orbiting sensors. Early prototypes have delivered various data products to end-users with latencies between 6 and 32 minutes. We have begun to replicate machine instances in the cloud, so as to reduce latency and maintain near-real time data access regardless of increased data input rates or user demand -- all at quite moderate monthly costs. Our service-based approach (in which users invoke software processes on a Web-accessible server) facilitates access into datasets of arbitrary size and resolution, and allows users to request and receive tailored and composite (e.g., false-color multiband) products on demand. To facilitate broad impact and adoption of our technology, we have emphasized open, industry-standard software interfaces and open source software. Through our work, we envision the widespread establishment of similar, derived, or interoperable systems for

  18. Software architecture and design of the web services facilitating climate model diagnostic analysis

    NASA Astrophysics Data System (ADS)

    Pan, L.; Lee, S.; Zhang, J.; Tang, B.; Zhai, C.; Jiang, J. H.; Wang, W.; Bao, Q.; Qi, M.; Kubar, T. L.; Teixeira, J.

    2015-12-01

    Climate model diagnostic analysis is a computationally- and data-intensive task because it involves multiple numerical model outputs and satellite observation data that can both be high resolution. We have built an online tool that facilitates this process. The tool is called Climate Model Diagnostic Analyzer (CMDA). It employs the web service technology and provides a web-based user interface. The benefits of these choices include: (1) No installation of any software other than a browser, hence it is platform compatable; (2) Co-location of computation and big data on the server side, and small results and plots to be downloaded on the client side, hence high data efficiency; (3) multi-threaded implementation to achieve parallel performance on multi-core servers; and (4) cloud deployment so each user has a dedicated virtual machine. In this presentation, we will focus on the computer science aspects of this tool, namely the architectural design, the infrastructure of the web services, the implementation of the web-based user interface, the mechanism of provenance collection, the approach to virtualization, and the Amazon Cloud deployment. As an example, We will describe our methodology to transform an existing science application code into a web service using a Python wrapper interface and Python web service frameworks (i.e., Flask, Gunicorn, and Tornado). Another example is the use of Docker, a light-weight virtualization container, to distribute and deploy CMDA onto an Amazon EC2 instance. Our tool of CMDA has been successfully used in the 2014 Summer School hosted by the JPL Center for Climate Science. Students had positive feedbacks in general and we will report their comments. An enhanced version of CMDA with several new features, some requested by the 2014 students, will be used in the 2015 Summer School soon.

  19. [Solution of PACS to communicate with the handheld mobile intelligent equipment based on web service].

    PubMed

    Xue, Weijing; Wang, Pengcheng; Meng, Jian; Wang, Haiyang

    2013-09-01

    Using the Web Service technology, the paper puts forward a solution of remote communication between PACS and the handheld mobile intelligent devices, the characteristics of the solution are simple, easy to implement, short development cycle, high practicability, and low cost etc. PMID:24409803

  20. Electronic Resources for Youth Services: A Print Bibliography and Web Site.

    ERIC Educational Resources Information Center

    Amey, Larry; Segal, Erez

    1996-01-01

    This article evaluates 57 World Wide Web sites related to children's literature and youth-oriented library services, in categories including award-winning books; book reviews; reading and storytelling; writing resources; online children's literature; educational entertainment; and authors, publishers, and booksellers. Also included is information…

  1. Design and implementation of web service-based mobile reply testing management system

    NASA Astrophysics Data System (ADS)

    Gao, Liang; Han, Yu-min

    2013-03-01

    This paper introduces the design and implementation of a reply testing management system that based-on C#, Web Service and ASP .Net. With the system, teachers can manage reply site through Windows Mobile, judges can grade onscene by visiting WAP site through GPRS. Then this paper discusses details of the implementation and key technologies of the system.

  2. Oh! Web 2.0, Virtual Reference Service 2.0, Tools & Techniques (II)

    ERIC Educational Resources Information Center

    Arya, Harsh Bardhan; Mishra, J. K.

    2012-01-01

    The paper describes the theory and definition of the practice of librarianship, specifically addressing how Web 2.0 technologies (tools) such as synchronous messaging, collaborative reference service and streaming media, blogs, wikis, social networks, social bookmarking tools, tagging, RSS feeds, and mashups might intimate changes and how…

  3. A Framework for Sharing and Integrating Remote Sensing and GIS Models Based on Web Service

    PubMed Central

    Chen, Zeqiang; Lin, Hui; Chen, Min; Liu, Deer; Bao, Ying; Ding, Yulin

    2014-01-01

    Sharing and integrating Remote Sensing (RS) and Geographic Information System/Science (GIS) models are critical for developing practical application systems. Facilitating model sharing and model integration is a problem for model publishers and model users, respectively. To address this problem, a framework based on a Web service for sharing and integrating RS and GIS models is proposed in this paper. The fundamental idea of the framework is to publish heterogeneous RS and GIS models into standard Web services for sharing and interoperation and then to integrate the RS and GIS models using Web services. For the former, a “black box” and a visual method are employed to facilitate the publishing of the models as Web services. For the latter, model integration based on the geospatial workflow and semantic supported marching method is introduced. Under this framework, model sharing and integration is applied for developing the Pearl River Delta water environment monitoring system. The results show that the framework can facilitate model sharing and model integration for model publishers and model users. PMID:24901016

  4. Freshman Admissions Predictor: An Interactive Self-Help Web Counseling Service

    ERIC Educational Resources Information Center

    Head, Joe F.; Hughes, Thomas M.

    2004-01-01

    Colleges and universities must seek or develop the most competitive enrollment management tools in order to reach and admit qualified students. However, institutions that utilize transactional Web features are more effective if they can personalize services by providing useful customized information in real time for the prospect. Well crafted high…

  5. Using Forecasting to Predict Long-Term Resource Utilization for Web Services

    ERIC Educational Resources Information Center

    Yoas, Daniel W.

    2013-01-01

    Researchers have spent years understanding resource utilization to improve scheduling, load balancing, and system management through short-term prediction of resource utilization. Early research focused primarily on single operating systems; later, interest shifted to distributed systems and, finally, into web services. In each case researchers…

  6. Prototype of a Mobile Social Network for Education Using Dynamic Web Service

    ERIC Educational Resources Information Center

    Hoentsch, Sandra Costa Pinto; Carvalho, Felipe Oliveira; Santos, Luiz Marcus Monteiro de Almeida; Ribeiro, Admilson de Ribamar Lima

    2012-01-01

    This article presents the proposal of a social network site SocialNetLab that belongs to the Department of Computing-Federal University of Sergipe and which aims to locate and notify users of a nearby friend independently of the location technology available in the equipment through dynamic Web Service; to serve as a laboratory for research in…

  7. The Experiences of Older Students' Use of Web-Based Student Services

    ERIC Educational Resources Information Center

    Ho, Katy W.

    2012-01-01

    The purpose of this phenomenological case study was to understand the experiences of older students' use of web-based student services in a community college setting. For the purpose of this study the term "older student" was defined as people born between the years 1943 and 1960. This group of people, often described as the Baby…

  8. A Service Oriented Web Application for Learner Knowledge Representation, Management and Sharing Conforming to IMS LIP

    ERIC Educational Resources Information Center

    Lazarinis, Fotis

    2014-01-01

    iLM is a Web based application for representation, management and sharing of IMS LIP conformant user profiles. The tool is developed using a service oriented architecture with emphasis on the easy data sharing. Data elicitation from user profiles is based on the utilization of XQuery scripts and sharing with other applications is achieved through…

  9. A Semantic Web Service and Simulation Framework to Intelligent Distributed Manufacturing

    SciTech Connect

    Son, Young Jun; Kulvatunyou, Boonserm; Cho, Hyunbo; Feng, Shaw

    2005-11-01

    To cope with today's fluctuating markets, a virtual enterprise (VE) concept can be employed to achieve the cooperation among independently operating enterprises. The success of VE depends on reliable interoperation among trading partners. This paper proposes a framework based on semantic web of manufacturing and simulation services to enable business and engineering collaborations between VE partners, particularly a design house and manufacturing suppliers.

  10. NDU Knowledge Net: A Web-Enabled Just-In-Time Information Service for Continuing Education.

    ERIC Educational Resources Information Center

    Alden, Jay

    This paper describes the development of a web-enabled information service for constituents of the Information Resources Management College (National Defense University, Washington, DC). The constituents of the College, who include graduates, current students, and prospective students, typically work in the Chief Information Officer (CIO) office of…

  11. A framework for sharing and integrating remote sensing and GIS models based on Web service.

    PubMed

    Chen, Zeqiang; Lin, Hui; Chen, Min; Liu, Deer; Bao, Ying; Ding, Yulin

    2014-01-01

    Sharing and integrating Remote Sensing (RS) and Geographic Information System/Science (GIS) models are critical for developing practical application systems. Facilitating model sharing and model integration is a problem for model publishers and model users, respectively. To address this problem, a framework based on a Web service for sharing and integrating RS and GIS models is proposed in this paper. The fundamental idea of the framework is to publish heterogeneous RS and GIS models into standard Web services for sharing and interoperation and then to integrate the RS and GIS models using Web services. For the former, a "black box" and a visual method are employed to facilitate the publishing of the models as Web services. For the latter, model integration based on the geospatial workflow and semantic supported marching method is introduced. Under this framework, model sharing and integration is applied for developing the Pearl River Delta water environment monitoring system. The results show that the framework can facilitate model sharing and model integration for model publishers and model users.

  12. Effects of light reduction on food webs and associated ecosystem services of Yaquina Bay

    EPA Science Inventory

    Reduced water clarity can affect estuarine primary production but little is known of its subsequent effects to consumer guilds or ecosystem services. We investigated those effects using inverse analysis of modeled food webs of the lower (polyhaline) and upper (mesohaline) reache...

  13. SABinder: A Web Service for Predicting Streptavidin-Binding Peptides

    PubMed Central

    Kang, Juanjuan; Ru, Beibei; Zhou, Peng

    2016-01-01

    Streptavidin is sometimes used as the intended target to screen phage-displayed combinatorial peptide libraries for streptavidin-binding peptides (SBPs). More often in the biopanning system, however, streptavidin is just a commonly used anchoring molecule that can efficiently capture the biotinylated target. In this case, SBPs creeping into the biopanning results are not desired binders but target-unrelated peptides (TUP). Taking them as intended binders may mislead subsequent studies. Therefore, it is important to find if a peptide is likely to be an SBP when streptavidin is either the intended target or just the anchoring molecule. In this paper, we describe an SVM-based ensemble predictor called SABinder. It is the first predictor for SBP. The model was built with the feature of optimized dipeptide composition. It was observed that 89.20% (MCC = 0.78; AUC = 0.93; permutation test, p < 0.001) of peptides were correctly classified. As a web server, SABinder is freely accessible. The tool provides a highly efficient way to exclude potential SBP when they are TUP or to facilitate identification of possibly new SBP when they are the desired binders. In either case, it will be helpful and can benefit related scientific community. PMID:27610387

  14. SABinder: A Web Service for Predicting Streptavidin-Binding Peptides.

    PubMed

    He, Bifang; Kang, Juanjuan; Ru, Beibei; Ding, Hui; Zhou, Peng; Huang, Jian

    2016-01-01

    Streptavidin is sometimes used as the intended target to screen phage-displayed combinatorial peptide libraries for streptavidin-binding peptides (SBPs). More often in the biopanning system, however, streptavidin is just a commonly used anchoring molecule that can efficiently capture the biotinylated target. In this case, SBPs creeping into the biopanning results are not desired binders but target-unrelated peptides (TUP). Taking them as intended binders may mislead subsequent studies. Therefore, it is important to find if a peptide is likely to be an SBP when streptavidin is either the intended target or just the anchoring molecule. In this paper, we describe an SVM-based ensemble predictor called SABinder. It is the first predictor for SBP. The model was built with the feature of optimized dipeptide composition. It was observed that 89.20% (MCC = 0.78; AUC = 0.93; permutation test, p < 0.001) of peptides were correctly classified. As a web server, SABinder is freely accessible. The tool provides a highly efficient way to exclude potential SBP when they are TUP or to facilitate identification of possibly new SBP when they are the desired binders. In either case, it will be helpful and can benefit related scientific community. PMID:27610387

  15. Microarray oligonucleotide probe designer (MOPeD): A web service

    PubMed Central

    Patel, Viren C; Mondal, Kajari; Shetty, Amol Carl; Horner, Vanessa L; Bedoyan, Jirair K; Martin, Donna; Caspary, Tamara; Cutler, David J; Zwick, Michael E

    2011-01-01

    Methods of genomic selection that combine high-density oligonucleotide microarrays with next-generation DNA sequencing allow investigators to characterize genomic variation in selected portions of complex eukaryotic genomes. Yet choosing which specific oligonucleotides to be use can pose a major technical challenge. To address this issue, we have developed a software package called MOPeD (Microarray Oligonucleotide Probe Designer), which automates the process of designing genomic selection microarrays. This web-based software allows individual investigators to design custom genomic selection microarrays optimized for synthesis with Roche NimbleGen’s maskless photolithography. Design parameters include uniqueness of the probe sequences, melting temperature, hairpin formation, and the presence of single nucleotide polymorphisms. We generated probe databases for the human, mouse, and rhesus macaque genomes and conducted experimental validation of MOPeD-designed microarrays in human samples by sequencing the human X chromosome exome, where relevant sequence metrics indicated superior performance relative to a microarray designed by the Roche NimbleGen proprietary algorithm. We also performed validation in the mouse to identify known mutations contained within a 487-kb region from mouse chromosome 16, the mouse chromosome 16 exome (1.7 Mb), and the mouse chromosome 12 exome (3.3 Mb). Our results suggest that the open source MOPeD software package and website (http://moped.genetics.emory.edu/) will make a valuable resource for investigators in their sequence-based studies of complex eukaryotic genomes. PMID:21379402

  16. SABinder: A Web Service for Predicting Streptavidin-Binding Peptides

    PubMed Central

    Kang, Juanjuan; Ru, Beibei; Zhou, Peng

    2016-01-01

    Streptavidin is sometimes used as the intended target to screen phage-displayed combinatorial peptide libraries for streptavidin-binding peptides (SBPs). More often in the biopanning system, however, streptavidin is just a commonly used anchoring molecule that can efficiently capture the biotinylated target. In this case, SBPs creeping into the biopanning results are not desired binders but target-unrelated peptides (TUP). Taking them as intended binders may mislead subsequent studies. Therefore, it is important to find if a peptide is likely to be an SBP when streptavidin is either the intended target or just the anchoring molecule. In this paper, we describe an SVM-based ensemble predictor called SABinder. It is the first predictor for SBP. The model was built with the feature of optimized dipeptide composition. It was observed that 89.20% (MCC = 0.78; AUC = 0.93; permutation test, p < 0.001) of peptides were correctly classified. As a web server, SABinder is freely accessible. The tool provides a highly efficient way to exclude potential SBP when they are TUP or to facilitate identification of possibly new SBP when they are the desired binders. In either case, it will be helpful and can benefit related scientific community.

  17. Issues in implementing services for a wireless web-enabled digital camera

    NASA Astrophysics Data System (ADS)

    Venkataraman, Shyam; Sampat, Nitin; Fisher, Yoram; Canosa, John; Noel, Nicholas

    2001-05-01

    The competition in the exploding digital photography market has caused vendors to explore new ways to increase their return on investment. A common view among industry analysts is that increasingly it will be services provided by these cameras, and not the cameras themselves, that will provide the revenue stream. These services will be coupled to e- Appliance based Communities. In addition, the rapidly increasing need to upload images to the Internet for photo- finishing services as well as the need to download software upgrades to the camera is driving many camera OEMs to evaluate the benefits of using the wireless web to extend their enterprise systems. Currently, creating a viable e- appliance such as a digital camera coupled with a wireless web service requires more than just a competency in product development. This paper will evaluate the system implications in the deployment of recurring revenue services and enterprise connectivity of a wireless, web-enabled digital camera. These include, among other things, an architectural design approach for services such as device management, synchronization, billing, connectivity, security, etc. Such an evaluation will assist, we hope, anyone designing or connecting a digital camera to the enterprise systems.

  18. UltiMatch-NL: A Web Service Matchmaker Based on Multiple Semantic Filters

    PubMed Central

    Mohebbi, Keyvan; Ibrahim, Suhaimi; Zamani, Mazdak; Khezrian, Mojtaba

    2014-01-01

    In this paper, a Semantic Web service matchmaker called UltiMatch-NL is presented. UltiMatch-NL applies two filters namely Signature-based and Description-based on different abstraction levels of a service profile to achieve more accurate results. More specifically, the proposed filters rely on semantic knowledge to extract the similarity between a given pair of service descriptions. Thus it is a further step towards fully automated Web service discovery via making this process more semantic-aware. In addition, a new technique is proposed to weight and combine the results of different filters of UltiMatch-NL, automatically. Moreover, an innovative approach is introduced to predict the relevance of requests and Web services and eliminate the need for setting a threshold value of similarity. In order to evaluate UltiMatch-NL, the repository of OWLS-TC is used. The performance evaluation based on standard measures from the information retrieval field shows that semantic matching of OWL-S services can be significantly improved by incorporating designed matching filters. PMID:25157872

  19. UltiMatch-NL: a Web service matchmaker based on multiple semantic filters.

    PubMed

    Mohebbi, Keyvan; Ibrahim, Suhaimi; Zamani, Mazdak; Khezrian, Mojtaba

    2014-01-01

    In this paper, a Semantic Web service matchmaker called UltiMatch-NL is presented. UltiMatch-NL applies two filters namely Signature-based and Description-based on different abstraction levels of a service profile to achieve more accurate results. More specifically, the proposed filters rely on semantic knowledge to extract the similarity between a given pair of service descriptions. Thus it is a further step towards fully automated Web service discovery via making this process more semantic-aware. In addition, a new technique is proposed to weight and combine the results of different filters of UltiMatch-NL, automatically. Moreover, an innovative approach is introduced to predict the relevance of requests and Web services and eliminate the need for setting a threshold value of similarity. In order to evaluate UltiMatch-NL, the repository of OWLS-TC is used. The performance evaluation based on standard measures from the information retrieval field shows that semantic matching of OWL-S services can be significantly improved by incorporating designed matching filters.

  20. UltiMatch-NL: a Web service matchmaker based on multiple semantic filters.

    PubMed

    Mohebbi, Keyvan; Ibrahim, Suhaimi; Zamani, Mazdak; Khezrian, Mojtaba

    2014-01-01

    In this paper, a Semantic Web service matchmaker called UltiMatch-NL is presented. UltiMatch-NL applies two filters namely Signature-based and Description-based on different abstraction levels of a service profile to achieve more accurate results. More specifically, the proposed filters rely on semantic knowledge to extract the similarity between a given pair of service descriptions. Thus it is a further step towards fully automated Web service discovery via making this process more semantic-aware. In addition, a new technique is proposed to weight and combine the results of different filters of UltiMatch-NL, automatically. Moreover, an innovative approach is introduced to predict the relevance of requests and Web services and eliminate the need for setting a threshold value of similarity. In order to evaluate UltiMatch-NL, the repository of OWLS-TC is used. The performance evaluation based on standard measures from the information retrieval field shows that semantic matching of OWL-S services can be significantly improved by incorporating designed matching filters. PMID:25157872

  1. Framework for ReSTful Web Services in OSGi

    NASA Technical Reports Server (NTRS)

    Shams, Khawaja S.; Norris, Jeffrey S.; Powell, Mark W.; Crockett, Thomas M.; Mittman, David S.; Fox, Jason M.; Joswig, Joseph C.; Wallick, Michael N.; Torres, Recaredo J.; Rabe, Kenneth

    2009-01-01

    Ensemble ReST is a software system that eases the development, deployment, and maintenance of server-side application programs to perform functions that would otherwise be performed by client software. Ensemble ReST takes advantage of the proven disciplines of ReST (Representational State Transfer. ReST leverages the standardized HTTP protocol to enable developers to offer services to a diverse variety of clients: from shell scripts to sophisticated Java application suites

  2. Web-Based Self-Service Systems for Managed IT Support: Service Provider Perspectives of Stakeholder-Based Issues

    NASA Astrophysics Data System (ADS)

    Cooper, Vanessa A.; Lichtenstein, Sharman; Smith, Ross

    This chapter explores the provision of after-sales information technology (IT) support services using Web-based self-service systems (WSSs) in a business-to-business (B2B) context. A recent study conducted at six large multi-national IT support organisations revealed a number of critical success factors (CSFs) and stakeholder-based issues. To better identify and understand these important enablers and barriers, we explain how WSSs should be considered within a complex network of service providers, business partners and customer firms. The CSFs and stakeholder-based issues are discussed. The chapter highlights that for more successful service provision using WSSs, IT service providers should collaborate more effectively with enterprise customers and business partners and should better integrate their WSSs.

  3. Analysing the primacy of distance in the utilization of health services in the Ahafo-Ano South district, Ghana.

    PubMed

    Buor, Daniel

    2003-01-01

    Although the distance factor has been identified as key in the utilization of health services in rural areas of developing countries, it has been analysed without recourse to related factors of travel time and transport cost. Also, the influence of distance on vulnerable groups in utilization has not been an object of survey by researchers. This paper addresses the impact of distance on utilization, and how distance compares with travel time and transport cost that are related to it in the utilization of health services in the Ahafo-Ano South (rural) district in Ghana. The study, a cross-sectional survey, also identifies the position of distance among other important factors of utilization. A sample of 400, drawn through systematic random technique, was used for the survey. Data were analysed using the regression model and some graphic techniques. The main instruments used in data collection were formal (face-by-face) interview and a questionnaire. The survey finds that distance is the most important factor that influences the utilization of health services in the Ahafo-Ano South district. Other key factors are income, service cost and education. The effect of travel time on utilization reflects that of distance and utilization. Recommendations to reduce distance coverage, improve formal education and reduce poverty have been made.

  4. Using Transactional Distance Theory to Redesign an Online Mathematics Education Course for Pre-Service Primary Teachers

    ERIC Educational Resources Information Center

    Larkin, Kevin; Jamieson-Proctor, Romina

    2015-01-01

    This paper examines the impact of a series of design changes to an online mathematics education course in terms of transactional distance between learner and teachers, pre-service education students' attitudes towards mathematics, and their development of mathematical pedagogical knowledge. Transactional distance theory (TDT) was utilised to…

  5. Package research based on COM+ technical of web service

    NASA Astrophysics Data System (ADS)

    Liu, Xiaoyue; Du, Liang

    2011-10-01

    Early information management systems are established for realization of specific business functions ordered by enterprise, including the control and management of production, supply, sales and customer service and so on. Enterprises in accordance with their various departments establish a variety of enterprise information systems used by them, and in the actual work play a huge role in the production. In this paper, from the overall view a variety of enterprise information management systems is considered, thereafter according to specific business needs to unify arrangement of these systems, and use a suitable platform for the organic integration of existing enterprise information management systems, so that the "information islands" link.

  6. Pilot evaluation of a web-based intervention targeting sexual health service access.

    PubMed

    Brown, K E; Newby, K; Caley, M; Danahay, A; Kehal, I

    2016-04-01

    Sexual health service access is fundamental to good sexual health, yet interventions designed to address this have rarely been implemented or evaluated. In this article, pilot evaluation findings for a targeted public health behavior change intervention, delivered via a website and web-app, aiming to increase uptake of sexual health services among 13-19-year olds are reported. A pre-post questionnaire-based design was used. Matched baseline and follow-up data were identified from 148 respondents aged 13-18 years. Outcome measures were self-reported service access, self-reported intention to access services and beliefs about services and service access identified through needs analysis. Objective service access data provided by local sexual health services were also analyzed. Analysis suggests the intervention had a significant positive effect on psychological barriers to and antecedents of service access among females. Males, who reported greater confidence in service access compared with females, significantly increased service access by time 2 follow-up. Available objective service access data support the assertion that the intervention may have led to increases in service access. There is real promise for this novel digital intervention. Further evaluation is planned as the model is licensed to and rolled out by other local authorities in the United Kingdom. PMID:26928566

  7. Technology Development, Implementation and Assessment: K-16 Pre-Service, In-Service and Distance Learning Initiatives

    NASA Technical Reports Server (NTRS)

    Williams, William B., Jr.

    1999-01-01

    The technologies associated with distance learning are evolving rapidly, giving to educators a potential tool for enhancing the educational experiences of large numbers of students simultaneously. This enhancement, in order to be effective, must take into account the various agendas of teachers, administrators, state systems, and of course students. It must also make use of the latest research on effective pedagogy. This combination, effective pedagogy and robust information technology, is a powerful vehicle for communicating, to a large audience of school children the excitement of mathematics and science--an excitement that for the most part is now well-hidden. This project,"Technology Development, Implementation and Assessment," proposed to bring to bear on the education of learners in grades 3 - 8 in science and mathematics both advances in information technology and in effective pedagogy. Specifically, the project developed components NASA CONNECT video series--problem-based learning modules that focus on the scientific method and that incorporate problem-based learning scenarios tied to national mathematics and science standards. These videos serve two purposes; they engage students in the excitement of hands-on learning and they model for the teachers of these students the problem-based learning practices that are proving to be excellent ways to teach science and mathematics to school students. Another component of NASA CONNECT is the accompanying web-site.

  8. Experimental evaluation of the impact of packet capturing tools for web services.

    SciTech Connect

    Choe, Yung Ryn; Mohapatra, Prasant; Chuah, Chen-Nee; Chen, Chao-Chih

    2011-10-01

    Network measurement is a discipline that provides the techniques to collect data that are fundamental to many branches of computer science. While many capturing tools and comparisons have made available in the literature and elsewhere, the impact of these packet capturing tools on existing processes have not been thoroughly studied. While not a concern for collection methods in which dedicated servers are used, many usage scenarios of packet capturing now requires the packet capturing tool to run concurrently with operational processes. In this work we perform experimental evaluations of the performance impact that packet capturing process have on web-based services; in particular, we observe the impact on web servers. We find that packet capturing processes indeed impact the performance of web servers, but on a multi-core system the impact varies depending on whether the packet capturing and web hosting processes are co-located or not. In addition, the architecture and behavior of the web server and process scheduling is coupled with the behavior of the packet capturing process, which in turn also affect the web server's performance.

  9. Development of a Dynamic Web Mapping Service for Vegetation Productivity Using Earth Observation and in situ Sensors in a Sensor Web Based Approach.

    PubMed

    Kooistra, Lammert; Bergsma, Aldo; Chuma, Beatus; de Bruin, Sytze

    2009-01-01

    This paper describes the development of a sensor web based approach which combines earth observation and in situ sensor data to derive typical information offered by a dynamic web mapping service (WMS). A prototype has been developed which provides daily maps of vegetation productivity for the Netherlands with a spatial resolution of 250 m. Daily available MODIS surface reflectance products and meteorological parameters obtained through a Sensor Observation Service (SOS) were used as input for a vegetation productivity model. This paper presents the vegetation productivity model, the sensor data sources and the implementation of the automated processing facility. Finally, an evaluation is made of the opportunities and limitations of sensor web based approaches for the development of web services which combine both satellite and in situ sensor sources.

  10. Development of a Dynamic Web Mapping Service for Vegetation Productivity Using Earth Observation and in situ Sensors in a Sensor Web Based Approach

    PubMed Central

    Kooistra, Lammert; Bergsma, Aldo; Chuma, Beatus; de Bruin, Sytze

    2009-01-01

    This paper describes the development of a sensor web based approach which combines earth observation and in situ sensor data to derive typical information offered by a dynamic web mapping service (WMS). A prototype has been developed which provides daily maps of vegetation productivity for the Netherlands with a spatial resolution of 250 m. Daily available MODIS surface reflectance products and meteorological parameters obtained through a Sensor Observation Service (SOS) were used as input for a vegetation productivity model. This paper presents the vegetation productivity model, the sensor data sources and the implementation of the automated processing facility. Finally, an evaluation is made of the opportunities and limitations of sensor web based approaches for the development of web services which combine both satellite and in situ sensor sources. PMID:22574019

  11. Composition of Web Services Using Markov Decision Processes and Dynamic Programming

    PubMed Central

    Uc-Cetina, Víctor; Moo-Mena, Francisco; Hernandez-Ucan, Rafael

    2015-01-01

    We propose a Markov decision process model for solving the Web service composition (WSC) problem. Iterative policy evaluation, value iteration, and policy iteration algorithms are used to experimentally validate our approach, with artificial and real data. The experimental results show the reliability of the model and the methods employed, with policy iteration being the best one in terms of the minimum number of iterations needed to estimate an optimal policy, with the highest Quality of Service attributes. Our experimental work shows how the solution of a WSC problem involving a set of 100,000 individual Web services and where a valid composition requiring the selection of 1,000 services from the available set can be computed in the worst case in less than 200 seconds, using an Intel Core i5 computer with 6 GB RAM. Moreover, a real WSC problem involving only 7 individual Web services requires less than 0.08 seconds, using the same computational power. Finally, a comparison with two popular reinforcement learning algorithms, sarsa and Q-learning, shows that these algorithms require one or two orders of magnitude and more time than policy iteration, iterative policy evaluation, and value iteration to handle WSC problems of the same complexity. PMID:25874247

  12. Business Models of E-Government: Research on Dynamic E-Government Based on Web Services

    NASA Astrophysics Data System (ADS)

    Li, Yan; Yang, Jiumin

    Government transcends all sectors in a society. It provides not only the legal, political and economic infrastructure to support other sectors, but also exerts significant influence on the social factors that contribute to their development. With its maturity of technologies and management, e-government will eventually enter into the time of 'one-stop' services. Among others, the technology of Web services is the major contributor to this achievement. Web services provides a new way of standard-based software technology, letting programmers combine existing computer system in new ways over the Internet within one business or across many, and would thereby bring about profound and far-reaching impacts on e-government. This paper introduced the business modes of e-government, architecture of dynamic e-government and its key technologies. Finally future prospect of dynamic e-government was also briefly discussed.

  13. A probabilistic sampling method (PSM) for estimating geographic distance to health services when only the region of residence is known

    PubMed Central

    2011-01-01

    Background The need to estimate the distance from an individual to a service provider is common in public health research. However, estimated distances are often imprecise and, we suspect, biased due to a lack of specific residential location data. In many cases, to protect subject confidentiality, data sets contain only a ZIP Code or a county. Results This paper describes an algorithm, known as "the probabilistic sampling method" (PSM), which was used to create a distribution of estimated distances to a health facility for a person whose region of residence was known, but for which demographic details and centroids were known for smaller areas within the region. From this distribution, the median distance is the most likely distance to the facility. The algorithm, using Monte Carlo sampling methods, drew a probabilistic sample of all the smaller areas (Census blocks) within each participant's reported region (ZIP Code), weighting these areas by the number of residents in the same age group as the participant. To test the PSM, we used data from a large cross-sectional study that screened women at a clinic for intimate partner violence (IPV). We had data on each woman's age and ZIP Code, but no precise residential address. We used the PSM to select a sample of census blocks, then calculated network distances from each census block's centroid to the closest IPV facility, resulting in a distribution of distances from these locations to the geocoded locations of known IPV services. We selected the median distance as the most likely distance traveled and computed confidence intervals that describe the shortest and longest distance within which any given percent of the distance estimates lie. We compared our results to those obtained using two other geocoding approaches. We show that one method overestimated the most likely distance and the other underestimated it. Neither of the alternative methods produced confidence intervals for the distance estimates. The algorithm

  14. MALINA: a web service for visual analytics of human gut microbiota whole-genome metagenomic reads.

    PubMed

    Tyakht, Alexander V; Popenko, Anna S; Belenikin, Maxim S; Altukhov, Ilya A; Pavlenko, Alexander V; Kostryukova, Elena S; Selezneva, Oksana V; Larin, Andrei K; Karpova, Irina Y; Alexeev, Dmitry G

    2012-01-01

    MALINA is a web service for bioinformatic analysis of whole-genome metagenomic data obtained from human gut microbiota sequencing. As input data, it accepts metagenomic reads of various sequencing technologies, including long reads (such as Sanger and 454 sequencing) and next-generation (including SOLiD and Illumina). It is the first metagenomic web service that is capable of processing SOLiD color-space reads, to authors' knowledge. The web service allows phylogenetic and functional profiling of metagenomic samples using coverage depth resulting from the alignment of the reads to the catalogue of reference sequences which are built into the pipeline and contain prevalent microbial genomes and genes of human gut microbiota. The obtained metagenomic composition vectors are processed by the statistical analysis and visualization module containing methods for clustering, dimension reduction and group comparison. Additionally, the MALINA database includes vectors of bacterial and functional composition for human gut microbiota samples from a large number of existing studies allowing their comparative analysis together with user samples, namely datasets from Russian Metagenome project, MetaHIT and Human Microbiome Project (downloaded from http://hmpdacc.org). MALINA is made freely available on the web at http://malina.metagenome.ru. The website is implemented in JavaScript (using Ext JS), Microsoft .NET Framework, MS SQL, Python, with all major browsers supported.

  15. Development and process evaluation of a web-based responsible beverage service training program

    PubMed Central

    2012-01-01

    Background Responsible beverage service (RBS) training designed to improve the appropriate service of alcohol in commercial establishments is typically delivered in workshops. Recently, Web-based RBS training programs have emerged. This report describes the formative development and subsequent design of an innovative Web-delivered RBS program, and evaluation of the impact of the program on servers’ knowledge, attitudes, and self-efficacy. Methods Formative procedures using focus groups and usability testing were used to develop a Web-based RBS training program. Professional alcohol servers (N = 112) who worked as servers and/or mangers in alcohol service settings were recruited to participate. A pre-post assessment design was used to assess changes associated with using the program. Results Participants who used the program showed significant improvements in their RBS knowledge, attitudes, and self-efficacy. Conclusions Although the current study did not directly observe and determine impact of the intervention on server behaviors, it demonstrated that the development process incorporating input from a multidisciplinary team in conjunction with feedback from end-users resulted in creation of a Web-based RBS program that was well-received by servers and that changed relevant knowledge, attitudes, and self-efficacy. The results also help to establish a needed evidence base in support of the use of online RBS training, which has been afforded little research attention. PMID:22999419

  16. OneGeology-Europe: architecture, portal and web services to provide a European geological map

    NASA Astrophysics Data System (ADS)

    Tellez-Arenas, Agnès.; Serrano, Jean-Jacques; Tertre, François; Laxton, John

    2010-05-01

    OneGeology-Europe is a large ambitious project to make geological spatial data further known and accessible. The OneGeology-Europe project develops an integrated system of data to create and make accessible for the first time through the internet the geological map of the whole of Europe. The architecture implemented by the project is web services oriented, based on the OGC standards: the geological map is not a centralized database but is composed by several web services, each of them hosted by a European country involved in the project. Since geological data are elaborated differently from country to country, they are difficult to share. OneGeology-Europe, while providing more detailed and complete information, will foster even beyond the geological community an easier exchange of data within Europe and globally. This implies an important work regarding the harmonization of the data, both model and the content. OneGeology-Europe is characterised by the high technological capacity of the EU Member States, and has the final goal to achieve the harmonisation of European geological survey data according to common standards. As a direct consequence Europe will make a further step in terms of innovation and information dissemination, continuing to play a world leading role in the development of geosciences information. The scope of the common harmonized data model was defined primarily by the requirements of the geological map of Europe, but in addition users were consulted and the requirements of both INSPIRE and ‘high-resolution' geological maps were considered. The data model is based on GeoSciML, developed since 2006 by a group of Geological Surveys. The data providers involved in the project implemented a new component that allows the web services to deliver the geological map expressed into GeoSciML. In order to capture the information describing the geological units of the map of Europe the scope of the data model needs to include lithology; age; genesis and

  17. Research on sudden environmental pollution public service platform construction based on WebGIS

    NASA Astrophysics Data System (ADS)

    Bi, T. P.; Gao, D. Y.; Zhong, X. Y.

    2016-08-01

    In order to actualize the social sharing and service of the emergency-response information for sudden pollution accidents, the public can share the risk source information service, dangerous goods control technology service and so on, The SQL Server and ArcSDE software are used to establish a spatial database to restore all kinds of information including risk sources, hazardous chemicals and handling methods in case of accidents. Combined with Chinese atmospheric environmental assessment standards, the SCREEN3 atmospheric dispersion model and one-dimensional liquid diffusion model are established to realize the query of related information and the display of the diffusion effect under B/S structure. Based on the WebGIS technology, C#.Net language is used to develop the sudden environmental pollution public service platform. As a result, the public service platform can make risk assessments and provide the best emergency processing services.

  18. HydroViewer: Utilizing Web-Based Hydrologic Data And Analytical Services

    NASA Astrophysics Data System (ADS)

    Ye, Z.; Djokic, D.; Armstrong, L.

    2011-12-01

    To conduct a hydrologic study in an area, the hydrologist needs to define the area of interest, collect the spatial, hydrological, and metrological data for the area, and finally perform the desired analysis. Services oriented architecture holds promise that these activities can be done using distributed services combined in a single lightweight application. Hydrologic time series data are collected, stored, and served by multiple agencies in different formats and are often difficult to find, acquire, and mobilize for analysis. CUAHSI's WaterOneFlow web service API provides functions for querying and collecting the temporal data in a consistent manner. Many agencies and Universities publish their time series data using WaterOneFlow services and thus make them available to a broad range of users. ArcGIS server allows publishing of spatial data and mapping services and is widely used in the government, academia, and industry. ArcGIS server can also be used to serve analytical services. By combining spatial services provided by ArcGIS server and temporal data provided through WaterOneFlow services, it is now possible to create web applications to explore the hydrologic time series data available in a given spatial area served by multiple agencies and perform analysis on them. A Web application, HydroViewer, is developed using ArcGIS Silverlight API to allow users to mobilize ArcGIS server and WaterOneFlow services in an integrated fashion. It performs the following tasks: (1) Delineating watershed for a user specified point of interest using Arc Hydro based watershed delineation service, (2) Exploring data collecting sites and data collected and served by different agencies for a given spatial area (watershed or viewing spatial extent) and time domain, (3) Viewing/graphing these data collected by these sites, (4) Collecting the metadata in for the data variables in a form of a data cart, (5) Downloading both spatial and time series data to create an Arc Hydro

  19. Increasing access to terrestrial ecology and remote sensing (MODIS) data through Web services and visualization tools

    NASA Astrophysics Data System (ADS)

    Santhana Vannan, S.; Cook, R. B.; Wei, Y.

    2012-12-01

    In recent years user access to data and information is increasingly handled through tools, services, and applications. Standards-based services have facilitated this development. These service-based methods to access data has boosted the use of data and in increasingly complex ways. The Oak Ridge National Laboratory Distributed Active Archive Center (ORNL DAAC) has taken the approach of service-based access to data and visualization for distribution and visualization of its terrestrial ecology data, including MODIS (Moderate Resolution Imaging Spectroradiometer) remote sensing data products. The MODIS data products are highly useful for field research. The spectral, spatial and temporal characteristics of MODIS products have made them an important data source for analyzing key science questions relating to Earth system processes at multiple spatial and temporal scales. However, MODIS data volume and the complexity in data format make it less usable in some cases. To solve this usability issue, the ORNL DAAC has developed a system that prepares and distributes subsets of selected MODIS land products in a scale and format useful for field researchers. Web and Web service tools provide MODIS subsets in comma-delimited text format and in GIS compatible GeoTIFF format. Users can download and visualize MODIS subsets for a set of pre-defined locations, order MODIS subsets for any land location or automate the process of subset extraction using a SOAP-based Web service. The MODIS tools and services can be extended to support the large volume of data that would be produced by the various decadal survey missions. http://daac.ornl.gov/MODIS . The ORNL DAAC has also created a Web-based Spatial Data Access Tool (SDAT) that enables users to browse, visualize, and download a wide variety of geospatial data in various user-selected spatial/temporal extents, formats, and projections. SDAT is based on Open Geospatial Consortium (OGC) Web service standards that allows users to

  20. Visualization of cloud physical properties derived from Meteosat using OGC Web Map Services.

    NASA Astrophysics Data System (ADS)

    Plieger, M.; Vegte, J. vd; Meirink, J. F.; Roebeling, R.

    2012-04-01

    The Cloud Physical Properties (CPP) algorithm is being developed at KNMI to derive cloud, precipitation and radiation products from satellite instruments such as SEVIRI onboard the METEOSAT Second Generation (MSG). The CPP algorithm development is largely done inside EUMETSAT's Climate Monitoring Satellite Application Facility (CM-SAF), but also in research projects like EURO4M, SYNTHESIS, MSGSIM, and SICCS. In the MSG-CPP project, data is converted to the network common data format (NetCDF) with Climate and Forecast conventions (CF conventions). Data can be accessed by using web services defined by the Open Geospatial Consortium (OGC). Web Map Services are offered for data visualization and Web Coverage Services are offered for data conversion and access (WMS & WCS). The OGC services can be accessed in the ADAGUC portal, a JavaScript based web tool which runs in most modern browsers. The portal allows viewing, zooming, querying and downloading of the data. Interesting situations can be captured and sent to others by creating status hyperlinks, and the data can be displayed in Google Earth if desired. The high resolution of the images and the large data volumes are challenging for real-time and web based visualization services. During the MSG-CPP project the ADAGUC OGC server has been optimized to enable quick rendering of visualizations from NetCDF data, the server reads the NetCDF4 files directly and interpolates the data to the desired geographic projection and location. A new feature of the server is the support for NetCDF flag values and flag meanings as specified in the Climate and Forecast conventions. Flag values and flag meanings allow for visualization of data with discrete values. This feature is used in the visualization of thermodynamic phase of particles at the top of the cloud, which values are indicated with meanings like no_cloud, liquid, ice and mixed. The legend of the WMS service displays these meanings individually and the getfeatureinfo

  1. The Taverna workflow suite: designing and executing workflows of Web Services on the desktop, web or in the cloud.

    PubMed

    Wolstencroft, Katherine; Haines, Robert; Fellows, Donal; Williams, Alan; Withers, David; Owen, Stuart; Soiland-Reyes, Stian; Dunlop, Ian; Nenadic, Aleksandra; Fisher, Paul; Bhagat, Jiten; Belhajjame, Khalid; Bacall, Finn; Hardisty, Alex; Nieva de la Hidalga, Abraham; Balcazar Vargas, Maria P; Sufi, Shoaib; Goble, Carole

    2013-07-01

    The Taverna workflow tool suite (http://www.taverna.org.uk) is designed to combine distributed Web Services and/or local tools into complex analysis pipelines. These pipelines can be executed on local desktop machines or through larger infrastructure (such as supercomputers, Grids or cloud environments), using the Taverna Server. In bioinformatics, Taverna workflows are typically used in the areas of high-throughput omics analyses (for example, proteomics or transcriptomics), or for evidence gathering methods involving text mining or data mining. Through Taverna, scientists have access to several thousand different tools and resources that are freely available from a large range of life science institutions. Once constructed, the workflows are reusable, executable bioinformatics protocols that can be shared, reused and repurposed. A repository of public workflows is available at http://www.myexperiment.org. This article provides an update to the Taverna tool suite, highlighting new features and developments in the workbench and the Taverna Server. PMID:23640334

  2. The Taverna workflow suite: designing and executing workflows of Web Services on the desktop, web or in the cloud.

    PubMed

    Wolstencroft, Katherine; Haines, Robert; Fellows, Donal; Williams, Alan; Withers, David; Owen, Stuart; Soiland-Reyes, Stian; Dunlop, Ian; Nenadic, Aleksandra; Fisher, Paul; Bhagat, Jiten; Belhajjame, Khalid; Bacall, Finn; Hardisty, Alex; Nieva de la Hidalga, Abraham; Balcazar Vargas, Maria P; Sufi, Shoaib; Goble, Carole

    2013-07-01

    The Taverna workflow tool suite (http://www.taverna.org.uk) is designed to combine distributed Web Services and/or local tools into complex analysis pipelines. These pipelines can be executed on local desktop machines or through larger infrastructure (such as supercomputers, Grids or cloud environments), using the Taverna Server. In bioinformatics, Taverna workflows are typically used in the areas of high-throughput omics analyses (for example, proteomics or transcriptomics), or for evidence gathering methods involving text mining or data mining. Through Taverna, scientists have access to several thousand different tools and resources that are freely available from a large range of life science institutions. Once constructed, the workflows are reusable, executable bioinformatics protocols that can be shared, reused and repurposed. A repository of public workflows is available at http://www.myexperiment.org. This article provides an update to the Taverna tool suite, highlighting new features and developments in the workbench and the Taverna Server.

  3. Service Quality and Students' Satisfaction with the Professional Teacher Development Programmes by Distance Mode in a South African University

    ERIC Educational Resources Information Center

    Oduaran, A. B.

    2011-01-01

    This article reports on the relationship between seven factors that described dimensions of education service quality and overall service quality on one hand, and students' satisfaction with the professional teacher development programmes by distance mode in a South African University on the other. We sought to find out whether students enrolled…

  4. DL-sQUAL: A Multiple-Item Scale for Measuring Service Quality of Online Distance Learning Programs

    ERIC Educational Resources Information Center

    Shaik, Naj; Lowe, Sue; Pinegar, Kem

    2006-01-01

    Education is a service with multiplicity of student interactions over time and across multiple touch points. Quality teaching needs to be supplemented by consistent quality supporting services for programs to succeed under the competitive distance learning landscape. ServQual and e-SQ scales have been proposed for measuring quality of traditional…

  5. A Mediator-Based Approach to Resolving Interface Heterogeneity of Web Services

    NASA Astrophysics Data System (ADS)

    Leitner, Philipp; Rosenberg, Florian; Michlmayr, Anton; Huber, Andreas; Dustdar, Schahram

    In theory, service-oriented architectures are based on the idea of increasing flexibility in the selection of internal and external business partners using loosely-coupled services. However, in practice this flexibility is limited by the fact that partners need not only to provide the same service, but to do so via virtually the same interface in order to actually be interchangeable easily. Invocation-level mediation may be used to overcome this issue — by using mediation interface differences can be resolved transparently at runtime. In this chapter we discuss the basic ideas of mediation, with a focus on interface-level mediation. We show how interface mediation is integrated into our dynamic Web service invocation framework DAIOS, and present three different mediation strategies, one based on structural message similarity, one based on semantically annotated WSDL, and one which is embedded into the VRESCo SOA runtime, a larger research project with explicit support for service mediation.

  6. Mining multilevel and location-aware service patterns in mobile web environments.

    PubMed

    Tseng, Shin-Mu; Tsui, Ching-Fu

    2004-12-01

    In this correspondence, we address the issue of efficiently mining multilevel and location-aware associated service patterns in a mobile web environment. In terms of multilevel concept, we consider the complex problem that locations and services are of hierarchical structures. We propose a new data mining method named two-dimensional multilevel (2-DML) association rules mining, which can efficiently discover the associated service request patterns by taking into account the multilevel properties of locations and services. The discovered patterns can be effectively utilized in real applications like location-based and personalized services. To the best of our knowledge, this is the first work addressing this research issue. Some variations of the 2-DML method with different properties in terms of execution efficiency and memory efficiency were also developed. Through empirical evaluation, the proposed methods are shown to deliver good performance in terms of efficiency and scalability under various system conditions.

  7. BioXSD: the common data-exchange format for everyday bioinformatics web services

    PubMed Central

    Kalaš, Matúš; Puntervoll, Pæl; Joseph, Alexandre; Bartaševičiūtė, Edita; Töpfer, Armin; Venkataraman, Prabakar; Pettifer, Steve; Bryne, Jan Christian; Ison, Jon; Blanchet, Christophe; Rapacki, Kristoffer; Jonassen, Inge

    2010-01-01

    Motivation: The world-wide community of life scientists has access to a large number of public bioinformatics databases and tools, which are developed and deployed using diverse technologies and designs. More and more of the resources offer programmatic web-service interface. However, efficient use of the resources is hampered by the lack of widely used, standard data-exchange formats for the basic, everyday bioinformatics data types. Results: BioXSD has been developed as a candidate for standard, canonical exchange format for basic bioinformatics data. BioXSD is represented by a dedicated XML Schema and defines syntax for biological sequences, sequence annotations, alignments and references to resources. We have adapted a set of web services to use BioXSD as the input and output format, and implemented a test-case workflow. This demonstrates that the approach is feasible and provides smooth interoperability. Semantics for BioXSD is provided by annotation with the EDAM ontology. We discuss in a separate section how BioXSD relates to other initiatives and approaches, including existing standards and the Semantic Web. Availability: The BioXSD 1.0 XML Schema is freely available at http://www.bioxsd.org/BioXSD-1.0.xsd under the Creative Commons BY-ND 3.0 license. The http://bioxsd.org web page offers documentation, examples of data in BioXSD format, example workflows with source codes in common programming languages, an updated list of compatible web services and tools and a repository of feature requests from the community. Contact: matus.kalas@bccs.uib.no; developers@bioxsd.org; support@bioxsd.org PMID:20823319

  8. [Building a virtual environment for distance learning: an in-service educational strategy].

    PubMed

    Grossi, Manoela Gomes; Kobayashi, Rika Miyahara

    2013-06-01

    This study aims to describe the construction of a virtual learning environment (VLE) in a social network for implementing distance learning (DL), developed in a public cardiology hospital by 23 nurses from the Education Group. The construction and implementation were carried out at the workplace, following the structuring phases of the education, development, and evaluation of the VLE for the DL Group from the perspectives of tutors and students. The learningand development of technological knowledge were found to occur alongside an increase in the knowledge of how to constructand utilize a VLE. The difficulties encountered were related to a lack of expertise, time, and infrastructure. Limitations relating to the required tools and internet access were also identified. To make the project possible, the nurses developed up-to-date skills, technological expertise, and creativity, as well as the ability to search for alternative resources to overcome structural difficulties, build team skills, and implement in-service innovative educational processes.

  9. An efficient and flexible web services-based multidisciplinary design optimisation framework for complex engineering systems

    NASA Astrophysics Data System (ADS)

    Li, Liansheng; Liu, Jihong

    2012-08-01

    Multidisciplinary design optimisation (MDO) involves multiple disciplines, multiple coupled relationships and multiple processes, which is implemented by different specialists dispersed geographically on heterogeneous platforms with different analysis and optimisation tools. The product design data integration and data sharing among the participants hampers the development and applications of MDO in enterprises seriously. Therefore, a multi-hierarchical integrated product design data model (MH-iPDM) supporting the MDO in the web environment and a web services-based multidisciplinary design optimisation (Web-MDO) framework are proposed in this article. Based on the enabling technologies including web services, ontology, workflow, agent, XML and evidence theory, the proposed framework enables the designers geographically dispersed to work collaboratively in the MDO environment. The ontology-based workflow enables the logical reasoning of MDO to be processed dynamically. The evidence theory-based uncertainty reasoning and analysis supports the quantification, aggregation and analysis of the conflicting epistemic uncertainty from multiple sources, which improves the quality of product. Finally, a proof-of-concept prototype system is developed using J2EE and an example of supersonic business jet is demonstrated to verify the autonomous execution of MDO strategies and the effectiveness of the proposed approach.

  10. A Web Service Model for Providing Weather Information through Sensor Networks Using a Fermat Point Based Data Forwarding Scheme

    NASA Astrophysics Data System (ADS)

    Ghosh, Kaushik; Rawat, Manoj; Das, Pradip K.

    2010-11-01

    Web services providing weather information are not new. The existing web services working on this kind of fields can provide more précised information if the concerned data is collected in a distributed fashion using a sensor network. Longer the lifetime of the sensor network, longer is the service provided without interruption. In this paper we propose a web service for providing weather information with a sensor network as the backbone. We have used a Fermat point based forwarding technique to minimize the energy consumption of the sensor network which eventually helps the web service work in an uninterrupted fashion for a longer duration, as the life time of the network has prolonged.

  11. Decentralized Orchestration of Composite Ogc Web Processing Services in the Cloud

    NASA Astrophysics Data System (ADS)

    Xiao, F.; Shea, G. Y. K.; Cao, J.

    2016-09-01

    Current web-based GIS or RS applications generally rely on centralized structure, which has inherent drawbacks such as single points of failure, network congestion, and data inconsistency, etc. The inherent disadvantages of traditional GISs need to be solved for new applications on Internet or Web. Decentralized orchestration offers performance improvements in terms of increased throughput and scalability and lower response time. This paper investigates build time and runtime issues related to decentralized orchestration of composite geospatial processing services based on OGC WPS standard specification. A case study of dust storm detection was demonstrated to evaluate the proposed method and the experimental results indicate that the method proposed in this study is effective for its ability to produce the high quality solution at a low cost of communications for geospatial processing service composition problem.

  12. The iceLogo web server and SOAP service for determining protein consensus sequences.

    PubMed

    Maddelein, Davy; Colaert, Niklaas; Buchanan, Iain; Hulstaert, Niels; Gevaert, Kris; Martens, Lennart

    2015-07-01

    The iceLogo web server and SOAP service implement the previously published iceLogo algorithm. iceLogo builds on probability theory to visualize protein consensus sequences in a format resembling sequence logos. Peptide sequences are compared against a reference sequence set that can be tailored to the studied system and the used protocol. As such, not only over- but also underrepresented residues can be visualized in a statistically sound manner, which further allows the user to easily analyse and interpret conserved sequence patterns in proteins. The web application and SOAP service can be found free and open to all users without the need for a login on http://iomics.ugent.be/icelogoserver/main.html.

  13. Protecting Web Services against DoS Attacks: A Case-Based Reasoning Approach

    NASA Astrophysics Data System (ADS)

    Pinzón, Cristian; de Paz, Juan F.; Zato, Carolina; Pérez, Javier

    The real-time detection is a key factor to detect and block DoS attacks within Web services. DoS attacks can be generated for different techniques that take advantage of points vulnerable within Web services. This paper describes a novel proposal based on a real time agent to classify user requests and detect and block malicious SOAP messages. The classification mechanism is based on a Case-Base Reasoning (CBR) model, where the different CBR phases are time bounded. Within the reuse phase of the CBR cycle is incorporated a mixture of experts to choose the most suitable technique of classification depending on the feature of the attack and the available time to solve the classification. A prototype of the architecture was developed and the results obtained are presented in this study.

  14. Pitfalls in Persuasion: How Do Users Experience Persuasive Techniques in a Web Service?

    NASA Astrophysics Data System (ADS)

    Segerståhl, Katarina; Kotro, Tanja; Väänänen-Vainio-Mattila, Kaisa

    Persuasive technologies are designed by utilizing a variety of interactive techniques that are believed to promote target behaviors. This paper describes a field study in which the aim was to discover possible pitfalls of persuasion, i.e., situations in which persuasive techniques do not function as expected. The study investigated persuasive functionality of a web service targeting weight loss. A qualitative online questionnaire was distributed through the web service and a total of 291 responses were extracted for interpretative analysis. The Persuasive Systems Design model (PSD) was used for supporting systematic analysis of persuasive functionality. Pitfalls were identified through situations that evoked negative user experiences. The primary pitfalls discovered were associated with manual logging of eating and exercise behaviors, appropriateness of suggestions and source credibility issues related to social facilitation. These pitfalls, when recognized, can be addressed in design by applying functional and facilitative persuasive techniques in meaningful combinations.

  15. TnpPred: A Web Service for the Robust Prediction of Prokaryotic Transposases

    PubMed Central

    Riadi, Gonzalo; Medina-Moenne, Cristobal; Holmes, David S.

    2012-01-01

    Transposases (Tnps) are enzymes that participate in the movement of insertion sequences (ISs) within and between genomes. Genes that encode Tnps are amongst the most abundant and widely distributed genes in nature. However, they are difficult to predict bioinformatically and given the increasing availability of prokaryotic genomes and metagenomes, it is incumbent to develop rapid, high quality automatic annotation of ISs. This need prompted us to develop a web service, termed TnpPred for Tnp discovery. It provides better sensitivity and specificity for Tnp predictions than given by currently available programs as determined by ROC analysis. TnpPred should be useful for improving genome annotation. The TnpPred web service is freely available for noncommercial use. PMID:23251097

  16. TnpPred: A Web Service for the Robust Prediction of Prokaryotic Transposases.

    PubMed

    Riadi, Gonzalo; Medina-Moenne, Cristobal; Holmes, David S

    2012-01-01

    Transposases (Tnps) are enzymes that participate in the movement of insertion sequences (ISs) within and between genomes. Genes that encode Tnps are amongst the most abundant and widely distributed genes in nature. However, they are difficult to predict bioinformatically and given the increasing availability of prokaryotic genomes and metagenomes, it is incumbent to develop rapid, high quality automatic annotation of ISs. This need prompted us to develop a web service, termed TnpPred for Tnp discovery. It provides better sensitivity and specificity for Tnp predictions than given by currently available programs as determined by ROC analysis. TnpPred should be useful for improving genome annotation. The TnpPred web service is freely available for noncommercial use. PMID:23251097

  17. MaXIC-Q Web: a fully automated web service using statistical and computational methods for protein quantitation based on stable isotope labeling and LC-MS.

    PubMed

    Tsou, Chih-Chiang; Tsui, Yin-Hao; Yian, Yi-Hwa; Chen, Yi-Ju; Yang, Han-Yin; Yu, Chuan-Yih; Lynn, Ke-Shiuan; Chen, Yu-Ju; Sung, Ting-Yi; Hsu, Wen-Lian

    2009-07-01

    Isotope labeling combined with liquid chromatography-mass spectrometry (LC-MS) provides a robust platform for analyzing differential protein expression in proteomics research. We present a web service, called MaXIC-Q Web (http://ms.iis.sinica.edu.tw/MaXIC-Q_Web/), for quantitation analysis of large-scale datasets generated from proteomics experiments using various stable isotope-labeling techniques, e.g. SILAC, ICAT and user-developed labeling methods. It accepts spectral files in the standard mzXML format and search results from SEQUEST, Mascot and ProteinProphet as input. Furthermore, MaXIC-Q Web uses statistical and computational methods to construct two kinds of elution profiles for each ion, namely, PIMS (projected ion mass spectrum) and XIC (extracted ion chromatogram) from MS data. Toward accurate quantitation, a stringent validation procedure is performed on PIMSs to filter out peptide ions interfered with co-eluting peptides or noise. The areas of XICs determine ion abundances, which are used to calculate peptide and protein ratios. Since MaXIC-Q Web adopts stringent validation on spectral data, it achieves high accuracy so that manual validation effort can be substantially reduced. Furthermore, it provides various visualization diagrams and comprehensive quantitation reports so that users can conveniently inspect quantitation results. In summary, MaXIC-Q Web is a user-friendly, interactive, robust, generic web service for quantitation based on ICAT and SILAC labeling techniques.

  18. A web service and android application for the distribution of rainfall estimates and Earth observation data

    NASA Astrophysics Data System (ADS)

    Mantas, V. M.; Liu, Z.; Pereira, A. J. S. C.

    2015-04-01

    The full potential of Satellite Rainfall Estimates (SRE) can only be realized if timely access to the datasets is possible. Existing data distribution web portals are often focused on global products and offer limited customization options, especially for the purpose of routine regional monitoring. Furthermore, most online systems are designed to meet the needs of desktop users, limiting the compatibility with mobile devices. In response to the growing demand for SRE and to address the current limitations of available web portals a project was devised to create a set of freely available applications and services, available at a common portal that can: (1) simplify cross-platform access to Tropical Rainfall Measuring Mission Online Visualization and Analysis System (TOVAS) data (including from Android mobile devices), (2) provide customized and continuous monitoring of SRE in response to user demands and (3) combine data from different online data distribution services, including rainfall estimates, river gauge measurements or imagery from Earth Observation missions at a single portal, known as the Tropical Rainfall Measuring Mission (TRMM) Explorer. The TRMM Explorer project suite includes a Python-based web service and Android applications capable of providing SRE and ancillary data in different intuitive formats with the focus on regional and continuous analysis. The outputs include dynamic plots, tables and data files that can also be used to feed downstream applications and services. A case study in Southern Angola is used to describe the potential of the TRMM Explorer for SRE distribution and analysis in the context of ungauged watersheds. The development of a collection of data distribution instances helped to validate the concept and identify the limitations of the program, in a real context and based on user feedback. The TRMM Explorer can successfully supplement existing web portals distributing SRE and provide a cost-efficient resource to small and medium

  19. Distance and utilisation of out-of-hours services in a Norwegian urban/rural district: an ecological study

    PubMed Central

    2013-01-01

    Background Long travel distances limit the utilisation of health services. We wanted to examine the relationship between the utilisation of a Norwegian out-of-hours service and the distance from the municipality population centroid to the associated casualty clinic. Methods All first contacts from ten municipalities in Arendal out-of-hours district were registered from 2007 through 2011. The main outcomes were contact and consultation rates for each municipality for each year. The associations between main outcomes and distance from the population centroid of the participating municipalities to the casualty clinic and were examined by linear regression. Demographic and socioeconomic factors were included in multivariate linear regression. Secondary endpoints include association between distance and rates of different first actions taken and priority grades assessed by triage nurses. Age and gender specific subgroup analyses were performed. Results 141 342 contacts were included in the analyses. Increasing distance was associated with marked lower rates of all contact types except telephone consultations by doctor. Moving 43 kilometres away from the casualty clinic led to a 50 per cent drop in the rate of face-to-face consultations with a doctor. Availability of primary care doctors and education level contributed to a limited extent to the variance in consultation rate. The rates of all priority grades decreased significantly with increasing distance. The rate of acute events was reduced by 22 per cent when moving 50 kilometres away. The proportion of patients above 66 years increased with increasing distance, while the proportion of 13- to 19 year olds decreased. The proportion of female patients decreased with increasing distance. Conclusions The results confirm that increasing distance is associated with lower utilisation of out-of-hours services, even for the most acute cases. Extremely long distances might compromise patient safety. This must be taken into

  20. GAM: a web-service for integrated transcriptional and metabolic network analysis

    PubMed Central

    Sergushichev, Alexey A.; Loboda, Alexander A.; Jha, Abhishek K.; Vincent, Emma E.; Driggers, Edward M.; Jones, Russell G.; Pearce, Edward J.; Artyomov, Maxim N.

    2016-01-01

    Novel techniques for high-throughput steady-state metabolomic profiling yield information about changes of nearly thousands of metabolites. Such metabolomic profiles, when analyzed together with transcriptional profiles, can reveal novel insights about underlying biological processes. While a number of conceptual approaches have been developed for data integration, easily accessible tools for integrated analysis of mammalian steady-state metabolomic and transcriptional data are lacking. Here we present GAM (‘genes and metabolites’): a web-service for integrated network analysis of transcriptional and steady-state metabolomic data focused on identification of the most changing metabolic subnetworks between two conditions of interest. In the web-service, we have pre-assembled metabolic networks for humans, mice, Arabidopsis and yeast and adapted exact solvers for an optimal subgraph search to work in the context of these metabolic networks. The output is the most regulated metabolic subnetwork of size controlled by false discovery rate parameters. The subnetworks are then visualized online and also can be downloaded in Cytoscape format for subsequent processing. The web-service is available at: https://artyomovlab.wustl.edu/shiny/gam/ PMID:27098040

  1. NaviCell Web Service for network-based data visualization.

    PubMed

    Bonnet, Eric; Viara, Eric; Kuperstein, Inna; Calzone, Laurence; Cohen, David P A; Barillot, Emmanuel; Zinovyev, Andrei

    2015-07-01

    Data visualization is an essential element of biological research, required for obtaining insights and formulating new hypotheses on mechanisms of health and disease. NaviCell Web Service is a tool for network-based visualization of 'omics' data which implements several data visual representation methods and utilities for combining them together. NaviCell Web Service uses Google Maps and semantic zooming to browse large biological network maps, represented in various formats, together with different types of the molecular data mapped on top of them. For achieving this, the tool provides standard heatmaps, barplots and glyphs as well as the novel map staining technique for grasping large-scale trends in numerical values (such as whole transcriptome) projected onto a pathway map. The web service provides a server mode, which allows automating visualization tasks and retrieving data from maps via RESTful (standard HTTP) calls. Bindings to different programming languages are provided (Python and R). We illustrate the purpose of the tool with several case studies using pathway maps created by different research groups, in which data visualization provides new insights into molecular mechanisms involved in systemic diseases such as cancer and neurodegenerative diseases.

  2. NaviCell Web Service for network-based data visualization

    PubMed Central

    Bonnet, Eric; Viara, Eric; Kuperstein, Inna; Calzone, Laurence; Cohen, David P. A.; Barillot, Emmanuel; Zinovyev, Andrei

    2015-01-01

    Data visualization is an essential element of biological research, required for obtaining insights and formulating new hypotheses on mechanisms of health and disease. NaviCell Web Service is a tool for network-based visualization of ‘omics’ data which implements several data visual representation methods and utilities for combining them together. NaviCell Web Service uses Google Maps and semantic zooming to browse large biological network maps, represented in various formats, together with different types of the molecular data mapped on top of them. For achieving this, the tool provides standard heatmaps, barplots and glyphs as well as the novel map staining technique for grasping large-scale trends in numerical values (such as whole transcriptome) projected onto a pathway map. The web service provides a server mode, which allows automating visualization tasks and retrieving data from maps via RESTful (standard HTTP) calls. Bindings to different programming languages are provided (Python and R). We illustrate the purpose of the tool with several case studies using pathway maps created by different research groups, in which data visualization provides new insights into molecular mechanisms involved in systemic diseases such as cancer and neurodegenerative diseases. PMID:25958393

  3. Nurses' opinions about a web-based distance course in a specialist education programme for the care of older people: a questionnaire study.

    PubMed

    Häggström, Elisabeth; Hofsten, Anna; Wadensten, Barbro

    2009-09-01

    Aim.  The aim of the present study was to investigate students' opinions about a web-based distance learning course that was part of a specialist education programme. Background.  Most registered nurses (RNs) who work with older people in Sweden do not have a specialist education in caring for older people. Design.  Survey. Method.  The study was a descriptive survey with both qualitative and quantitative questions. Results.  The students found the web-based course very useful for their theoretical and professional development. New perspectives on ageing, increased knowledge about analysing and describing theories as well as practising interviews and qualitative analysis as methods were appreciated by the students. Conclusions.  One way for RNs in Sweden to get a university specialist education in the care of older people is to have a distance web-based course, as it allows flexibility and improves the students' theoretical and professional knowledge and communication skills. Relevance for clinical practice.  Managers working in care for older people settings and educators will be interested in this study's finding that distance learning is a useful strategy for providing education for RNs who work with older people.

  4. Symmetrical compression distance for arrhythmia discrimination in cloud-based big-data services.

    PubMed

    Lillo-Castellano, J M; Mora-Jiménez, I; Santiago-Mozos, R; Chavarría-Asso, F; Cano-González, A; García-Alberola, A; Rojo-Álvarez, J L

    2015-07-01

    The current development of cloud computing is completely changing the paradigm of data knowledge extraction in huge databases. An example of this technology in the cardiac arrhythmia field is the SCOOP platform, a national-level scientific cloud-based big data service for implantable cardioverter defibrillators. In this scenario, we here propose a new methodology for automatic classification of intracardiac electrograms (EGMs) in a cloud computing system, designed for minimal signal preprocessing. A new compression-based similarity measure (CSM) is created for low computational burden, so-called weighted fast compression distance, which provides better performance when compared with other CSMs in the literature. Using simple machine learning techniques, a set of 6848 EGMs extracted from SCOOP platform were classified into seven cardiac arrhythmia classes and one noise class, reaching near to 90% accuracy when previous patient arrhythmia information was available and 63% otherwise, hence overcoming in all cases the classification provided by the majority class. Results show that this methodology can be used as a high-quality service of cloud computing, providing support to physicians for improving the knowledge on patient diagnosis.

  5. Symmetrical compression distance for arrhythmia discrimination in cloud-based big-data services.

    PubMed

    Lillo-Castellano, J M; Mora-Jiménez, I; Santiago-Mozos, R; Chavarría-Asso, F; Cano-González, A; García-Alberola, A; Rojo-Álvarez, J L

    2015-07-01

    The current development of cloud computing is completely changing the paradigm of data knowledge extraction in huge databases. An example of this technology in the cardiac arrhythmia field is the SCOOP platform, a national-level scientific cloud-based big data service for implantable cardioverter defibrillators. In this scenario, we here propose a new methodology for automatic classification of intracardiac electrograms (EGMs) in a cloud computing system, designed for minimal signal preprocessing. A new compression-based similarity measure (CSM) is created for low computational burden, so-called weighted fast compression distance, which provides better performance when compared with other CSMs in the literature. Using simple machine learning techniques, a set of 6848 EGMs extracted from SCOOP platform were classified into seven cardiac arrhythmia classes and one noise class, reaching near to 90% accuracy when previous patient arrhythmia information was available and 63% otherwise, hence overcoming in all cases the classification provided by the majority class. Results show that this methodology can be used as a high-quality service of cloud computing, providing support to physicians for improving the knowledge on patient diagnosis. PMID:25823046

  6. Global-Scale Resource Survey and Performance Monitoring of Public OGC Web Map Services

    NASA Astrophysics Data System (ADS)

    Gui, Zhipeng; Cao, Jun; Liu, Xiaojing; Cheng, Xiaoqiang; Wu, Huayi

    2016-06-01

    One of the most widely-implemented service standards provided by the Open Geospatial Consortium (OGC) to the user community is the Web Map Service (WMS). WMS is widely employed globally, but there is limited knowledge of the global distribution, adoption status or the service quality of these online WMS resources. To fill this void, we investigated global WMSs resources and performed distributed performance monitoring of these services. This paper explicates a distributed monitoring framework that was used to monitor 46,296 WMSs continuously for over one year and a crawling method to discover these WMSs. We analyzed server locations, provider types, themes, the spatiotemporal coverage of map layers and the service versions for 41,703 valid WMSs. Furthermore, we appraised the stability and performance of basic operations for 1210 selected WMSs (i.e., GetCapabilities and GetMap). We discuss the major reasons for request errors and performance issues, as well as the relationship between service response times and the spatiotemporal distribution of client monitoring sites. This paper will help service providers, end users and developers of standards to grasp the status of global WMS resources, as well as to understand the adoption status of OGC standards. The conclusions drawn in this paper can benefit geospatial resource discovery, service performance evaluation and guide service performance improvements.

  7. A Real-Time Web Services Hub to Improve Situation Awareness during Flash Flood Events

    NASA Astrophysics Data System (ADS)

    Salas, F. R.; Liu, F.; Maidment, D. R.; Hodges, B. R.

    2011-12-01

    The central Texas corridor is one of the most flash flood-prone regions in the United States. Over the years, flash floods have resulted in hundreds of flood fatalities and billions of dollars in property damage. In order to mitigate risk to residents and infrastructure during flood events, both citizens and emergency responders need to exhibit proactive behavior instead of reactive. Real-time and forecasted flood information is fairly limited and hard to come by at varying spatial scales. The University of Texas at Austin has collaborated with IBM Research-Austin and ESRI to build a distributed real-time flood information system through a framework that leverages large scale data management and distribution, Open Geospatial Consortium standardized web services, and smart map applications. Within this paradigm, observed precipitation data encoded in WaterML is ingested into HEC-HMS and then delivered to a high performance hydraulic routing software package developed by IBM that utilizes the latest advancements in VLSI design, numerical linear algebra and numerical integration techniques on contemporary multicore architecture to solve fully dynamic Saint Venant equations at both small and large scales. In this paper we present a real-time flood inundation map application that in conjunction with a web services Hub, seamlessly integrates hydrologic information available through both public and private data services, model services and mapping services. As a case study for this project, we demonstrate how this system has been implemented in the City of Austin, Texas.

  8. TogoWS: integrated SOAP and REST APIs for interoperable bioinformatics Web services.

    PubMed

    Katayama, Toshiaki; Nakao, Mitsuteru; Takagi, Toshihisa

    2010-07-01

    Web services have become widely used in bioinformatics analysis, but there exist incompatibilities in interfaces and data types, which prevent users from making full use of a combination of these services. Therefore, we have developed the TogoWS service to provide an integrated interface with advanced features. In the TogoWS REST (REpresentative State Transfer) API (application programming interface), we introduce a unified access method for major database resources through intuitive URIs that can be used to search, retrieve, parse and convert the database entries. The TogoWS SOAP API resolves compatibility issues found on the server and client-side SOAP implementations. The TogoWS service is freely available at: http://togows.dbcls.jp/.

  9. Measuring the impact of the approach to migration in the quality of web service interfaces

    NASA Astrophysics Data System (ADS)

    Mateos, Cristian; Crasso, Marco; Rodriguez, Juan M.; Zunino, Alejandro; Campo, Marcelo

    2015-01-01

    There is a good consensus on the strategic value of service-oriented architecture (SOA) as a way of structuring systems, and a common trend is to migrate legacy applications that use outdated technologies and architectures to SOA. We study the effects in the resulting Web Service interfaces of applying two traditional migration approaches combined with common ways of building services, namely, direct migration with code-first and indirect migration with contract-first. The migrated system was a 35-year-old COBOL system of a government agency that serves several millions of users. In addition, we provide a deep explanation of the trade-offs involved in following either combinations. Results confirm that the 'fast and cheap' approach to move into SOA, which is commonplace in the industry, may deliver poor service interfaces, and interface quality is also subject to the tools supporting the migration process.

  10. Building the Service-Based Library Web Site: A Step-by-Step Guide to Design and Options.

    ERIC Educational Resources Information Center

    Garlock, Kristen L.; Piontek, Sherry

    The World Wide Web, with its captivating multimedia features and hypertext capabilities, has brought millions of new users to the Internet. Library staff who could create a home page on the Web could present basic information about the library and its services, showcase its resources, create links to quality material inside and outside the…

  11. 78 FR 26664 - Submission for Review: CyberCorps®: Scholarship For Service (SFS) Registration Web Site

    Federal Register 2010, 2011, 2012, 2013, 2014

    2013-05-07

    ... From the Federal Register Online via the Government Publishing Office OFFICE OF PERSONNEL MANAGEMENT Submission for Review: CyberCorps : Scholarship For Service (SFS) Registration Web Site AGENCY: U... of the scholarship or one year, whichever is longer. Approval of the Web page is necessary...

  12. Oh! Web 2.0, Virtual Reference Service 2.0, Tools and Techniques (I): A Basic Approach

    ERIC Educational Resources Information Center

    Arya, Harsh Bardhan; Mishra, J. K.

    2011-01-01

    This study targets librarians and information professionals who use Web 2.0 tools and applications with a view to providing snapshots on how Web 2.0 technologies are used. It also aims to identify values and impact that such tools have exerted on libraries and their services, as well as to detect various issues associated with the implementation…

  13. I Help, Therefore, I Learn: Service Learning on Web 2.0 in an EFL Speaking Class

    ERIC Educational Resources Information Center

    Sun, Yu-Chih; Yang, Fang-Ying

    2015-01-01

    The present study integrates service learning into English as a Foreign Language (EFL) speaking class using Web 2.0 tools--YouTube and Facebook--as platforms. Fourteen undergraduate students participated in the study. The purpose of the service-learning project was to link service learning with oral communication training in an EFL speaking class…

  14. The Footprint Database and Web Services of the Herschel Space Observatory

    NASA Astrophysics Data System (ADS)

    Dobos, László; Varga-Verebélyi, Erika; Verdugo, Eva; Teyssier, David; Exter, Katrina; Valtchanov, Ivan; Budavári, Tamás; Kiss, Csaba

    2016-07-01

    Data from the Herschel Space Observatory is freely available to the public but no uniformly processed catalogue of the observations has been published so far. To date, the Herschel Science Archive does not contain the exact sky coverage (footprint) of individual observations and supports search for measurements based on bounding circles only. Drawing on previous experience in implementing footprint databases, we built the Herschel Footprint Database and Web Services for the Herschel Space Observatory to provide efficient search capabilities for typical astronomical queries. The database was designed with the following main goals in mind: (a) provide a unified data model for meta-data of all instruments and observational modes, (b) quickly find observations covering a selected object and its neighbourhood, (c) quickly find every observation in a larger area of the sky, (d) allow for finding solar system objects crossing observation fields. As a first step, we developed a unified data model of observations of all three Herschel instruments for all pointing and instrument modes. Then, using telescope pointing information and observational meta-data, we compiled a database of footprints. As opposed to methods using pixellation of the sphere, we represent sky coverage in an exact geometric form allowing for precise area calculations. For easier handling of Herschel observation footprints with rather complex shapes, two algorithms were implemented to reduce the outline. Furthermore, a new visualisation tool to plot footprints with various spherical projections was developed. Indexing of the footprints using Hierarchical Triangular Mesh makes it possible to quickly find observations based on sky coverage, time and meta-data. The database is accessible via a web site http://herschel.vo.elte.hu and also as a set of REST web service functions, which makes it readily usable from programming environments such as Python or IDL. The web service allows downloading footprint data

  15. The Footprint Database and Web Services of the Herschel Space Observatory

    NASA Astrophysics Data System (ADS)

    Dobos, László; Varga-Verebélyi, Erika; Verdugo, Eva; Teyssier, David; Exter, Katrina; Valtchanov, Ivan; Budavári, Tamás; Kiss, Csaba

    2016-10-01

    Data from the Herschel Space Observatory is freely available to the public but no uniformly processed catalogue of the observations has been published so far. To date, the Herschel Science Archive does not contain the exact sky coverage (footprint) of individual observations and supports search for measurements based on bounding circles only. Drawing on previous experience in implementing footprint databases, we built the Herschel Footprint Database and Web Services for the Herschel Space Observatory to provide efficient search capabilities for typical astronomical queries. The database was designed with the following main goals in mind: (a) provide a unified data model for meta-data of all instruments and observational modes, (b) quickly find observations covering a selected object and its neighbourhood, (c) quickly find every observation in a larger area of the sky, (d) allow for finding solar system objects crossing observation fields. As a first step, we developed a unified data model of observations of all three Herschel instruments for all pointing and instrument modes. Then, using telescope pointing information and observational meta-data, we compiled a database of footprints. As opposed to methods using pixellation of the sphere, we represent sky coverage in an exact geometric form allowing for precise area calculations. For easier handling of Herschel observation footprints with rather complex shapes, two algorithms were implemented to reduce the outline. Furthermore, a new visualisation tool to plot footprints with various spherical projections was developed. Indexing of the footprints using Hierarchical Triangular Mesh makes it possible to quickly find observations based on sky coverage, time and meta-data. The database is accessible via a web site http://herschel.vo.elte.hu and also as a set of REST web service functions, which makes it readily usable from programming environments such as Python or IDL. The web service allows downloading footprint data

  16. The promise of the internet for disability: a study of on-line services and web site accessibility at Centers for Independent Living.

    PubMed

    Ritchie, Heather; Blanck, Peter

    2003-01-01

    The Internet provides individuals with disabilities numerous tools to live independently. In the convenience of the home, a person can access an abundance of information, an electronic community, updates on the latest disability advocacy news, education through distance-learning classes, and on-line shopping for books, clothes, assistive technology, and a host of other consumer goods. Centers for Independent Living (CILs) are consumer-run, non-profit grassroots disability service organizations at the forefront of the disability rights movement. Providing services to individuals across the range of disabilities, CILs have begun to use the Internet as a complement to their traditional service delivery methods. This article examines the emerging trend of independent living services on the web. The investigation examines 200 CIL Internet sites across the United States during the period of April to August 2001. Information is collected and analyzed about how CILs are using the Internet to provide their services and programs. In addition, the article examines the technological accessibility of their web sites. Implications of the findings for CILs, consumers with disabilities, and disability policy are examined.

  17. Web servers and services for electrostatics calculations with APBS and PDB2PQR

    SciTech Connect

    Unni, Samir; Huang, Yong; Hanson, Robert M.; Tobias, Malcolm; Krishnan, Sriram; Li, Wilfred; Nielsen, Jens E.; Baker, Nathan A.

    2011-04-02

    APBS and PDB2PQR are widely utilized free software packages for biomolecular electrostatics calculations. Using the Opal toolkit, we have developed a web services framework for these software packages that enables the use of APBS and PDB2PQR by users who do not have local access to the necessary amount of computational capabilities. This not only increases accessibility of the software to a wider range of scientists, educators, and students but it also increases the availability of electrostatics calculations on portable computing platforms. Users can access this new functionality in two ways. First, an Opal-enabled version of APBS is provided in current distributions, available freely on the web. Second, we have extended the PDB2PQR web server to provide an interface for the setup, execution, and visualization electrostatics potentials as calculated by APBS. This web interface also uses the Opal framework which ensures the scalability needed to support the large APBS user community. Both of these resources are available from the APBS/PDB2PQR website: http://www.poissonboltzmann.org/.

  18. Design and Development of a Framework Based on Ogc Web Services for the Visualization of Three Dimensional Large-Scale Geospatial Data Over the Web

    NASA Astrophysics Data System (ADS)

    Roccatello, E.; Nozzi, A.; Rumor, M.

    2013-05-01

    This paper illustrates the key concepts behind the design and the development of a framework, based on OGC services, capable to visualize 3D large scale geospatial data streamed over the web. WebGISes are traditionally bounded to a bi-dimensional simplified representation of the reality and though they are successfully addressing the lack of flexibility and simplicity of traditional desktop clients, a lot of effort is still needed to reach desktop GIS features, like 3D visualization. The motivations behind this work lay in the widespread availability of OGC Web Services inside government organizations and in the technology support to HTML 5 and WebGL standard of the web browsers. This delivers an improved user experience, similar to desktop applications, therefore allowing to augment traditional WebGIS features with a 3D visualization framework. This work could be seen as an extension of the Cityvu project, started in 2008 with the aim of a plug-in free OGC CityGML viewer. The resulting framework has also been integrated in existing 3DGIS software products and will be made available in the next months.

  19. UniPrime2: a web service providing easier Universal Primer design

    PubMed Central

    Boutros, Robin; Stokes, Nicola; Bekaert, Michaël; Teeling, Emma C.

    2009-01-01

    The UniPrime2 web server is a publicly available online resource which automatically designs large sets of universal primers when given a gene reference ID or Fasta sequence input by a user. UniPrime2 works by automatically retrieving and aligning homologous sequences from GenBank, identifying regions of conservation within the alignment, and generating suitable primers that can be used to amplify variable genomic regions. In essence, UniPrime2 is a suite of publicly available software packages (Blastn, T-Coffee, GramAlign, Primer3), which reduces the laborious process of primer design, by integrating these programs into a single software pipeline. Hence, UniPrime2 differs from previous primer design web services in that all steps are automated, linked, saved and phylogenetically delimited, only requiring a single user-defined gene reference ID or input sequence. We provide an overview of the web service and wet-laboratory validation of the primers generated. The system is freely accessible at: http://uniprime.batlab.eu. UniPrime2 is licenced under a Creative Commons Attribution Noncommercial-Share Alike 3.0 Licence. PMID:19401435

  20. A Web Service Protocol Realizing Interoperable Internet of Things Tasking Capability.

    PubMed

    Huang, Chih-Yuan; Wu, Cheng-Hung

    2016-08-31

    The Internet of Things (IoT) is an infrastructure that interconnects uniquely-identifiable devices using the Internet. By interconnecting everyday appliances, various monitoring, and physical mashup applications can be constructed to improve human's daily life. In general, IoT devices provide two main capabilities: sensing and tasking capabilities. While the sensing capability is similar to the World-Wide Sensor Web, this research focuses on the tasking capability. However, currently, IoT devices created by different manufacturers follow different proprietary protocols and are locked in many closed ecosystems. This heterogeneity issue impedes the interconnection between IoT devices and damages the potential of the IoT. To address this issue, this research aims at proposing an interoperable solution called tasking capability description that allows users to control different IoT devices using a uniform web service interface. This paper demonstrates the contribution of the proposed solution by interconnecting different IoT devices for different applications. In addition, the proposed solution is integrated with the OGC SensorThings API standard, which is a Web service standard defined for the IoT sensing capability. Consequently, the Extended SensorThings API can realize both IoT sensing and tasking capabilities in an integrated and interoperable manner.

  1. A Web Service Protocol Realizing Interoperable Internet of Things Tasking Capability

    PubMed Central

    Huang, Chih-Yuan; Wu, Cheng-Hung

    2016-01-01

    The Internet of Things (IoT) is an infrastructure that interconnects uniquely-identifiable devices using the Internet. By interconnecting everyday appliances, various monitoring, and physical mashup applications can be constructed to improve human’s daily life. In general, IoT devices provide two main capabilities: sensing and tasking capabilities. While the sensing capability is similar to the World-Wide Sensor Web, this research focuses on the tasking capability. However, currently, IoT devices created by different manufacturers follow different proprietary protocols and are locked in many closed ecosystems. This heterogeneity issue impedes the interconnection between IoT devices and damages the potential of the IoT. To address this issue, this research aims at proposing an interoperable solution called tasking capability description that allows users to control different IoT devices using a uniform web service interface. This paper demonstrates the contribution of the proposed solution by interconnecting different IoT devices for different applications. In addition, the proposed solution is integrated with the OGC SensorThings API standard, which is a Web service standard defined for the IoT sensing capability. Consequently, the Extended SensorThings API can realize both IoT sensing and tasking capabilities in an integrated and interoperable manner. PMID:27589759

  2. Twitter web-service for soft agent reporting in persistent surveillance systems

    NASA Astrophysics Data System (ADS)

    Rababaah, Haroun; Shirkhodaie, Amir

    2010-04-01

    Persistent surveillance is an intricate process requiring monitoring, gathering, processing, tracking, and characterization of many spatiotemporal events occurring concurrently. Data associated with events can be readily attained by networking of hard (physical) sensors. Sensors may have homogeneous or heterogeneous (hybrid) sensing modalities with different communication bandwidth requirements. Complimentary to hard sensors are human observers or "soft sensors" that can report occurrences of evolving events via different communication devices (e.g., texting, cell phones, emails, instant messaging, etc.) to the command control center. However, networking of human observers in ad-hoc way is rather a difficult task. In this paper, we present a Twitter web-service for soft agent reporting in persistent surveillance systems (called Web-STARS). The objective of this web-service is to aggregate multi-source human observations in hybrid sensor networks rapidly. With availability of Twitter social network, such a human networking concept can not only be realized for large scale persistent surveillance systems (PSS), but also, it can be employed with proper interfaces to expedite rapid events reporting by human observers. The proposed technique is particularly suitable for large-scale persistent surveillance systems with distributed soft and hard sensor networks. The efficiency and effectiveness of the proposed technique is measured experimentally by conducting several simulated persistent surveillance scenarios. It is demonstrated that by fusion of information from hard and soft agents improves understanding of common operating picture and enhances situational awareness.

  3. A Web Service Protocol Realizing Interoperable Internet of Things Tasking Capability.

    PubMed

    Huang, Chih-Yuan; Wu, Cheng-Hung

    2016-01-01

    The Internet of Things (IoT) is an infrastructure that interconnects uniquely-identifiable devices using the Internet. By interconnecting everyday appliances, various monitoring, and physical mashup applications can be constructed to improve human's daily life. In general, IoT devices provide two main capabilities: sensing and tasking capabilities. While the sensing capability is similar to the World-Wide Sensor Web, this research focuses on the tasking capability. However, currently, IoT devices created by different manufacturers follow different proprietary protocols and are locked in many closed ecosystems. This heterogeneity issue impedes the interconnection between IoT devices and damages the potential of the IoT. To address this issue, this research aims at proposing an interoperable solution called tasking capability description that allows users to control different IoT devices using a uniform web service interface. This paper demonstrates the contribution of the proposed solution by interconnecting different IoT devices for different applications. In addition, the proposed solution is integrated with the OGC SensorThings API standard, which is a Web service standard defined for the IoT sensing capability. Consequently, the Extended SensorThings API can realize both IoT sensing and tasking capabilities in an integrated and interoperable manner. PMID:27589759

  4. Grid enablement of OpenGeospatial Web Services: the G-OWS Working Group

    NASA Astrophysics Data System (ADS)

    Mazzetti, Paolo

    2010-05-01

    In last decades two main paradigms for resource sharing emerged and reached maturity: the Web and the Grid. They both demonstrate suitable for building Distributed Computing Infrastructures (DCIs) supporting the coordinated sharing of resources (i.e. data, information, services, etc) on the Internet. Grid and Web DCIs have much in common as a result of their underlying Internet technology (protocols, models and specifications). However, being based on different requirements and architectural approaches, they show some differences as well. The Web's "major goal was to be a shared information space through which people and machines could communicate" [Berners-Lee 1996]. The success of the Web, and its consequent pervasiveness, made it appealing for building specialized systems like the Spatial Data Infrastructures (SDIs). In this systems the introduction of Web-based geo-information technologies enables specialized services for geospatial data sharing and processing. The Grid was born to achieve "flexible, secure, coordinated resource sharing among dynamic collections of individuals, institutions, and resources" [Foster 2001]. It specifically focuses on large-scale resource sharing, innovative applications, and, in some cases, high-performance orientation. In the Earth and Space Sciences (ESS) the most part of handled information is geo-referred (geo-information) since spatial and temporal meta-information is of primary importance in many application domains: Earth Sciences, Disasters Management, Environmental Sciences, etc. On the other hand, in several application areas there is the need of running complex models which require the large processing and storage capabilities that the Grids are able to provide. Therefore the integration of geo-information and Grid technologies might be a valuable approach in order to enable advanced ESS applications. Currently both geo-information and Grid technologies have reached a high level of maturity, allowing to build such an

  5. Apollo: Giving application developers a single point of access to public health models using structured vocabularies and Web services

    PubMed Central

    Wagner, Michael M.; Levander, John D.; Brown, Shawn; Hogan, William R.; Millett, Nicholas; Hanna, Josh

    2013-01-01

    This paper describes the Apollo Web Services and Apollo-SV, its related ontology. The Apollo Web Services give an end-user application a single point of access to multiple epidemic simulators. An end user can specify an analytic problem—which we define as a configuration and a query of results—exactly once and submit it to multiple epidemic simulators. The end user represents the analytic problem using a standard syntax and vocabulary, not the native languages of the simulators. We have demonstrated the feasibility of this design by implementing a set of Apollo services that provide access to two epidemic simulators and two visualizer services. PMID:24551417

  6. Theoretical Virtual Observatory and Grid Web services: VisIVO and new capabilities

    NASA Astrophysics Data System (ADS)

    Costa, A.; Becciani, U.; Gheller, C.; Comparato, M.; Larsson, B.

    2007-08-01

    VisIVO is a tool for 3D visualization, it provides an effective and intuitive way of managing, visualizing and analysing the large amount of data produced by observations and numerical simulations. The software is specifically designed to deal with multidimensional data. Catalogues and numerical simulations represent the basic target of VisIVO. The package is written in C++. This poster describes VIsIVO's Grid Web Service (VWS) and its counterpart client side developed in VisIVO. VisIVO's Grid Web Service is developed as a part of the Italian Virtual Observatory, it allows to run applications in grid using the web service technology. The VWS has been designed to work within a Virtual Observatory environment so that the interface for the current application, shown in this poster, and the interfaces for other applications are described by a small and constant piece of WSDL code. Our first application is HOP: an algorithm for finding groups of particles based on the one developed and coded by Daniel Eisenstein & Piet Hut, Institute for Advanced Study, Princeton, NJ. We have developed VWS using Java AXIS libraries for the server side and C++ AXIS libraries for the client side. The access to the computational resources and storage areas is based on grid services in the INFN Production Grid and from this the VWS inherits asynchronous features, scheduling and matching algorithms. HOP was distributed as RPM Package and was installed in the Worker Node Elements of the INFN Production Grid. The idea is to use external tools through the grid avoiding the integration in our application. We can focus on the I/O management of the jobs and on the standardization of the access methods for the different analysis tools improving the scalability of our solution.

  7. Semantic-JSON: a lightweight web service interface for Semantic Web contents integrating multiple life science databases.

    PubMed

    Kobayashi, Norio; Ishii, Manabu; Takahashi, Satoshi; Mochizuki, Yoshiki; Matsushima, Akihiro; Toyoda, Tetsuro

    2011-07-01

    Global cloud frameworks for bioinformatics research databases become huge and heterogeneous; solutions face various diametric challenges comprising cross-integration, retrieval, security and openness. To address this, as of March 2011 organizations including RIKEN published 192 mammalian, plant and protein life sciences databases having 8.2 million data records, integrated as Linked Open or Private Data (LOD/LPD) using SciNetS.org, the Scientists' Networking System. The huge quantity of linked data this database integration framework covers is based on the Semantic Web, where researchers collaborate by managing metadata across public and private databases in a secured data space. This outstripped the data query capacity of existing interface tools like SPARQL. Actual research also requires specialized tools for data analysis using raw original data. To solve these challenges, in December 2009 we developed the lightweight Semantic-JSON interface to access each fragment of linked and raw life sciences data securely under the control of programming languages popularly used by bioinformaticians such as Perl and Ruby. Researchers successfully used the interface across 28 million semantic relationships for biological applications including genome design, sequence processing, inference over phenotype databases, full-text search indexing and human-readable contents like ontology and LOD tree viewers. Semantic-JSON services of SciNetS.org are provided at http://semanticjson.org.

  8. Semantic-JSON: a lightweight web service interface for Semantic Web contents integrating multiple life science databases

    PubMed Central

    Kobayashi, Norio; Ishii, Manabu; Takahashi, Satoshi; Mochizuki, Yoshiki; Matsushima, Akihiro; Toyoda, Tetsuro

    2011-01-01

    Global cloud frameworks for bioinformatics research databases become huge and heterogeneous; solutions face various diametric challenges comprising cross-integration, retrieval, security and openness. To address this, as of March 2011 organizations including RIKEN published 192 mammalian, plant and protein life sciences databases having 8.2 million data records, integrated as Linked Open or Private Data (LOD/LPD) using SciNetS.org, the Scientists' Networking System. The huge quantity of linked data this database integration framework covers is based on the Semantic Web, where researchers collaborate by managing metadata across public and private databases in a secured data space. This outstripped the data query capacity of existing interface tools like SPARQL. Actual research also requires specialized tools for data analysis using raw original data. To solve these challenges, in December 2009 we developed the lightweight Semantic-JSON interface to access each fragment of linked and raw life sciences data securely under the control of programming languages popularly used by bioinformaticians such as Perl and Ruby. Researchers successfully used the interface across 28 million semantic relationships for biological applications including genome design, sequence processing, inference over phenotype databases, full-text search indexing and human-readable contents like ontology and LOD tree viewers. Semantic-JSON services of SciNetS.org are provided at http://semanticjson.org. PMID:21632604

  9. Semantic enrichment of medical forms - semi-automated coding of ODM-elements via web services.

    PubMed

    Breil, Bernhard; Watermann, Andreas; Haas, Peter; Dziuballe, Philipp; Dugas, Martin

    2012-01-01

    Semantic interoperability is an unsolved problem which occurs while working with medical forms from different information systems or institutions. Standards like ODM or CDA assure structural homogenization but in order to compare elements from different data models it is necessary to use semantic concepts and codes on an item level of those structures. We developed and implemented a web-based tool which enables a domain expert to perform semi-automated coding of ODM-files. For each item it is possible to inquire web services which result in unique concept codes without leaving the context of the document. Although it was not feasible to perform a totally automated coding we have implemented a dialog based method to perform an efficient coding of all data elements in the context of the whole document. The proportion of codable items was comparable to results from previous studies.

  10. The Lexicon Builder Web service: Building Custom Lexicons from two hundred Biomedical Ontologies

    PubMed Central

    Parai, Gautam K.; Jonquet, Clement; Xu, Rong; Musen, Mark A.; Shah, Nigam H.

    2010-01-01

    Domain specific biomedical lexicons are extensively used by researchers for natural language processing tasks. Currently these lexicons are created manually by expert curators and there is a pressing need for automated methods to compile such lexicons. The Lexicon Builder Web service addresses this need and reduces the investment of time and effort involved in lexicon maintenance. The service has three components: Inclusion – selects one or several ontologies (or its branches) and includes preferred names and synonym terms; Exclusion – filters terms based on the term’s Medline frequency, syntactic type, UMLS semantic type and match with stopwords; Output – aggregates information, handles compression and output formats. Evaluation demonstrates that the service has high accuracy and runtime performance. It is currently being evaluated for several use cases to establish its utility in biomedical information processing tasks. The Lexicon Builder promotes collaboration, sharing and standardization of lexicons amongst researchers by automating the creation, maintainence and cross referencing of custom lexicons. PMID:21347046

  11. Cross-Dataset Analysis and Visualization Driven by Expressive Web Services

    NASA Astrophysics Data System (ADS)

    Alexandru Dumitru, Mircea; Catalin Merticariu, Vlad

    2015-04-01

    The deluge of data that is hitting us every day from satellite and airborne sensors is changing the workflow of environmental data analysts and modelers. Web geo-services play now a fundamental role, and are no longer needed to preliminary download and store the data, but rather they interact in real-time with GIS applications. Due to the very large amount of data that is curated and made available by web services, it is crucial to deploy smart solutions for optimizing network bandwidth, reducing duplication of data and moving the processing closer to the data. In this context we have created a visualization application for analysis and cross-comparison of aerosol optical thickness datasets. The application aims to help researchers identify and visualize discrepancies between datasets coming from various sources, having different spatial and time resolutions. It also acts as a proof of concept for integration of OGC Web Services under a user-friendly interface that provides beautiful visualizations of the explored data. The tool was built on top of the World Wind engine, a Java based virtual globe built by NASA and the open source community. For data retrieval and processing we exploited the OGC Web Coverage Service potential: the most exciting aspect being its processing extension, a.k.a. the OGC Web Coverage Processing Service (WCPS) standard. A WCPS-compliant service allows a client to execute a processing query on any coverage offered by the server. By exploiting a full grammar, several different kinds of information can be retrieved from one or more datasets together: scalar condensers, cross-sectional profiles, comparison maps and plots, etc. This combination of technology made the application versatile and portable. As the processing is done on the server-side, we ensured that the minimal amount of data is transferred and that the processing is done on a fully-capable server, leaving the client hardware resources to be used for rendering the visualization

  12. Integrating Distributed Physical and Biological Marine data using OGC Web Services

    NASA Astrophysics Data System (ADS)

    Gemmell, A. L.; Blower, J. D.; Haines, K.; Price, M.; Millard, K.; Harpham, Q.

    2008-12-01

    Earth scientists use highly diverse sources of data, including in-situ measurements, remotely-sensed information and the results of numerical simulations. The ability to access, visualize, combine and compare these datasets is at the core of scientific investigation, but such tasks have hitherto been very difficult or impossible due to a fundamental lack of harmonization of data products. As a result, much valuable data remains underused. We present a web portal that visualizes and compares physical and biological marine data from both numerical models and in-situ observations. The model data are obtained via an Open Geospatial Consortium (OGC)-compatible Web Map Service (WMS), and the observed data are obtained via an OGC Web Feature Service (WFS). The physical model WMS, the biological model WMS and the WFS are located at three different institutes. This ability to display in-situ point observations alongside model data facilitates much valuable work on model validation. As models become increasingly complex, and sources of observed data become more numerous, it is important to be able to access and compare this growing amount of data efficiently, to ensure cross-checking and consistency between models and observations. The web portal is being applied in a large European operational oceanography project (ECOOP), where it is used to provide support to ecosystem modellers, and specifically to aid detection of potentially harmful algal blooms in coastal areas. The development of this system has been enabled by the conceptual framework of the Climate Science Modelling Language (CSML), which provides a common view onto all these datasets, independent of their storage format or physical location. CSML is based upon emerging international standards, enabling interoperability with other standards-based infrastructures. By creating a reusable Java library that embodies the CSML concepts we are able to apply these techniques to a number of other projects.

  13. Data Access and Web Services at the EarthScope Plate Boundary Observatory

    NASA Astrophysics Data System (ADS)

    Matykiewicz, J.; Anderson, G.; Henderson, D.; Hodgkinson, K.; Hoyt, B.; Lee, E.; Persson, E.; Torrez, D.; Smith, J.; Wright, J.; Jackson, M.

    2007-12-01

    The EarthScope Plate Boundary Observatory (PBO) at UNAVCO, Inc., part of the NSF-funded EarthScope project, is designed to study the three-dimensional strain field resulting from deformation across the active boundary zone between the Pacific and North American plates in the western United States. To meet these goals, PBO will install 880 continuous GPS stations, 103 borehole strainmeter stations, and five laser strainmeters, as well as manage data for 209 previously existing continuous GPS stations and one previously existing laser strainmeter. UNAVCO provides access to data products from these stations, as well as general information about the PBO project, via the PBO web site (http://pboweb.unavco.org). GPS and strainmeter data products can be found using a variety of access methods, incuding map searches, text searches, and station specific data retrieval. In addition, the PBO construction status is available via multiple mapping interfaces, including custom web based map widgets and Google Earth. Additional construction details can be accessed from PBO operational pages and station specific home pages. The current state of health for the PBO network is available with the statistical snap-shot, full map interfaces, tabular web based reports, and automatic data mining and alerts. UNAVCO is currently working to enhance the community access to this information by developing a web service framework for the discovery of data products, interfacing with operational engineers, and exposing data services to third party participants. In addition, UNAVCO, through the PBO project, provides advanced data management and monitoring systems for use by the community in operating geodetic networks in the United States and beyond. We will demonstrate these systems during the AGU meeting, and we welcome inquiries from the community at any time.

  14. Sharing environmental models: An Approach using GitHub repositories and Web Processing Services

    NASA Astrophysics Data System (ADS)

    Stasch, Christoph; Nuest, Daniel; Pross, Benjamin

    2016-04-01

    The GLUES (Global Assessment of Land Use Dynamics, Greenhouse Gas Emissions and Ecosystem Services) project established a spatial data infrastructure for scientific geospatial data and metadata (http://geoportal-glues.ufz.de), where different regional collaborative projects researching the impacts of climate and socio-economic changes on sustainable land management can share their underlying base scenarios and datasets. One goal of the project is to ease the sharing of computational models between institutions and to make them easily executable in Web-based infrastructures. In this work, we present such an approach for sharing computational models relying on GitHub repositories (http://github.com) and Web Processing Services. At first, model providers upload their model implementations to GitHub repositories in order to share them with others. The GitHub platform allows users to submit changes to the model code. The changes can be discussed and reviewed before merging them. However, while GitHub allows sharing and collaborating of model source code, it does not actually allow running these models, which requires efforts to transfer the implementation to a model execution framework. We thus have extended an existing implementation of the OGC Web Processing Service standard (http://www.opengeospatial.org/standards/wps), the 52°North Web Processing Service (http://52north.org/wps) platform to retrieve all model implementations from a git (http://git-scm.com) repository and add them to the collection of published geoprocesses. The current implementation is restricted to models implemented as R scripts using WPS4R annotations (Hinz et al.) and to Java algorithms using the 52°North WPS Java API. The models hence become executable through a standardized Web API by multiple clients such as desktop or browser GIS and modelling frameworks. If the model code is changed on the GitHub platform, the changes are retrieved by the service and the processes will be updated

  15. Web-based data delivery services in support of disaster-relief applications

    USGS Publications Warehouse

    Jones, B.K.; Risty, R.R.; Buswell, M.

    2003-01-01

    The U.S. Geological Survey Earth Resources Observation Systems Data Center responds to emergencies in support of various government agencies for human-induced and natural disasters. This response consists of satellite tasking and acquisitions, satellite image registrations, disaster-extent maps analysis and creation, base image provision and support, Web-based mapping services for product delivery, and predisaster and postdisaster data archiving. The emergency response staff are on call 24 hours a day, 7 days a week, and have access to many commercial and government satellite and aerial photography tasking authorities. They have access to value-added data processing and photographic laboratory services for off-hour emergency requests. They work with various Federal agencies for preparedness planning, which includes providing base imagery. These data may include digital elevation models, hydrographic models, base satellite images, vector data layers such as roads, aerial photographs, and other predisaster data. These layers are incorporated into a Web-based browser and data delivery service that is accessible either to the general public or to select customers. As usage declines, the data are moved to a postdisaster nearline archive that is still accessible, but not in real time.

  16. SemBiosphere: a semantic web approach to recommending microarray clustering services.

    PubMed

    Yip, Kevin Y; Qi, Peishen; Schultz, Martin; Cheung, David W; Cheung, Kei-Hoi

    2006-01-01

    Clustering is a popular method for analyzing microarray data. Given the large number of clustering algorithms being available, it is difficult to identify the most suitable ones for a particular task. It is also difficult to locate, download, install and run the algorithms. This paper describes a matchmaking system, SemBiosphere, which solves both problems. It recommends clustering algorithms based on some minimal user requirement inputs and the data properties. An ontology was developed in OWL, an expressive ontological language, for describing what the algorithms are and how they perform, in addition to how they can be invoked. This allows machines to "understand" the algorithms and make the recommendations. The algorithm can be implemented by different groups and in different languages, and run on different platforms at geographically distributed sites. Through the use of XML-based web services, they can all be invoked in the same standard way. The current clustering services were transformed from the non-semantic web services of the Biosphere system, which includes a variety of algorithms that have been applied to microarray gene expression data analysis. New algorithms can be incorporated into the system without too much effort. The SemBiosphere system and the complete clustering ontology can be accessed at http://yeasthub2.gersteinlab. org/sembiosphere/.

  17. Service-Learning from a Distance: Partnering Multiple Universities and Local Governments in a Large Scale Initiative

    ERIC Educational Resources Information Center

    Poindexter, Sandra; Arnold, Pamela; Osterhout, Christopher

    2009-01-01

    Service-learning can be academically effective even when the distances between students and client organizations prevent face-to-face interchanges and site visits. Working with the State of Michigan and Michigan Townships Association, Michigan students from five universities learned about local government while helping Michigan townships develop…

  18. A Survey of Use and Evaluation of Support Services by External Students, 1983. Working Papers in Distance Education, No. 10.

    ERIC Educational Resources Information Center

    Kelly, Mavis E.

    This report presents the results of a 1983 survey of external (distance education) students' attitudes toward and use of publications and study materials issued by the School of External Studies and Continuing Education at the University of Queensland. The use and evaluation of regional and central support services were also studied. Data from the…

  19. Achieving the Impossible? Teaching Practice Component of a Pre-Service Distance English Language Teacher Training Program in Turkey

    ERIC Educational Resources Information Center

    Kecik, Ilknur; Aydin, Belgin

    2011-01-01

    The aim of this article is to describe the model developed for the teaching practice component of the pre-service Distance English Language Teacher Training Program (DELTT) at Anadolu University, Eskisehir, Turkey. The steps taken to improve the model over a six-year period will be explained and the recent developments in the teaching practice…

  20. Service Teachers' Academic Achievements in Online Distance Education: The Roles of Online Self-Regulation and Attitudes

    ERIC Educational Resources Information Center

    Uzun, Ahmet Murat; Unal, Erhan; Yamac, Ahmet

    2013-01-01

    The purpose of this study is to examine pre-service teachers' academic achievements in terms of different variables such as online self-regulated learning skills and attitudes towards distance education. This study is descriptive in nature. Survey study design was implemented to examine the relationships among variables. The study was…

  1. Listen to What They Have to Say! Assessing Distance Learners' Satisfaction with Library Services Using a Transactional Survey

    ERIC Educational Resources Information Center

    Alewine, Michael C.

    2012-01-01

    This paper examines the evolution and findings of an on-going longitudinal study that is assessing the satisfaction of distance education students with library reference services through the use of a transaction-level survey. The survey's purpose is two-fold: first, it is used to garner valuable input from these students; and second, it also…

  2. Evaluation of the Distance Education Pre-Service Teachers' Opinions about Teaching Practice Course (Case of Izmir City)

    ERIC Educational Resources Information Center

    Guven, Meral; Kurum, Dilruba; Saglam, Mustafa

    2012-01-01

    The aim of this study was to determine the distance education pre-service teachers' opinions about the teaching practice course. The study was conducted with descriptive method. For data collection, analysis and interpretation, qualitative research method was used. Out of the students enrolled at Open Education Faculty, Department of Pre-school…

  3. Clever generation of rich SPARQL queries from annotated relational schema: application to Semantic Web Service creation for biological databases

    PubMed Central

    2013-01-01

    Background In recent years, a large amount of “-omics” data have been produced. However, these data are stored in many different species-specific databases that are managed by different institutes and laboratories. Biologists often need to find and assemble data from disparate sources to perform certain analyses. Searching for these data and assembling them is a time-consuming task. The Semantic Web helps to facilitate interoperability across databases. A common approach involves the development of wrapper systems that map a relational database schema onto existing domain ontologies. However, few attempts have been made to automate the creation of such wrappers. Results We developed a framework, named BioSemantic, for the creation of Semantic Web Services that are applicable to relational biological databases. This framework makes use of both Semantic Web and Web Services technologies and can be divided into two main parts: (i) the generation and semi-automatic annotation of an RDF view; and (ii) the automatic generation of SPARQL queries and their integration into Semantic Web Services backbones. We have used our framework to integrate genomic data from different plant databases. Conclusions BioSemantic is a framework that was designed to speed integration of relational databases. We present how it can be used to speed the development of Semantic Web Services for existing relational biological databases. Currently, it creates and annotates RDF views that enable the automatic generation of SPARQL queries. Web Services are also created and deployed automatically, and the semantic annotations of our Web Services are added automatically using SAWSDL attributes. BioSemantic is downloadable at http://southgreen.cirad.fr/?q=content/Biosemantic. PMID:23586394

  4. Designing and Implementing WebCT-Based Courses Online: Distance English Language Teacher Training Program (DELTT) Model

    ERIC Educational Resources Information Center

    Kurubacak, Gulsun

    2002-01-01

    Many undergraduate and graduate courses in Open Education Faculty (OEF) of Anadolu University in Turkey have been delivered for over the two decades. Most of distance programs have been distributed via traditional distance education approaches and philosophy, such as TV and radio programs, printed materials, etc. Today, however, OEF has been…

  5. The Evolution of Distance Education: Implications for Instructional Design on the Potential of the Web. Part 3: K-12

    ERIC Educational Resources Information Center

    Huett, Jason; Moller, Leslie; Foshay, Wellesley R.; Coleman, Craig

    2008-01-01

    This article is the last of a three-part series on distance education. Part 1 focused on Training and Development; Part 2 on Higher Education. Distance education in the K-12 arena is often referred to as "virtual schooling" and learning through virtual schooling is one of the fastest growing areas for K-12 schools. Virtual schools offer distance…

  6. Efficient Authorization of Rich Presence Using Secure and Composed Web Services

    NASA Astrophysics Data System (ADS)

    Li, Li; Chou, Wu

    This paper presents an extended Role-Based Access Control (RBAC) model for efficient authorization of rich presence using secure web services composed with an abstract presence data model. Following the information symmetry principle, the standard RBAC model is extended to support context sensitive social relations and cascaded authority. In conjunction with the extended RBAC model, we introduce an extensible presence architecture prototype using WS-Security and WS-Eventing to secure rich presence information exchanges based on PKI certificates. Applications and performance measurements of our presence system are presented to show that the proposed RBAC framework for presence and collaboration is well suited for real-time communication and collaboration.

  7. Professional's Information Link (PiL): a web-based asynchronous consultation service.

    PubMed Central

    Hersh, William; Miller, Robin; Olson, Daniel; Sacherek, Lynetta; Cross, Ping

    2002-01-01

    Despite the massive growth of on-line information resources, many clinicians still prefer to obtain information from other clinicians, particularly those to whom they refer patients. Based on this notion and an information needs assessment, we developed the Professional's Information Link (PiL), a Web-based asynchronous consultation service to facilitate question answering between rural clinicians and an academic medical center. The system aims to provide answers to questions within two working days. It also includes patient handouts and continuing medical education resources. Our preliminary evaluation demonstrates modest but enthusiastic use of the system. PMID:12463840

  8. The National 3-D Geospatial Information Web-Based Service of Korea

    NASA Astrophysics Data System (ADS)

    Lee, D. T.; Kim, C. W.; Kang, I. G.

    2013-09-01

    3D geospatial information systems should provide efficient spatial analysis tools and able to use all capabilities of the third dimension, and a visualization. Currently, many human activities make steps toward the third dimension like land use, urban and landscape planning, cadastre, environmental monitoring, transportation monitoring, real estate market, military applications, etc. To reflect this trend, the Korean government has been started to construct the 3D geospatial data and service platform. Since the geospatial information was introduced in Korea, the construction of geospatial information (3D geospatial information, digital maps, aerial photographs, ortho photographs, etc.) has been led by the central government. The purpose of this study is to introduce the Korean government-lead 3D geospatial information web-based service for the people who interested in this industry and we would like to introduce not only the present conditions of constructed 3D geospatial data but methodologies and applications of 3D geospatial information. About 15% (about 3,278.74 km2) of the total urban area's 3D geospatial data have been constructed by the national geographic information institute (NGII) of Korea from 2005 to 2012. Especially in six metropolitan cities and Dokdo (island belongs to Korea) on level of detail (LOD) 4 which is photo-realistic textured 3D models including corresponding ortho photographs were constructed in 2012. In this paper, we represented web-based 3D map service system composition and infrastructure and comparison of V-world with Google Earth service will be presented. We also represented Open API based service cases and discussed about the protection of location privacy when we construct 3D indoor building models. In order to prevent an invasion of privacy, we processed image blurring, elimination and camouflage. The importance of public-private cooperation and advanced geospatial information policy is emphasized in Korea. Thus, the progress of

  9. The DBCLS BioHackathon: standardization and interoperability for bioinformatics web services and workflows. The DBCLS BioHackathon Consortium*.

    PubMed

    Katayama, Toshiaki; Arakawa, Kazuharu; Nakao, Mitsuteru; Ono, Keiichiro; Aoki-Kinoshita, Kiyoko F; Yamamoto, Yasunori; Yamaguchi, Atsuko; Kawashima, Shuichi; Chun, Hong-Woo; Aerts, Jan; Aranda, Bruno; Barboza, Lord Hendrix; Bonnal, Raoul Jp; Bruskiewich, Richard; Bryne, Jan C; Fernández, José M; Funahashi, Akira; Gordon, Paul Mk; Goto, Naohisa; Groscurth, Andreas; Gutteridge, Alex; Holland, Richard; Kano, Yoshinobu; Kawas, Edward A; Kerhornou, Arnaud; Kibukawa, Eri; Kinjo, Akira R; Kuhn, Michael; Lapp, Hilmar; Lehvaslaiho, Heikki; Nakamura, Hiroyuki; Nakamura, Yasukazu; Nishizawa, Tatsuya; Nobata, Chikashi; Noguchi, Tamotsu; Oinn, Thomas M; Okamoto, Shinobu; Owen, Stuart; Pafilis, Evangelos; Pocock, Matthew; Prins, Pjotr; Ranzinger, René; Reisinger, Florian; Salwinski, Lukasz; Schreiber, Mark; Senger, Martin; Shigemoto, Yasumasa; Standley, Daron M; Sugawara, Hideaki; Tashiro, Toshiyuki; Trelles, Oswaldo; Vos, Rutger A; Wilkinson, Mark D; York, William; Zmasek, Christian M; Asai, Kiyoshi; Takagi, Toshihisa

    2010-01-01

    Web services have become a key technology for bioinformatics, since life science databases are globally decentralized and the exponential increase in the amount of available data demands for efficient systems without the need to transfer entire databases for every step of an analysis. However, various incompatibilities among database resources and analysis services make it difficult to connect and integrate these into interoperable workflows. To resolve this situation, we invited domain specialists from web service providers, client software developers, Open Bio* projects, the BioMoby project and researchers of emerging areas where a standard exchange data format is not well established, for an intensive collaboration entitled the BioHackathon 2008. The meeting was hosted by the Database Center for Life Science (DBCLS) and Computational Biology Research Center (CBRC) and was held in Tokyo from February 11th to 15th, 2008. In this report we highlight the work accomplished and the common issues arisen from this event, including the standardization of data exchange formats and services in the emerging fields of glycoinformatics, biological interaction networks, text mining, and phyloinformatics. In addition, common shared object development based on BioSQL, as well as technical challenges in large data management, asynchronous services, and security are discussed. Consequently, we improved interoperability of web services in several fields, however, further cooperation among major database centers and continued collaborative efforts between service providers and software developers are still necessary for an effective advance in bioinformatics web service technologies. PMID:20727200

  10. The DBCLS BioHackathon: standardization and interoperability for bioinformatics web services and workflows. The DBCLS BioHackathon Consortium*

    PubMed Central

    2010-01-01

    Web services have become a key technology for bioinformatics, since life science databases are globally decentralized and the exponential increase in the amount of available data demands for efficient systems without the need to transfer entire databases for every step of an analysis. However, various incompatibilities among database resources and analysis services make it difficult to connect and integrate these into interoperable workflows. To resolve this situation, we invited domain specialists from web service providers, client software developers, Open Bio* projects, the BioMoby project and researchers of emerging areas where a standard exchange data format is not well established, for an intensive collaboration entitled the BioHackathon 2008. The meeting was hosted by the Database Center for Life Science (DBCLS) and Computational Biology Research Center (CBRC) and was held in Tokyo from February 11th to 15th, 2008. In this report we highlight the work accomplished and the common issues arisen from this event, including the standardization of data exchange formats and services in the emerging fields of glycoinformatics, biological interaction networks, text mining, and phyloinformatics. In addition, common shared object development based on BioSQL, as well as technical challenges in large data management, asynchronous services, and security are discussed. Consequently, we improved interoperability of web services in several fields, however, further cooperation among major database centers and continued collaborative efforts between service providers and software developers are still necessary for an effective advance in bioinformatics web service technologies. PMID:20727200

  11. A History of the NASA Planetary Data System (PDS) Imaging Node's Map-A-Planet Legacy Web Services

    NASA Astrophysics Data System (ADS)

    Garcia, P. A.; Isbell, C. E.; Gaddis, L. R.

    2015-06-01

    NASA Planetary Data System (PDS) Imaging Node’s Map-A-Planet Legacy Web Services have served the planetary data community for more than fifteen years. Here we look back at the evolution and development of the services over the that time.

  12. Exploring Pre-Service Teachers' Beliefs about Using Web 2.0 Technologies in K-12 Classroom

    ERIC Educational Resources Information Center

    Sadaf, Ayesha; Newby, Timothy J.; Ertmer, Peggy A.

    2012-01-01

    This qualitative study explored pre-service teachers' behavioral, normative, and control beliefs regarding their intentions to use Web 2.0 technologies in their future classrooms. The Theory of Planned Behavior (TPB) was used as the theoretical framework (Ajzen, 1991) to understand these beliefs and pre-service teachers' intentions for why they…

  13. Effective electron-density map improvement and structure validation on a Linux multi-CPU web cluster: The TB Structural Genomics Consortium Bias Removal Web Service.

    PubMed

    Reddy, Vinod; Swanson, Stanley M; Segelke, Brent; Kantardjieff, Katherine A; Sacchettini, James C; Rupp, Bernhard

    2003-12-01

    Anticipating a continuing increase in the number of structures solved by molecular replacement in high-throughput crystallography and drug-discovery programs, a user-friendly web service for automated molecular replacement, map improvement, bias removal and real-space correlation structure validation has been implemented. The service is based on an efficient bias-removal protocol, Shake&wARP, and implemented using EPMR and the CCP4 suite of programs, combined with various shell scripts and Fortran90 routines. The service returns improved maps, converted data files and real-space correlation and B-factor plots. User data are uploaded through a web interface and the CPU-intensive iteration cycles are executed on a low-cost Linux multi-CPU cluster using the Condor job-queuing package. Examples of map improvement at various resolutions are provided and include model completion and reconstruction of absent parts, sequence correction, and ligand validation in drug-target structures.

  14. A flexible statistics web processing service--added value for information systems for experiment data.

    PubMed

    Heimann, Dennis; Nieschulze, Jens; König-Ries, Birgitta

    2010-01-01

    Data management in the life sciences has evolved from simple storage of data to complex information systems providing additional functionalities like analysis and visualization capabilities, demanding the integration of statistical tools. In many cases the used statistical tools are hard-coded within the system. That leads to an expensive integration, substitution, or extension of tools because all changes have to be done in program code. Other systems are using generic solutions for tool integration but adapting them to another system is mostly rather extensive work. This paper shows a way to provide statistical functionality over a statistics web service, which can be easily integrated in any information system and set up using XML configuration files. The statistical functionality is extendable by simply adding the description of a new application to a configuration file. The service architecture as well as the data exchange process between client and service and the adding of analysis applications to the underlying service provider are described. Furthermore a practical example demonstrates the functionality of the service. PMID:20410555

  15. A Proposal for a Thesaurus for Web Services in Solar Radiation

    NASA Technical Reports Server (NTRS)

    Gschwind, Benoit; Menard, Lionel; Ranchin, Thierry; Wald, Lucien; Stackhouse, Paul W., Jr.

    2007-01-01

    Metadata are necessary to discover, describe and exchange any type of information, resource and service at a large scale. A significant amount of effort has been made in the field of geography and environment to establish standards. Efforts still remain to address more specific domains such as renewable energies. This communication focuses on solar energy and more specifically on aspects in solar radiation that relate to geography and meteorology. A thesaurus in solar radiation is proposed for the keys elements in solar radiation namely time, space and radiation types. The importance of time-series in solar radiation is outlined and attributes of the key elements are discussed. An XML schema for encoding metadata is proposed. The exploitation of such a schema in web services is discussed. This proposal is a first attempt at establishing a thesaurus for describing data and applications in solar radiation.

  16. Metadata research and design of ocean color remote sensing data based on web service

    NASA Astrophysics Data System (ADS)

    Kang, Yan; Pan, Delu; He, Xianqiang; Wang, Difeng; Chen, Jianyu

    2010-10-01

    The ocean color remote sensing metadata describes the content, quality, condition, and other characteristics of ocean color remote sensing data. Paper presents a metadata standard draft based on XML, and gives the details of main ocean color remote sensing metadata XML elements. The ocean color remote sensing data platform-sharing is in developments as a part of the digital ocean system, on this basis, the ocean color remote sensing metadata directory service system based on web service is put forward, which aims to store and manage the ocean color remote sensing metadata effectively. The metadata of the ocean color remote sensing data become the most important event for the ocean color remote sensing information more retrieved and used.

  17. Middleware and Web Services for the Collaborative Information Portal of NASA's Mars Exploration Rovers Mission

    NASA Technical Reports Server (NTRS)

    Sinderson, Elias; Magapu, Vish; Mak, Ronald

    2004-01-01

    We describe the design and deployment of the middleware for the Collaborative Information Portal (CIP), a mission critical J2EE application developed for NASA's 2003 Mars Exploration Rover mission. CIP enabled mission personnel to access data and images sent back from Mars, staff and event schedules, broadcast messages and clocks displaying various Earth and Mars time zones. We developed the CIP middleware in less than two years time usins cutting-edge technologies, including EJBs, servlets, JDBC, JNDI and JMS. The middleware was designed as a collection of independent, hot-deployable web services, providing secure access to back end file systems and databases. Throughout the middleware we enabled crosscutting capabilities such as runtime service configuration, security, logging and remote monitoring. This paper presents our approach to mitigating the challenges we faced, concluding with a review of the lessons we learned from this project and noting what we'd do differently and why.

  18. 2D Maps, 3D Globes, and OGC Web Services Supporting Arctic Science through the Arctic Research Mapping Application (ARMAP)

    NASA Astrophysics Data System (ADS)

    Johnson, G. W.; Gaylord, A. G.; Brady, J.; Cody, R.; Ramirez, G.; Gonzalez, J. C.; Rubio, C.; Dover, M.; Garcia-Lavigne, D.; Manley, W.; Score, R.; Tweedie, C.

    2008-12-01

    The Arctic Research Mapping Application (ARMAP) is a suite of online services designed to provide support for Arctic science. These services include: a text based online search utility, 2D Internet Map Server (IMS), 3D globe applications (Google Earth and ArcGIS Explorer), Open Geospatial Consortium (OGC) Web Map Service (WMS) and Keyhole Markup Language (KML) Service , and a prototype 2D ArcGIS Server Web Mapping Application (WMA). Avoiding a duplication of effort has been a primary objective of the ARMAP project which incorporates best practices (e.g. OGC standard web services and metadata) and off the shelf technologies. The Arctic Research Logistics Support Service (ARLSS) database is the foundation of all the ARMAP services and includes US research funded by the National Science Foundation, National Aeronautics and Space Administration and National Oceanic and Atmospheric Administration. With ARMAP's 2D maps and 3D globes, users can navigate to areas of interest, view a variety of map layers, and explore U.S. federally funded research projects. Projects can be queried by location, year, funding program, discipline, and keyword. Links to specific information and other web sites associated with particular research projects are included. . The ARMAP suite provides tools for users of various levels of technical ability to interact with data by running text based queries, browsing in 2D or 3D, or importing the KML and OGC web services directly into their own GIS applications and virtual globes. With special emphasis on the International Polar Year (IPY), ARMAP has targeted science planners, scientists, educators, and the general public. In sum, ARMAP goes beyond a simple map display to enable analysis, synthesis, and coordination of Arctic research. ARMAP may be accessed via the gateway web site at http://www.armap.org.

  19. Effecting Change in In-Service Teachers through Distance Education and Communication Technology.

    ERIC Educational Resources Information Center

    Knapczyk, Dennis R.; Rodes, Paul

    1995-01-01

    Describes a cooperative program that uses low-cost communication technology and distance education to give inservice teachers access to information on teaching practices, and to provide a vehicle for improving the quality of teacher preparation and educational programs. Discusses choosing appropriate technology for distance education and…

  20. Student Perceptions of Support Services and the Influence of Targeted Interventions on Retention in Distance Education

    ERIC Educational Resources Information Center

    Nichols, Mark

    2010-01-01

    To improve student retention in distance education, Simpson suggested in 2003 that institutions analyse their own retention characteristics and "spot the leaks." In 2008 the Centre for Distance Learning at Laidlaw College, New Zealand, employed two part-time academic support coordinators in an effort to improve student retention and success. This…

  1. The Role of Student Affairs in Distance Education: Cyber-Services or Virtual Communities

    ERIC Educational Resources Information Center

    Kretovics, Mark

    2003-01-01

    As distance education technology enables institutions of higher education to offer courses to students throughout the country, it is important for student affairs to offer opportunities for these students to connect with the institution. This article reviews the relevant literature on distance education and discusses differences between providing…

  2. Idiographic Roles of Cooperating Teachers as Mentors in Pre-Service Distance Teacher Education

    ERIC Educational Resources Information Center

    Koc, Ebru Melek

    2012-01-01

    This study aims to define the roles of cooperating teachers as mentors in the context of distance-learning teacher education. The participants included 358 cooperating teachers who mentored 4th-year student teachers in a Distance English Language Teacher Training Program in Turkey. To determine the roles that were perceived as mentoring roles by…

  3. Grid enablement of OpenGeospatial Web Services: the G-OWS Working Group

    NASA Astrophysics Data System (ADS)

    Mazzetti, Paolo

    2010-05-01

    In last decades two main paradigms for resource sharing emerged and reached maturity: the Web and the Grid. They both demonstrate suitable for building Distributed Computing Infrastructures (DCIs) supporting the coordinated sharing of resources (i.e. data, information, services, etc) on the Internet. Grid and Web DCIs have much in common as a result of their underlying Internet technology (protocols, models and specifications). However, being based on different requirements and architectural approaches, they show some differences as well. The Web's "major goal was to be a shared information space through which people and machines could communicate" [Berners-Lee 1996]. The success of the Web, and its consequent pervasiveness, made it appealing for building specialized systems like the Spatial Data Infrastructures (SDIs). In this systems the introduction of Web-based geo-information technologies enables specialized services for geospatial data sharing and processing. The Grid was born to achieve "flexible, secure, coordinated resource sharing among dynamic collections of individuals, institutions, and resources" [Foster 2001]. It specifically focuses on large-scale resource sharing, innovative applications, and, in some cases, high-performance orientation. In the Earth and Space Sciences (ESS) the most part of handled information is geo-referred (geo-information) since spatial and temporal meta-information is of primary importance in many application domains: Earth Sciences, Disasters Management, Environmental Sciences, etc. On the other hand, in several application areas there is the need of running complex models which require the large processing and storage capabilities that the Grids are able to provide. Therefore the integration of geo-information and Grid technologies might be a valuable approach in order to enable advanced ESS applications. Currently both geo-information and Grid technologies have reached a high level of maturity, allowing to build such an

  4. INTAMAP: The design and implementation of an interoperable automated interpolation web service

    NASA Astrophysics Data System (ADS)

    Pebesma, Edzer; Cornford, Dan; Dubois, Gregoire; Heuvelink, Gerard B. M.; Hristopulos, Dionisis; Pilz, Jürgen; Stöhlker, Ulrich; Morin, Gary; Skøien, Jon O.

    2011-03-01

    INTAMAP is a Web Processing Service for the automatic spatial interpolation of measured point data. Requirements were (i) using open standards for spatial data such as developed in the context of the Open Geospatial Consortium (OGC), (ii) using a suitable environment for statistical modelling and computation, and (iii) producing an integrated, open source solution. The system couples an open-source Web Processing Service (developed by 52°North), accepting data in the form of standardised XML documents (conforming to the OGC Observations and Measurements standard) with a computing back-end realised in the R statistical environment. The probability distribution of interpolation errors is encoded with UncertML, a markup language designed to encode uncertain data. Automatic interpolation needs to be useful for a wide range of applications and the algorithms have been designed to cope with anisotropy, extreme values, and data with known error distributions. Besides a fully automatic mode, the system can be used with different levels of user control over the interpolation process.

  5. The NORM technology connection web site : streamlined access to NORM-related service company and regulatory information.

    SciTech Connect

    Smith, K. P.; Richmond, P.; LePoire, D. J.; Arnish, J. J.; Johnson, R.

    2000-11-08

    Argonne National Laboratory has developed an Internet web site providing access to critical information needed to support decisions on the management and disposal of wastes containing naturally occurring radioactive material (NORM). The NORM Technology Connection web site provides current information on (1) service companies that provide support on NORM issues (e.g., site characterization and remediation, sample analysis, radiation safety training, disposal) and (2) existing applicable NORM regulations and guidelines. A third element of the site is an electronic mail list that allows users to post or respond to questions about the management of NORM. Development of the NORM Technology Connection web site was funded by the U.S. Department of Energy, Office of Fossil Energy. It is hosted and maintained by the Interstate Oil and Gas Compact Commission. The web site is publicly available; access is free, as is participation by any of the service companies.

  6. AWSCS-A System to Evaluate Different Approaches for the Automatic Composition and Execution of Web Services Flows

    PubMed Central

    Tardiole Kuehne, Bruno; Estrella, Julio Cezar; Nunes, Luiz Henrique; Martins de Oliveira, Edvard; Hideo Nakamura, Luis; Gomes Ferreira, Carlos Henrique; Carlucci Santana, Regina Helena; Reiff-Marganiec, Stephan; Santana, Marcos José

    2015-01-01

    This paper proposes a system named AWSCS (Automatic Web Service Composition System) to evaluate different approaches for automatic composition of Web services, based on QoS parameters that are measured at execution time. The AWSCS is a system to implement different approaches for automatic composition of Web services and also to execute the resulting flows from these approaches. Aiming at demonstrating the results of this paper, a scenario was developed, where empirical flows were built to demonstrate the operation of AWSCS, since algorithms for automatic composition are not readily available to test. The results allow us to study the behaviour of running composite Web services, when flows with the same functionality but different problem-solving strategies were compared. Furthermore, we observed that the influence of the load applied on the running system as the type of load submitted to the system is an important factor to define which approach for the Web service composition can achieve the best performance in production. PMID:26068216

  7. Enhancing the Teaching of Digital Processing of Remote Sensing Image Course through Geospatial Web Processing Services

    NASA Astrophysics Data System (ADS)

    di, L.; Deng, M.

    2010-12-01

    Remote sensing (RS) is an essential method to collect data for Earth science research. Huge amount of remote sensing data, most of them in the image form, have been acquired. Almost all geography departments in the world offer courses in digital processing of remote sensing images. Such courses place emphasis on how to digitally process large amount of multi-source images for solving real world problems. However, due to the diversity and complexity of RS images and the shortcomings of current data and processing infrastructure, obstacles for effectively teaching such courses still remain. The major obstacles include 1) difficulties in finding, accessing, integrating and using massive RS images by students and educators, and 2) inadequate processing functions and computing facilities for students to freely explore the massive data. Recent development in geospatial Web processing service systems, which make massive data, computing powers, and processing capabilities to average Internet users anywhere in the world, promises the removal of the obstacles. The GeoBrain system developed by CSISS is an example of such systems. All functions available in GRASS Open Source GIS have been implemented as Web services in GeoBrain. Petabytes of remote sensing images in NASA data centers, the USGS Landsat data archive, and NOAA CLASS are accessible transparently and processable through GeoBrain. The GeoBrain system is operated on a high performance cluster server with large disk storage and fast Internet connection. All GeoBrain capabilities can be accessed by any Internet-connected Web browser. Dozens of universities have used GeoBrain as an ideal platform to support data-intensive remote sensing education. This presentation gives a specific example of using GeoBrain geoprocessing services to enhance the teaching of GGS 588, Digital Remote Sensing taught at the Department of Geography and Geoinformation Science, George Mason University. The course uses the textbook "Introductory

  8. Collaborative Science Using Web Services and the SciFlo Grid Dataflow Engine

    NASA Astrophysics Data System (ADS)

    Wilson, B. D.; Manipon, G.; Xing, Z.; Yunck, T.

    2006-12-01

    The General Earth Science Investigation Suite (GENESIS) project is a NASA-sponsored partnership between the Jet Propulsion Laboratory, academia, and NASA data centers to develop a new suite of Web Services tools to facilitate multi-sensor investigations in Earth System Science. The goal of GENESIS is to enable large-scale, multi-instrument atmospheric science using combined datasets from the AIRS, MODIS, MISR, and GPS sensors. Investigations include cross-comparison of spaceborne climate sensors, cloud spectral analysis, study of upper troposphere-stratosphere water transport, study of the aerosol indirect cloud effect, and global climate model validation. The challenges are to bring together very large datasets, reformat and understand the individual instrument retrievals, co-register or re-grid the retrieved physical parameters, perform computationally-intensive data fusion and data mining operations, and accumulate complex statistics over months to years of data. To meet these challenges, we have developed a Grid computing and dataflow framework, named SciFlo, in which we are deploying a set of versatile and reusable operators for data access, subsetting, registration, mining, fusion, compression, and advanced statistical analysis. SciFlo leverages remote Web Services, called via Simple Object Access Protocol (SOAP) or REST (one-line) URLs, and the Grid Computing standards (WS-* &Globus Alliance toolkits), and enables scientists to do multi-instrument Earth Science by assembling reusable Web Services and native executables into a distributed computing flow (tree of operators). The SciFlo client &server engines optimize the execution of such distributed data flows and allow the user to transparently find and use datasets and operators without worrying about the actual location of the Grid resources. In particular, SciFlo exploits the wealth of datasets accessible by OpenGIS Consortium (OGC) Web Mapping Servers & Web Coverage Servers (WMS/WCS), and by Open Data

  9. Analysis of a Customized Intervention for the Development of a Web-Based Lesson by Pre-Service Teachers.

    ERIC Educational Resources Information Center

    Murry, G. Brandon; Murry, Francie R.

    This study compared the use of two developmental alternatives: a Web Editor (WE) in combination with a customized template/shell (Teaching Not Teaching, T-N-T) and a WE only, for development of a Web-based lesson by pre-service teachers. Six hypotheses were tested to find whether the WE and T-N-T alternative was more efficient, effective, and…

  10. Improving the quality of e-commerce web service: what is important for the request scheduling algorithm?

    NASA Astrophysics Data System (ADS)

    Suchacka, Grazyna

    2005-02-01

    The paper concerns a new research area that is Quality of Web Service (QoWS). The need for QoWS is motivated by a still growing number of Internet users, by a steady development and diversification of Web services, and especially by popularization of e-commerce applications. The goal of the paper is a critical analysis of the literature concerning scheduling algorithms for e-commerce Web servers. The paper characterizes factors affecting the load of the Web servers and discusses ways of improving their efficiency. Crucial QoWS requirements of the business Web server are identified: serving requests before their individual deadlines, supporting user session integrity, supporting different classes of users and minimizing a number of rejected requests. It is justified that meeting these requirements and implementing them in an admission control (AC) and scheduling algorithm for the business Web server is crucial to the functioning of e-commerce Web sites and revenue generated by them. The paper presents results of the literature analysis and discusses algorithms that implement these important QoWS requirements. The analysis showed that very few algorithms take into consideration the above mentioned factors and that there is a need for designing an algorithm implementing them.

  11. IRRIMET: a web 2.0 advisory service for irrigation water management

    NASA Astrophysics Data System (ADS)

    De Michele, Carlo; Anzano, Enrico; Colandrea, Marco; Marotta, Luigi; Mula, Ileana; Pelosi, Anna; D'Urso, Guido; Battista Chirico, Giovanni

    2016-04-01

    Irrigation agriculture is one the biggest consumer of water in Europe, especially in southern regions, where it accounts for up to 70% of the total water consumption. The EU Common Agricultural Policy, combined with the Water Framework Directive, imposes to farmers and irrigation managers a substantial increase of the efficiency in the use of water in agriculture for the next decade. Irrigating according to reliable crop water requirement estimates is one of the most convincing solution to decrease agricultural water use. Here we present an innovative irrigation advisory service, applied in Campania region (Southern Italy), where a satellite assisted irrigation advisory service has been operating since 2006. The advisory service is based on the optimal combination of VIS-NIR high resolution satellite images (Landsat, Deimos, Rapideye) to map crop vigour, and high resolution numerical weather prediction for assessing the meteorological variables driving the crop water needs in the short-medium range. The advisory service is broadcasted with a simple and intuitive web app interface which makes daily real time irrigation and evapotranspiration maps and customized weather forecasts (based on Cosmo Leps model) accessible from desktop computers, tablets and smartphones.

  12. Hydrological models as web services: Experiences from the Environmental Virtual Observatory project

    NASA Astrophysics Data System (ADS)

    Buytaert, W.; Vitolo, C.; Reaney, S. M.; Beven, K.

    2012-12-01

    Data availability in environmental sciences is expanding at a rapid pace. From the constant stream of high-resolution satellite images to the local efforts of citizen scientists, there is an increasing need to process the growing stream of heterogeneous data and turn it into useful information for decision-making. Environmental models, ranging from simple rainfall - runoff relations to complex climate models, can be very useful tools to process data, identify patterns, and help predict the potential impact of management scenarios. Recent technological innovations in networking, computing and standardization may bring a new generation of interactive models plugged into virtual environments closer to the end-user. They are the driver of major funding initiatives such as the UK's Virtual Observatory program, and the U.S. National Science Foundation's Earth Cube. In this study we explore how hydrological models, being an important subset of environmental models, have to be adapted in order to function within a broader environment of web-services and user interactions. Historically, hydrological models have been developed for very different purposes. Typically they have a rigid model structure, requiring a very specific set of input data and parameters. As such, the process of implementing a model for a specific catchment requires careful collection and preparation of the input data, extensive calibration and subsequent validation. This procedure seems incompatible with a web-environment, where data availability is highly variable, heterogeneous and constantly changing in time, and where the requirements of end-users may be not necessarily align with the original intention of the model developer. We present prototypes of models that are web-enabled using the web standards of the Open Geospatial Consortium, and implemented in online decision-support systems. We identify issues related to (1) optimal use of available data; (2) the need for flexible and adaptive structures

  13. Implementation of Web Processing Services (WPS) over IPSL Earth System Grid Federation (ESGF) node

    NASA Astrophysics Data System (ADS)

    Kadygrov, Nikolay; Denvil, Sebastien; Carenton, Nicolas; Levavasseur, Guillaume; Hempelmann, Nils; Ehbrecht, Carsten

    2016-04-01

    The Earth System Grid Federation (ESGF) is aimed to provide access to climate data for the international climate community. ESGF is a system of distributed and federated nodes that dynamically interact with each other. ESGF user may search and download climatic data, geographically distributed over the world, from one common web interface and through standardized API. With the continuous development of the climate models and the beginning of the sixth phase of the Coupled Model Intercomparison Project (CMIP6), the amount of data available from ESGF will continuously increase during the next 5 years. IPSL holds a replication of the different global and regional climate models output, observations and reanalysis data (CMIP5, CORDEX, obs4MIPs, etc) that are available on the IPSL ESGF node. In order to let scientists perform analysis of the models without downloading vast amount of data the Web Processing Services (WPS) were installed at IPSL compute node. The work is part of the CONVERGENCE project founded by French National Research Agency (ANR). PyWPS implementation of the Web processing Service standard from Open Geospatial Consortium (OGC) in the framework of birdhouse software is used. The processes could be run by user remotely through web-based WPS client or by using command-line tool. All the calculations are performed on the server side close to the data. If the models/observations are not available at IPSL it will be downloaded and cached by WPS process from ESGF network using synda tool. The outputs of the WPS processes are available for download as plots, tar-archives or as NetCDF files. We present the architecture of WPS at IPSL along with the processes for evaluation of the model performance, on-site diagnostics and post-analysis processing of the models output, e.g.: - regriding/interpolation/aggregation - ocgis (OpenClimateGIS) based polygon subsetting of the data - average seasonal cycle, multimodel mean, multimodel mean bias - calculation of the

  14. Provision of Distance Learning Services over Interactive Digital TV with MHP

    ERIC Educational Resources Information Center

    Pazos-Arias, Jose J.; Lopez-Nores, Martin; Garcia-Duque, Jorge; Diaz-Redondo, Rebeca P.; Blanco-Fernandez, Yolanda; Ramos-Cabrer, Manuel; Gil-Solla, Alberto; Fernandez-Vilas, Ana

    2008-01-01

    E-learning technologies have developed greatly in recent years, with considerable success. However, there is increasing evidence that web-based learning is not reaching the social sectors which are more reluctant to contact with the new technologies, thus leading to inequalities in the access to education and knowledge in the Information Society.…

  15. Integration between solar and space science data for space weather forecast using web services

    NASA Astrophysics Data System (ADS)

    Kato, S.

    2007-08-01

    As the technology develops, the opportunity that the human beings behave in space, and it is still understood that the solar activities (especially the solar flare) influence the airlines communication, the ship communication and the power generator of the electric power company, etc. Forecasting the effects of the solar activities is becoming very important because there is such a background. Our goal is that constructs the detailed model from the Sun to the magnetosphere of the earth and simulates the solar activities and the effects. We try to integrate the existing observational data including the ground observational data and satellite observational data using by web service technology as a base to construct the model. We introduce our activity to combine the solar and space science data in Japan. Methods Generally, it is difficult to develop the virtual common database, but web service makes interconnection among different databases comparatively easy. We try to connect some databases in the portal site. Each different data objects is aggregated to a common data object. We can develop more complex services. We use RELAX NG in order to develop these applications easily. We begin the trial of the interconnection among the solar and space science data in Japan. In the case of solar observational data, we find the activity such as VO, for example, VSO and EGSO, but space science data seems to be very complex. In addition to this, there is time lag that solar activity has an effect on the magnetosphere of the Earth. We discuss these characteristic in the data analysis between the solar and space data. This work was supported by the Grant-in-Aid for Creative Scientific Research `The Basic Study of Space Weather Prediction' (17GS0208) from the Ministry of Education, Science, Sports, Technology, and Culture of Japan

  16. Evaluation of flood hazard maps in print and web mapping services as information tools in flood risk communication

    NASA Astrophysics Data System (ADS)

    Hagemeier-Klose, M.; Wagner, K.

    2009-04-01

    Flood risk communication with the general public and the population at risk is getting increasingly important for flood risk management, especially as a precautionary measure. This is also underlined by the EU Flood Directive. The flood related authorities therefore have to develop adjusted information tools which meet the demands of different user groups. This article presents the formative evaluation of flood hazard maps and web mapping services according to the specific requirements and needs of the general public using the dynamic-transactional approach as a theoretical framework. The evaluation was done by a mixture of different methods; an analysis of existing tools, a creative workshop with experts and laymen and an online survey. The currently existing flood hazard maps or web mapping services or web GIS still lack a good balance between simplicity and complexity with adequate readability and usability for the public. Well designed and associative maps (e.g. using blue colours for water depths) which can be compared with past local flood events and which can create empathy in viewers, can help to raise awareness, to heighten the activity and knowledge level or can lead to further information seeking. Concerning web mapping services, a linkage between general flood information like flood extents of different scenarios and corresponding water depths and real time information like gauge levels is an important demand by users. Gauge levels of these scenarios are easier to understand than the scientifically correct return periods or annualities. The recently developed Bavarian web mapping service tries to integrate these requirements.

  17. McMaster Premium Literature Service (PLUS): An Evidence-based Medicine Information Service Delivered on the Web

    PubMed Central

    Holland, Jennifer; Haynes, R. Brian

    2005-01-01

    Background Evidence-based medicine (EBM) attempts to narrow the gap between knowledge and practice, but ready access to evidence-based resources remains a challenge to practicing physicians. Objective To describe a new, EBM information delivery service, its trial design, and baseline data. Methods McMaster PLUS (Premium Literature Service), composed of a continuously updated database and web-based interface, delivers scientifically rigorous and clinically relevant research literature matched to individual physicians’ clinical interests. A cluster randomized controlled trial is currently underway, comparing 2 versions of the PLUS system. Results As of Feb 2005, the PLUS database contained over 5700 scientifically sound, clinically relevant articles, published from 2001 to present. 68% of articles have full text links. Over 200 physicians in Northern Ontario have been randomized to one of 2 PLUS interfaces. Conclusion McMaster PLUS has been designed to aid physicians to home in on high quality research that is highly relevant and important to their own clinical practice. PMID:16779058

  18. Expanding access to gerontological education via distance learning: the Management of Aging Services Masters Program at UMass Boston.

    PubMed

    Nadash, Pamela; Miller, Edward Alan; Porell, Frank W; Birchander, Ellen; Glickman, Lillian; Burr, Jeffrey A

    2014-01-01

    This article describes the online Management of Aging Services Masters Program at the University of Massachusetts Boston and reports on a recent Program review. The Program has experienced rapid growth, evolving from seven matriculating students in 2003 to 108 in 2012. It has graduated 125 students and boasts a 78% completion rate. The authors describe the Program and report on faculty and student perceptions of performance. The Program demonstrates sound pedagogical practice for online education, incorporating techniques to foster community and encourage students and faculty interaction. Distance learning holds considerable promise for expanding access to gerontological education to reach future aging services professionals.

  19. Building web service interfaces to geoscience data sets: EarthCube GeoWS project activities at the IRIS DMC

    NASA Astrophysics Data System (ADS)

    Trabant, C. M.; Ahern, T. K.; Stults, M.

    2015-12-01

    At the IRIS Data Management Center (DMC) we have been developing web service data access interfaces for our, primarily seismological, repositories for five years. These interfaces have become the primary access mechanisms for all data extraction from the DMC. For the last two years the DMC has been a principal participant in the GeoWS project, which aims to develop common web service interfaces for data access across hydrology, geodesy, seismology, marine geophysics, atmospheric and other geoscience disciplines. By extending our approach we have converged, along with other project members, on a web service interface and presentation design appropriate for geoscience and other data. The key principles of the approach include using a simple subset of RESTful concepts, common calling conventions whenever possible, a common tabular text data set convention, human-readable documentation and tools to help scientific end users learn how to use the interfaces. The common tabular text format, called GeoCSV, has been incorporated into the DMC's seismic station and event (earthquake) services. In addition to modifying our existing services, we have developed prototype GeoCSV web services for data sets managed by external (unfunded) collaborators. These prototype services include interfaces for data sets at NGDC/NCEI (water level tides and meteorological satellite measurements), INTERMAGNET repository and UTEP gravity and magnetic measurements. In progress are interfaces for WOVOdat (volcano observatory measurements), NEON (ecological observatory measurements) and more. An important goal of our work is to build interfaces usable by non-technologist end users. We find direct usability by researchers to be a major factor in cross-discipline data use, which itself is a key to solving complex research questions. In addition to data discovery and collection by end users, these interfaces provide a foundation upon which federated data access and brokering systems are already being

  20. The 2nd DBCLS BioHackathon: interoperable bioinformatics Web services for integrated applications

    PubMed Central

    2011-01-01

    Background The interaction between biological researchers and the bioinformatics tools they use is still hampered by incomplete interoperability between such tools. To ensure interoperability initiatives are effectively deployed, end-user applications need to be aware of, and support, best practices and standards. Here, we report on an initiative in which software developers and genome biologists came together to explore and raise awareness of these issues: BioHackathon 2009. Results Developers in attendance came from diverse backgrounds, with experts in Web services, workflow tools, text mining and visualization. Genome biologists provided expertise and exemplar data from the domains of sequence and pathway analysis and glyco-informatics. One goal of the meeting was to evaluate the ability to address real world use cases in these domains using the tools that the developers represented. This resulted in i) a workflow to annotate 100,000 sequences from an invertebrate species; ii) an integrated system for analysis of the transcription factor binding sites (TFBSs) enriched based on differential gene expression data obtained from a microarray experiment; iii) a workflow to enumerate putative physical protein interactions among enzymes in a metabolic pathway using protein structure data; iv) a workflow to analyze glyco-gene-related diseases by searching for human homologs of glyco-genes in other species, such as fruit flies, and retrieving their phenotype-annotated SNPs. Conclusions Beyond deriving prototype solutions for each use-case, a second major purpose of the BioHackathon was to highlight areas of insufficiency. We discuss the issues raised by our exploration of the problem/solution space, concluding that there are still problems with the way Web services are modeled and annotated, including: i) the absence of several useful data or analysis functions in the Web service "space"; ii) the lack of documentation of methods; iii) lack of compliance with the SOAP