Sample records for large scale web

  1. Large-Scale Multiobjective Static Test Generation for Web-Based Testing with Integer Programming

    ERIC Educational Resources Information Center

    Nguyen, M. L.; Hui, Siu Cheung; Fong, A. C. M.

    2013-01-01

    Web-based testing has become a ubiquitous self-assessment method for online learning. One useful feature that is missing from today's web-based testing systems is the reliable capability to fulfill different assessment requirements of students based on a large-scale question data set. A promising approach for supporting large-scale web-based…

  2. Designing and developing portable large-scale JavaScript web applications within the Experiment Dashboard framework

    NASA Astrophysics Data System (ADS)

    Andreeva, J.; Dzhunov, I.; Karavakis, E.; Kokoszkiewicz, L.; Nowotka, M.; Saiz, P.; Tuckett, D.

    2012-12-01

    Improvements in web browser performance and web standards compliance, as well as the availability of comprehensive JavaScript libraries, provides an opportunity to develop functionally rich yet intuitive web applications that allow users to access, render and analyse data in novel ways. However, the development of such large-scale JavaScript web applications presents new challenges, in particular with regard to code sustainability and team-based work. We present an approach that meets the challenges of large-scale JavaScript web application design and development, including client-side model-view-controller architecture, design patterns, and JavaScript libraries. Furthermore, we show how the approach leads naturally to the encapsulation of the data source as a web API, allowing applications to be easily ported to new data sources. The Experiment Dashboard framework is used for the development of applications for monitoring the distributed computing activities of virtual organisations on the Worldwide LHC Computing Grid. We demonstrate the benefits of the approach for large-scale JavaScript web applications in this context by examining the design of several Experiment Dashboard applications for data processing, data transfer and site status monitoring, and by showing how they have been ported for different virtual organisations and technologies.

  3. Macroscopic characterisations of Web accessibility

    NASA Astrophysics Data System (ADS)

    Lopes, Rui; Carriço, Luis

    2010-12-01

    The Web Science framework poses fundamental questions on the analysis of the Web, by focusing on how microscopic properties (e.g. at the level of a Web page or Web site) emerge into macroscopic properties and phenomena. One research topic on the analysis of the Web is Web accessibility evaluation, which centres on understanding how accessible a Web page is for people with disabilities. However, when framing Web accessibility evaluation on Web Science, we have found that existing research stays at the microscopic level. This article presents an experimental study on framing Web accessibility evaluation into Web Science's goals. This study resulted in novel accessibility properties of the Web not found at microscopic levels, as well as of Web accessibility evaluation processes themselves. We observed at large scale some of the empirical knowledge on how accessibility is perceived by designers and developers, such as the disparity of interpretations of accessibility evaluation tools warnings. We also found a direct relation between accessibility quality and Web page complexity. We provide a set of guidelines for designing Web pages, education on Web accessibility, as well as on the computational limits of large-scale Web accessibility evaluations.

  4. Robopedia: Leveraging Sensorpedia for Web-Enabled Robot Control

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Resseguie, David R

    There is a growing interest in building Internetscale sensor networks that integrate sensors from around the world into a single unified system. In contrast, robotics application development has primarily focused on building specialized systems. These specialized systems take scalability and reliability into consideration, but generally neglect exploring the key components required to build a large scale system. Integrating robotic applications with Internet-scale sensor networks will unify specialized robotics applications and provide answers to large scale implementation concerns. We focus on utilizing Internet-scale sensor network technology to construct a framework for unifying robotic systems. Our framework web-enables a surveillance robot smore » sensor observations and provides a webinterface to the robot s actuators. This lets robots seamlessly integrate into web applications. In addition, the framework eliminates most prerequisite robotics knowledge, allowing for the creation of general web-based robotics applications. The framework also provides mechanisms to create applications that can interface with any robot. Frameworks such as this one are key to solving large scale mobile robotics implementation problems. We provide an overview of previous Internetscale sensor networks, Sensorpedia (an ad-hoc Internet-scale sensor network), our framework for integrating robots with Sensorpedia, two applications which illustrate our frameworks ability to support general web-based robotic control, and offer experimental results that illustrate our framework s scalability, feasibility, and resource requirements.« less

  5. EPA Facilities and Regional Boundaries Service, US, 2012, US EPA, SEGS

    EPA Pesticide Factsheets

    This SEGS web service contains EPA facilities, EPA facilities labels, small- and large-scale versions of EPA region boundaries, and EPA region boundaries extended to the 200nm Exclusive Economic Zone (EEZ). Small scale EPA boundaries and boundaries extended to the EEZ render at scales of less than 5 million, large scale EPA boundaries draw at scales greater than or equal to 5 million. EPA facilities labels draw at scales greater than 2 million. Data used to create this web service are available as a separate download at the Secondary Linkage listed above. Full FGDC metadata records for each layer may be found by clicking the layer name in the web service table of contents (available through the online link provided above) and viewing the layer description. This SEGS dataset was produced by EPA through the Office of Environmental Information.

  6. Scale-free characteristics of random networks: the topology of the world-wide web

    NASA Astrophysics Data System (ADS)

    Barabási, Albert-László; Albert, Réka; Jeong, Hawoong

    2000-06-01

    The world-wide web forms a large directed graph, whose vertices are documents and edges are links pointing from one document to another. Here we demonstrate that despite its apparent random character, the topology of this graph has a number of universal scale-free characteristics. We introduce a model that leads to a scale-free network, capturing in a minimal fashion the self-organization processes governing the world-wide web.

  7. Bayesian analysis of the dynamic cosmic web in the SDSS galaxy survey

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Leclercq, Florent; Wandelt, Benjamin; Jasche, Jens, E-mail: florent.leclercq@polytechnique.org, E-mail: jasche@iap.fr, E-mail: wandelt@iap.fr

    Recent application of the Bayesian algorithm \\textsc(borg) to the Sloan Digital Sky Survey (SDSS) main sample galaxies resulted in the physical inference of the formation history of the observed large-scale structure from its origin to the present epoch. In this work, we use these inferences as inputs for a detailed probabilistic cosmic web-type analysis. To do so, we generate a large set of data-constrained realizations of the large-scale structure using a fast, fully non-linear gravitational model. We then perform a dynamic classification of the cosmic web into four distinct components (voids, sheets, filaments, and clusters) on the basis of themore » tidal field. Our inference framework automatically and self-consistently propagates typical observational uncertainties to web-type classification. As a result, this study produces accurate cosmographic classification of large-scale structure elements in the SDSS volume. By also providing the history of these structure maps, the approach allows an analysis of the origin and growth of the early traces of the cosmic web present in the initial density field and of the evolution of global quantities such as the volume and mass filling fractions of different structures. For the problem of web-type classification, the results described in this work constitute the first connection between theory and observations at non-linear scales including a physical model of structure formation and the demonstrated capability of uncertainty quantification. A connection between cosmology and information theory using real data also naturally emerges from our probabilistic approach. Our results constitute quantitative chrono-cosmography of the complex web-like patterns underlying the observed galaxy distribution.« less

  8. Collaborative Working for Large Digitisation Projects

    ERIC Educational Resources Information Center

    Yeates, Robin; Guy, Damon

    2006-01-01

    Purpose: To explore the effectiveness of large-scale consortia for disseminating local heritage via the web. To describe the creation of a large geographically based cultural heritage consortium in the South East of England and management lessons resulting from a major web site digitisation project. To encourage the improved sharing of experience…

  9. Large orb-webs adapted to maximise total biomass not rare, large prey

    PubMed Central

    Harmer, Aaron M. T.; Clausen, Philip D.; Wroe, Stephen; Madin, Joshua S.

    2015-01-01

    Spider orb-webs are the ultimate anti-ballistic devices, capable of dissipating the relatively massive kinetic energy of flying prey. Increased web size and prey stopping capacity have co-evolved in a number orb-web taxa, but the selective forces driving web size and performance increases are under debate. The rare, large prey hypothesis maintains that the energetic benefits of rare, very large prey are so much greater than the gains from smaller, more common prey that smaller prey are irrelevant for reproduction. Here, we integrate biophysical and ecological data and models to test a major prediction of the rare, large prey hypothesis, that selection should favour webs with increased stopping capacity and that large prey should comprise a significant proportion of prey stopped by a web. We find that larger webs indeed have a greater capacity to stop large prey. However, based on prey ecology, we also find that these large prey make up a tiny fraction of the total biomass (=energy) potentially captured. We conclude that large webs are adapted to stop more total biomass, and that the capacity to stop rare, but very large, prey is an incidental consequence of the longer radial silks that scale with web size. PMID:26374379

  10. Development of a metal-clad advanced composite shear web design concept

    NASA Technical Reports Server (NTRS)

    Laakso, J. H.

    1974-01-01

    An advanced composite web concept was developed for potential application to the Space Shuttle Orbiter main engine thrust structure. The program consisted of design synthesis, analysis, detail design, element testing, and large scale component testing. A concept was sought that offered significant weight saving by the use of Boron/Epoxy (B/E) reinforced titanium plate structure. The desired concept was one that was practical and that utilized metal to efficiently improve structural reliability. The resulting development of a unique titanium-clad B/E shear web design concept is described. Three large scale components were fabricated and tested to demonstrate the performance of the concept: a titanium-clad plus or minus 45 deg B/E web laminate stiffened with vertical B/E reinforced aluminum stiffeners.

  11. Large scale rigidity-based flexibility analysis of biomolecules

    PubMed Central

    Streinu, Ileana

    2016-01-01

    KINematics And RIgidity (KINARI) is an on-going project for in silico flexibility analysis of proteins. The new version of the software, Kinari-2, extends the functionality of our free web server KinariWeb, incorporates advanced web technologies, emphasizes the reproducibility of its experiments, and makes substantially improved tools available to the user. It is designed specifically for large scale experiments, in particular, for (a) very large molecules, including bioassemblies with high degree of symmetry such as viruses and crystals, (b) large collections of related biomolecules, such as those obtained through simulated dilutions, mutations, or conformational changes from various types of dynamics simulations, and (c) is intended to work as seemlessly as possible on the large, idiosyncratic, publicly available repository of biomolecules, the Protein Data Bank. We describe the system design, along with the main data processing, computational, mathematical, and validation challenges underlying this phase of the KINARI project. PMID:26958583

  12. Web-based Visual Analytics for Extreme Scale Climate Science

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Steed, Chad A; Evans, Katherine J; Harney, John F

    In this paper, we introduce a Web-based visual analytics framework for democratizing advanced visualization and analysis capabilities pertinent to large-scale earth system simulations. We address significant limitations of present climate data analysis tools such as tightly coupled dependencies, ineffi- cient data movements, complex user interfaces, and static visualizations. Our Web-based visual analytics framework removes critical barriers to the widespread accessibility and adoption of advanced scientific techniques. Using distributed connections to back-end diagnostics, we minimize data movements and leverage HPC platforms. We also mitigate system dependency issues by employing a RESTful interface. Our framework embraces the visual analytics paradigm via newmore » visual navigation techniques for hierarchical parameter spaces, multi-scale representations, and interactive spatio-temporal data mining methods that retain details. Although generalizable to other science domains, the current work focuses on improving exploratory analysis of large-scale Community Land Model (CLM) and Community Atmosphere Model (CAM) simulations.« less

  13. geoknife: Reproducible web-processing of large gridded datasets

    USGS Publications Warehouse

    Read, Jordan S.; Walker, Jordan I.; Appling, Alison P.; Blodgett, David L.; Read, Emily K.; Winslow, Luke A.

    2016-01-01

    Geoprocessing of large gridded data according to overlap with irregular landscape features is common to many large-scale ecological analyses. The geoknife R package was created to facilitate reproducible analyses of gridded datasets found on the U.S. Geological Survey Geo Data Portal web application or elsewhere, using a web-enabled workflow that eliminates the need to download and store large datasets that are reliably hosted on the Internet. The package provides access to several data subset and summarization algorithms that are available on remote web processing servers. Outputs from geoknife include spatial and temporal data subsets, spatially-averaged time series values filtered by user-specified areas of interest, and categorical coverage fractions for various land-use types.

  14. Building to Scale: An Analysis of Web-Based Services in CIC (Big Ten) Libraries.

    ERIC Educational Resources Information Center

    Dewey, Barbara I.

    Advancing library services in large universities requires creative approaches for "building to scale." This is the case for CIC, Committee on Institutional Cooperation (Big Ten), libraries whose home institutions serve thousands of students, faculty, staff, and others. Developing virtual Web-based services is an increasingly viable…

  15. Data partitioning enables the use of standard SOAP Web Services in genome-scale workflows.

    PubMed

    Sztromwasser, Pawel; Puntervoll, Pål; Petersen, Kjell

    2011-07-26

    Biological databases and computational biology tools are provided by research groups around the world, and made accessible on the Web. Combining these resources is a common practice in bioinformatics, but integration of heterogeneous and often distributed tools and datasets can be challenging. To date, this challenge has been commonly addressed in a pragmatic way, by tedious and error-prone scripting. Recently however a more reliable technique has been identified and proposed as the platform that would tie together bioinformatics resources, namely Web Services. In the last decade the Web Services have spread wide in bioinformatics, and earned the title of recommended technology. However, in the era of high-throughput experimentation, a major concern regarding Web Services is their ability to handle large-scale data traffic. We propose a stream-like communication pattern for standard SOAP Web Services, that enables efficient flow of large data traffic between a workflow orchestrator and Web Services. We evaluated the data-partitioning strategy by comparing it with typical communication patterns on an example pipeline for genomic sequence annotation. The results show that data-partitioning lowers resource demands of services and increases their throughput, which in consequence allows to execute in-silico experiments on genome-scale, using standard SOAP Web Services and workflows. As a proof-of-principle we annotated an RNA-seq dataset using a plain BPEL workflow engine.

  16. QuickEval: a web application for psychometric scaling experiments

    NASA Astrophysics Data System (ADS)

    Van Ngo, Khai; Storvik, Jehans J.; Dokkeberg, Christopher A.; Farup, Ivar; Pedersen, Marius

    2015-01-01

    QuickEval is a web application for carrying out psychometric scaling experiments. It offers the possibility of running controlled experiments in a laboratory, or large scale experiment over the web for people all over the world. It is a unique one of a kind web application, and it is a software needed in the image quality field. It is also, to the best of knowledge, the first software that supports the three most common scaling methods; paired comparison, rank order, and category judgement. It is also the first software to support rank order. Hopefully, a side effect of this newly created software is that it will lower the threshold to perform psychometric experiments, improve the quality of the experiments being carried out, make it easier to reproduce experiments, and increase research on image quality both in academia and industry. The web application is available at www.colourlab.no/quickeval.

  17. A Web-based Distributed Voluntary Computing Platform for Large Scale Hydrological Computations

    NASA Astrophysics Data System (ADS)

    Demir, I.; Agliamzanov, R.

    2014-12-01

    Distributed volunteer computing can enable researchers and scientist to form large parallel computing environments to utilize the computing power of the millions of computers on the Internet, and use them towards running large scale environmental simulations and models to serve the common good of local communities and the world. Recent developments in web technologies and standards allow client-side scripting languages to run at speeds close to native application, and utilize the power of Graphics Processing Units (GPU). Using a client-side scripting language like JavaScript, we have developed an open distributed computing framework that makes it easy for researchers to write their own hydrologic models, and run them on volunteer computers. Users will easily enable their websites for visitors to volunteer sharing their computer resources to contribute running advanced hydrological models and simulations. Using a web-based system allows users to start volunteering their computational resources within seconds without installing any software. The framework distributes the model simulation to thousands of nodes in small spatial and computational sizes. A relational database system is utilized for managing data connections and queue management for the distributed computing nodes. In this paper, we present a web-based distributed volunteer computing platform to enable large scale hydrological simulations and model runs in an open and integrated environment.

  18. Orientation of cosmic web filaments with respect to the underlying velocity field

    NASA Astrophysics Data System (ADS)

    Tempel, E.; Libeskind, N. I.; Hoffman, Y.; Liivamägi, L. J.; Tamm, A.

    2014-01-01

    The large-scale structure of the Universe is characterized by a web-like structure made of voids, sheets, filaments and knots. The structure of this so-called cosmic web is dictated by the local velocity shear tensor. In particular, the local direction of a filament should be strongly aligned with hat{e}_3, the eigenvector associated with the smallest eigenvalue of the tensor. That conjecture is tested here on the basis of a cosmological simulation. The cosmic web delineated by the halo distribution is probed by a marked point process with interactions (the Bisous model), detecting filaments directly from the halo distribution (P-web). The detected P-web filaments are found to be strongly aligned with the local hat{e}_3: the alignment is within 30° for ˜80 per cent of the elements. This indicates that large-scale filaments defined purely from the distribution of haloes carry more than just morphological information, although the Bisous model does not make any prior assumption on the underlying shear tensor. The P-web filaments are also compared to the structure revealed from the velocity shear tensor itself (V-web). In the densest regions, the P- and V-web filaments overlap well (90 per cent), whereas in lower density regions, the P-web filaments preferentially mark sheets in the V-web.

  19. Assessing the Effect of Web-Based Learning Tools on Student Understanding of Stoichiometry Using Knowledge Space Theory

    ERIC Educational Resources Information Center

    Arasasingham, Ramesh D.; Taagepera, Mare; Potter, Frank; Martorell, Ingrid; Lonjers, Stacy

    2005-01-01

    Student achievement in web-based learning tools is assessed by using in-class examination, pretests, and posttests. The study reveals that using mastering chemistry web software in large-scale instruction provides an overall benefit to introductory chemistry students.

  20. Design and Development of a Framework Based on Ogc Web Services for the Visualization of Three Dimensional Large-Scale Geospatial Data Over the Web

    NASA Astrophysics Data System (ADS)

    Roccatello, E.; Nozzi, A.; Rumor, M.

    2013-05-01

    This paper illustrates the key concepts behind the design and the development of a framework, based on OGC services, capable to visualize 3D large scale geospatial data streamed over the web. WebGISes are traditionally bounded to a bi-dimensional simplified representation of the reality and though they are successfully addressing the lack of flexibility and simplicity of traditional desktop clients, a lot of effort is still needed to reach desktop GIS features, like 3D visualization. The motivations behind this work lay in the widespread availability of OGC Web Services inside government organizations and in the technology support to HTML 5 and WebGL standard of the web browsers. This delivers an improved user experience, similar to desktop applications, therefore allowing to augment traditional WebGIS features with a 3D visualization framework. This work could be seen as an extension of the Cityvu project, started in 2008 with the aim of a plug-in free OGC CityGML viewer. The resulting framework has also been integrated in existing 3DGIS software products and will be made available in the next months.

  1. Security and Efficiency Concerns With Distributed Collaborative Networking Environments

    DTIC Science & Technology

    2003-09-01

    have the ability to access Web communications services of the WebEx MediaTone Network from a single login. [24] WebEx provides a range of secure...Web. WebEx services enable secure data, voice and video communications through the browser and are supported by the WebEx MediaTone Network, a global...designed to host large-scale, structured events and conferences, featuring a Q&A Manager that allows multiple moderators to handle questions while

  2. SLIDE - a web-based tool for interactive visualization of large-scale -omics data.

    PubMed

    Ghosh, Soumita; Datta, Abhik; Tan, Kaisen; Choi, Hyungwon

    2018-06-28

    Data visualization is often regarded as a post hoc step for verifying statistically significant results in the analysis of high-throughput data sets. This common practice leaves a large amount of raw data behind, from which more information can be extracted. However, existing solutions do not provide capabilities to explore large-scale raw datasets using biologically sensible queries, nor do they allow user interaction based real-time customization of graphics. To address these drawbacks, we have designed an open-source, web-based tool called Systems-Level Interactive Data Exploration, or SLIDE to visualize large-scale -omics data interactively. SLIDE's interface makes it easier for scientists to explore quantitative expression data in multiple resolutions in a single screen. SLIDE is publicly available under BSD license both as an online version as well as a stand-alone version at https://github.com/soumitag/SLIDE. Supplementary Information are available at Bioinformatics online.

  3. Secure web-based invocation of large-scale plasma simulation codes

    NASA Astrophysics Data System (ADS)

    Dimitrov, D. A.; Busby, R.; Exby, J.; Bruhwiler, D. L.; Cary, J. R.

    2004-12-01

    We present our design and initial implementation of a web-based system for running, both in parallel and serial, Particle-In-Cell (PIC) codes for plasma simulations with automatic post processing and generation of visual diagnostics.

  4. Collaborative Visualization for Large-Scale Accelerator Electromagnetic Modeling (Final Report)

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    William J. Schroeder

    2011-11-13

    This report contains the comprehensive summary of the work performed on the SBIR Phase II, Collaborative Visualization for Large-Scale Accelerator Electromagnetic Modeling at Kitware Inc. in collaboration with Stanford Linear Accelerator Center (SLAC). The goal of the work was to develop collaborative visualization tools for large-scale data as illustrated in the figure below. The solutions we proposed address the typical problems faced by geographicallyand organizationally-separated research and engineering teams, who produce large data (either through simulation or experimental measurement) and wish to work together to analyze and understand their data. Because the data is large, we expect that it cannotmore » be easily transported to each team member's work site, and that the visualization server must reside near the data. Further, we also expect that each work site has heterogeneous resources: some with large computing clients, tiled (or large) displays and high bandwidth; others sites as simple as a team member on a laptop computer. Our solution is based on the open-source, widely used ParaView large-data visualization application. We extended this tool to support multiple collaborative clients who may locally visualize data, and then periodically rejoin and synchronize with the group to discuss their findings. Options for managing session control, adding annotation, and defining the visualization pipeline, among others, were incorporated. We also developed and deployed a Web visualization framework based on ParaView that enables the Web browser to act as a participating client in a collaborative session. The ParaView Web Visualization framework leverages various Web technologies including WebGL, JavaScript, Java and Flash to enable interactive 3D visualization over the web using ParaView as the visualization server. We steered the development of this technology by teaming with the SLAC National Accelerator Laboratory. SLAC has a computationally-intensive problem important to the nations scientific progress as described shortly. Further, SLAC researchers routinely generate massive amounts of data, and frequently collaborate with other researchers located around the world. Thus SLAC is an ideal teammate through which to develop, test and deploy this technology. The nature of the datasets generated by simulations performed at SLAC presented unique visualization challenges especially when dealing with higher-order elements that were addressed during this Phase II. During this Phase II, we have developed a strong platform for collaborative visualization based on ParaView. We have developed and deployed a ParaView Web Visualization framework that can be used for effective collaboration over the Web. Collaborating and visualizing over the Web presents the community with unique opportunities for sharing and accessing visualization and HPC resources that hitherto with either inaccessible or difficult to use. The technology we developed in here will alleviate both these issues as it becomes widely deployed and adopted.« less

  5. Life as an emergent phenomenon: studies from a large-scale boid simulation and web data.

    PubMed

    Ikegami, Takashi; Mototake, Yoh-Ichi; Kobori, Shintaro; Oka, Mizuki; Hashimoto, Yasuhiro

    2017-12-28

    A large group with a special structure can become the mother of emergence. We discuss this hypothesis in relation to large-scale boid simulations and web data. In the boid swarm simulations, the nucleation, organization and collapse dynamics were found to be more diverse in larger flocks than in smaller flocks. In the second analysis, large web data, consisting of shared photos with descriptive tags, tended to group together users with similar tendencies, allowing the network to develop a core-periphery structure. We show that the generation rate of novel tags and their usage frequencies are high in the higher-order cliques. In this case, novelty is not considered to arise randomly; rather, it is generated as a result of a large and structured network. We contextualize these results in terms of adjacent possible theory and as a new way to understand collective intelligence. We argue that excessive information and material flow can become a source of innovation.This article is part of the themed issue 'Reconceptualizing the origins of life'. © 2017 The Author(s).

  6. Life as an emergent phenomenon: studies from a large-scale boid simulation and web data

    NASA Astrophysics Data System (ADS)

    Ikegami, Takashi; Mototake, Yoh-ichi; Kobori, Shintaro; Oka, Mizuki; Hashimoto, Yasuhiro

    2017-11-01

    A large group with a special structure can become the mother of emergence. We discuss this hypothesis in relation to large-scale boid simulations and web data. In the boid swarm simulations, the nucleation, organization and collapse dynamics were found to be more diverse in larger flocks than in smaller flocks. In the second analysis, large web data, consisting of shared photos with descriptive tags, tended to group together users with similar tendencies, allowing the network to develop a core-periphery structure. We show that the generation rate of novel tags and their usage frequencies are high in the higher-order cliques. In this case, novelty is not considered to arise randomly; rather, it is generated as a result of a large and structured network. We contextualize these results in terms of adjacent possible theory and as a new way to understand collective intelligence. We argue that excessive information and material flow can become a source of innovation. This article is part of the themed issue 'Reconceptualizing the origins of life'.

  7. WebCIS: large scale deployment of a Web-based clinical information system.

    PubMed

    Hripcsak, G; Cimino, J J; Sengupta, S

    1999-01-01

    WebCIS is a Web-based clinical information system. It sits atop the existing Columbia University clinical information system architecture, which includes a clinical repository, the Medical Entities Dictionary, an HL7 interface engine, and an Arden Syntax based clinical event monitor. WebCIS security features include authentication with secure tokens, authorization maintained in an LDAP server, SSL encryption, permanent audit logs, and application time outs. WebCIS is currently used by 810 physicians at the Columbia-Presbyterian center of New York Presbyterian Healthcare to review and enter data into the electronic medical record. Current deployment challenges include maintaining adequate database performance despite complex queries, replacing large numbers of computers that cannot run modern Web browsers, and training users that have never logged onto the Web. Although the raised expectations and higher goals have increased deployment costs, the end result is a far more functional, far more available system.

  8. An evaluation of multi-probe locality sensitive hashing for computing similarities over web-scale query logs.

    PubMed

    Cormode, Graham; Dasgupta, Anirban; Goyal, Amit; Lee, Chi Hoon

    2018-01-01

    Many modern applications of AI such as web search, mobile browsing, image processing, and natural language processing rely on finding similar items from a large database of complex objects. Due to the very large scale of data involved (e.g., users' queries from commercial search engines), computing such near or nearest neighbors is a non-trivial task, as the computational cost grows significantly with the number of items. To address this challenge, we adopt Locality Sensitive Hashing (a.k.a, LSH) methods and evaluate four variants in a distributed computing environment (specifically, Hadoop). We identify several optimizations which improve performance, suitable for deployment in very large scale settings. The experimental results demonstrate our variants of LSH achieve the robust performance with better recall compared with "vanilla" LSH, even when using the same amount of space.

  9. Assessing the trophic position and ecological role of squids in marine ecosystems by means of food-web models

    NASA Astrophysics Data System (ADS)

    Coll, Marta; Navarro, Joan; Olson, Robert J.; Christensen, Villy

    2013-10-01

    We synthesized available information from ecological models at local and regional scales to obtain a global picture of the trophic position and ecological role of squids in marine ecosystems. First, static food-web models were used to analyze basic ecological parameters and indicators of squids: biomass, production, consumption, trophic level, omnivory index, predation mortality diet, and the ecological role. In addition, we developed various dynamic temporal simulations using two food-web models that included squids in their parameterization, and we investigated potential impacts of fishing pressure and environmental conditions for squid populations and, consequently, for marine food webs. Our results showed that squids occupy a large range of trophic levels in marine food webs and show a large trophic width, reflecting the versatility in their feeding behaviors and dietary habits. Models illustrated that squids are abundant organisms in marine ecosystems, and have high growth and consumption rates, but these parameters are highly variable because squids are adapted to a large variety of environmental conditions. Results also show that squids can have a large trophic impact on other elements of the food web, and top-down control from squids to their prey can be high. In addition, some squid species are important prey of apical predators and may be keystone species in marine food webs. In fact, we found strong interrelationships between neritic squids and the populations of their prey and predators in coastal and shelf areas, while the role of squids in open ocean and upwelling ecosystems appeared more constrained to a bottom-up impact on their predators. Therefore, large removals of squids will likely have large-scale effects on marine ecosystems. In addition, simulations confirm that squids are able to benefit from a general increase in fishing pressure, mainly due to predation release, and quickly respond to changes triggered by the environment. Squids may thus be very sensitive to the effects of fishing and climate change.

  10. Large-Scale Overlays and Trends: Visually Mining, Panning and Zooming the Observable Universe.

    PubMed

    Luciani, Timothy Basil; Cherinka, Brian; Oliphant, Daniel; Myers, Sean; Wood-Vasey, W Michael; Labrinidis, Alexandros; Marai, G Elisabeta

    2014-07-01

    We introduce a web-based computing infrastructure to assist the visual integration, mining and interactive navigation of large-scale astronomy observations. Following an analysis of the application domain, we design a client-server architecture to fetch distributed image data and to partition local data into a spatial index structure that allows prefix-matching of spatial objects. In conjunction with hardware-accelerated pixel-based overlays and an online cross-registration pipeline, this approach allows the fetching, displaying, panning and zooming of gigabit panoramas of the sky in real time. To further facilitate the integration and mining of spatial and non-spatial data, we introduce interactive trend images-compact visual representations for identifying outlier objects and for studying trends within large collections of spatial objects of a given class. In a demonstration, images from three sky surveys (SDSS, FIRST and simulated LSST results) are cross-registered and integrated as overlays, allowing cross-spectrum analysis of astronomy observations. Trend images are interactively generated from catalog data and used to visually mine astronomy observations of similar type. The front-end of the infrastructure uses the web technologies WebGL and HTML5 to enable cross-platform, web-based functionality. Our approach attains interactive rendering framerates; its power and flexibility enables it to serve the needs of the astronomy community. Evaluation on three case studies, as well as feedback from domain experts emphasize the benefits of this visual approach to the observational astronomy field; and its potential benefits to large scale geospatial visualization in general.

  11. SOCRAT Platform Design: A Web Architecture for Interactive Visual Analytics Applications

    PubMed Central

    Kalinin, Alexandr A.; Palanimalai, Selvam; Dinov, Ivo D.

    2018-01-01

    The modern web is a successful platform for large scale interactive web applications, including visualizations. However, there are no established design principles for building complex visual analytics (VA) web applications that could efficiently integrate visualizations with data management, computational transformation, hypothesis testing, and knowledge discovery. This imposes a time-consuming design and development process on many researchers and developers. To address these challenges, we consider the design requirements for the development of a module-based VA system architecture, adopting existing practices of large scale web application development. We present the preliminary design and implementation of an open-source platform for Statistics Online Computational Resource Analytical Toolbox (SOCRAT). This platform defines: (1) a specification for an architecture for building VA applications with multi-level modularity, and (2) methods for optimizing module interaction, re-usage, and extension. To demonstrate how this platform can be used to integrate a number of data management, interactive visualization, and analysis tools, we implement an example application for simple VA tasks including raw data input and representation, interactive visualization and analysis. PMID:29630069

  12. SOCRAT Platform Design: A Web Architecture for Interactive Visual Analytics Applications.

    PubMed

    Kalinin, Alexandr A; Palanimalai, Selvam; Dinov, Ivo D

    2017-04-01

    The modern web is a successful platform for large scale interactive web applications, including visualizations. However, there are no established design principles for building complex visual analytics (VA) web applications that could efficiently integrate visualizations with data management, computational transformation, hypothesis testing, and knowledge discovery. This imposes a time-consuming design and development process on many researchers and developers. To address these challenges, we consider the design requirements for the development of a module-based VA system architecture, adopting existing practices of large scale web application development. We present the preliminary design and implementation of an open-source platform for Statistics Online Computational Resource Analytical Toolbox (SOCRAT). This platform defines: (1) a specification for an architecture for building VA applications with multi-level modularity, and (2) methods for optimizing module interaction, re-usage, and extension. To demonstrate how this platform can be used to integrate a number of data management, interactive visualization, and analysis tools, we implement an example application for simple VA tasks including raw data input and representation, interactive visualization and analysis.

  13. WebViz:A Web-based Collaborative Interactive Visualization System for large-Scale Data Sets

    NASA Astrophysics Data System (ADS)

    Yuen, D. A.; McArthur, E.; Weiss, R. M.; Zhou, J.; Yao, B.

    2010-12-01

    WebViz is a web-based application designed to conduct collaborative, interactive visualizations of large data sets for multiple users, allowing researchers situated all over the world to utilize the visualization services offered by the University of Minnesota’s Laboratory for Computational Sciences and Engineering (LCSE). This ongoing project has been built upon over the last 3 1/2 years .The motivation behind WebViz lies primarily with the need to parse through an increasing amount of data produced by the scientific community as a result of larger and faster multicore and massively parallel computers coming to the market, including the use of general purpose GPU computing. WebViz allows these large data sets to be visualized online by anyone with an account. The application allows users to save time and resources by visualizing data ‘on the fly’, wherever he or she may be located. By leveraging AJAX via the Google Web Toolkit (http://code.google.com/webtoolkit/), we are able to provide users with a remote, web portal to LCSE's (http://www.lcse.umn.edu) large-scale interactive visualization system already in place at the University of Minnesota. LCSE’s custom hierarchical volume rendering software provides high resolution visualizations on the order of 15 million pixels and has been employed for visualizing data primarily from simulations in astrophysics to geophysical fluid dynamics . In the current version of WebViz, we have implemented a highly extensible back-end framework built around HTTP "server push" technology. The web application is accessible via a variety of devices including netbooks, iPhones, and other web and javascript-enabled cell phones. Features in the current version include the ability for users to (1) securely login (2) launch multiple visualizations (3) conduct collaborative visualization sessions (4) delegate control aspects of a visualization to others and (5) engage in collaborative chats with other users within the user interface of the web application. These features are all in addition to a full range of essential visualization functions including 3-D camera and object orientation, position manipulation, time-stepping control, and custom color/alpha mapping.

  14. Spider webs designed for rare but life-saving catches

    PubMed Central

    Venner, Samuel; Casas, Jérôme

    2005-01-01

    The impact of rare but positive events on the design of organisms has been largely ignored, probably due to the paucity of recordings of such events and to the difficulty of estimating their impact on lifetime reproductive success. In this respect, we investigated the size of spider webs in relation to rare but large prey catches. First, we collected field data on a short time-scale using the common orb-weaving spider Zygiella x-notata to determine the distribution of the size of prey caught and to quantify the relationship between web size and daily capture success. Second, we explored, with an energetic model, the consequences of an increase in web size on spider fitness. Our results showed that (i) the great majority of prey caught are quite small (body length less than 2 mm) while large prey (length greater than 10 mm) are rare, (ii) spiders cannot survive or produce eggs without catching these large but rare prey and (iii) increasing web size increases the daily number of prey caught and thus long-term survival and fecundity. Spider webs seem, therefore, designed for making the best of the rare but crucial event of catching large prey. PMID:16048774

  15. Large-Scale Astrophysical Visualization on Smartphones

    NASA Astrophysics Data System (ADS)

    Becciani, U.; Massimino, P.; Costa, A.; Gheller, C.; Grillo, A.; Krokos, M.; Petta, C.

    2011-07-01

    Nowadays digital sky surveys and long-duration, high-resolution numerical simulations using high performance computing and grid systems produce multidimensional astrophysical datasets in the order of several Petabytes. Sharing visualizations of such datasets within communities and collaborating research groups is of paramount importance for disseminating results and advancing astrophysical research. Moreover educational and public outreach programs can benefit greatly from novel ways of presenting these datasets by promoting understanding of complex astrophysical processes, e.g., formation of stars and galaxies. We have previously developed VisIVO Server, a grid-enabled platform for high-performance large-scale astrophysical visualization. This article reviews the latest developments on VisIVO Web, a custom designed web portal wrapped around VisIVO Server, then introduces VisIVO Smartphone, a gateway connecting VisIVO Web and data repositories for mobile astrophysical visualization. We discuss current work and summarize future developments.

  16. Rotation invariant fast features for large-scale recognition

    NASA Astrophysics Data System (ADS)

    Takacs, Gabriel; Chandrasekhar, Vijay; Tsai, Sam; Chen, David; Grzeszczuk, Radek; Girod, Bernd

    2012-10-01

    We present an end-to-end feature description pipeline which uses a novel interest point detector and Rotation- Invariant Fast Feature (RIFF) descriptors. The proposed RIFF algorithm is 15× faster than SURF1 while producing large-scale retrieval results that are comparable to SIFT.2 Such high-speed features benefit a range of applications from Mobile Augmented Reality (MAR) to web-scale image retrieval and analysis.

  17. CImbinator: a web-based tool for drug synergy analysis in small- and large-scale datasets.

    PubMed

    Flobak, Åsmund; Vazquez, Miguel; Lægreid, Astrid; Valencia, Alfonso

    2017-08-01

    Drug synergies are sought to identify combinations of drugs particularly beneficial. User-friendly software solutions that can assist analysis of large-scale datasets are required. CImbinator is a web-service that can aid in batch-wise and in-depth analyzes of data from small-scale and large-scale drug combination screens. CImbinator offers to quantify drug combination effects, using both the commonly employed median effect equation, as well as advanced experimental mathematical models describing dose response relationships. CImbinator is written in Ruby and R. It uses the R package drc for advanced drug response modeling. CImbinator is available at http://cimbinator.bioinfo.cnio.es , the source-code is open and available at https://github.com/Rbbt-Workflows/combination_index . A Docker image is also available at https://hub.docker.com/r/mikisvaz/rbbt-ci_mbinator/ . asmund.flobak@ntnu.no or miguel.vazquez@cnio.es. Supplementary data are available at Bioinformatics online. © The Author(s) 2017. Published by Oxford University Press.

  18. An evaluation of multi-probe locality sensitive hashing for computing similarities over web-scale query logs

    PubMed Central

    2018-01-01

    Many modern applications of AI such as web search, mobile browsing, image processing, and natural language processing rely on finding similar items from a large database of complex objects. Due to the very large scale of data involved (e.g., users’ queries from commercial search engines), computing such near or nearest neighbors is a non-trivial task, as the computational cost grows significantly with the number of items. To address this challenge, we adopt Locality Sensitive Hashing (a.k.a, LSH) methods and evaluate four variants in a distributed computing environment (specifically, Hadoop). We identify several optimizations which improve performance, suitable for deployment in very large scale settings. The experimental results demonstrate our variants of LSH achieve the robust performance with better recall compared with “vanilla” LSH, even when using the same amount of space. PMID:29346410

  19. Computing the universe: how large-scale simulations illuminate galaxies and dark energy

    NASA Astrophysics Data System (ADS)

    O'Shea, Brian

    2015-04-01

    High-performance and large-scale computing is absolutely to understanding astronomical objects such as stars, galaxies, and the cosmic web. This is because these are structures that operate on physical, temporal, and energy scales that cannot be reasonably approximated in the laboratory, and whose complexity and nonlinearity often defies analytic modeling. In this talk, I show how the growth of computing platforms over time has facilitated our understanding of astrophysical and cosmological phenomena, focusing primarily on galaxies and large-scale structure in the Universe.

  20. The value of the Semantic Web in the laboratory.

    PubMed

    Frey, Jeremy G

    2009-06-01

    The Semantic Web is beginning to impact on the wider chemical and physical sciences, beyond the earlier adopted bio-informatics. While useful in large-scale data driven science with automated processing, these technologies can also help integrate the work of smaller scale laboratories producing diverse data. The semantics aid the discovery, reliable re-use of data, provide improved provenance and facilitate automated processing by increased resilience to changes in presentation and reduced ambiguity. The Semantic Web, its tools and collections are not yet competitive with well-established solutions to current problems. It is in the reduced cost of instituting solutions to new problems that the versatility of Semantic Web-enabled data and resources will make their mark once the more general-purpose tools are more available.

  1. Evaluation of a metal shear web selectively reinforced with filamentary composites for space shuttle application. Phase 2: summary report: Shear web component fabrication

    NASA Technical Reports Server (NTRS)

    Laakso, J. H.; Smith, D. D.; Zimmerman, D. K.

    1973-01-01

    The fabrication of two shear web test elements and three large scale shear web test components are reported. In addition, the fabrication of test fixtures for the elements and components is described. The center-loaded beam test fixtures were configured to have a test side and a dummy or permanent side. The test fixtures were fabricated from standard extruded aluminum sections and plates and were designed to be reuseable.

  2. Lyα-emitting galaxies as a probe of reionization: large-scale bubble morphology and small-scale absorbers

    NASA Astrophysics Data System (ADS)

    Kakiichi, Koki; Dijkstra, Mark; Ciardi, Benedetta; Graziani, Luca

    2016-12-01

    The visibility of Lyα-emitting galaxies during the Epoch of Reionization is controlled by both diffuse H I patches in large-scale bubble morphology and small-scale absorbers. To investigate their impacts on Lyα transfer, we apply a novel combination of analytic modelling and cosmological hydrodynamical, radiative transfer simulations to three reionization models: (I) the `bubble' model, where only diffuse H I outside ionized bubbles is present; (II) the `web' model, where H I exists only in overdense self-shielded gas; and (III) the hybrid `web-bubble' model. The three models can explain the observed Lyα luminosity function equally well, but with very different H I fractions. This confirms a degeneracy between the ionization topology of the intergalactic medium (IGM) and the H I fraction inferred from Lyα surveys. We highlight the importance of the clustering of small-scale absorbers around galaxies. A combined analysis of the Lyα luminosity function and the Lyα fraction can break this degeneracy and provide constraints on the reionization history and its topology. Constraints can be improved by analysing the full MUV-dependent redshift evolution of the Lyα fraction of Lyman break galaxies. We find that the IGM-transmission probability distribution function is unimodal for bubble models and bimodal in web models. Comparing our models to observations, we infer that the neutral fraction at z ˜ 7 is likely to be of the order of tens of per cent when interpreted with bubble or web-bubble models, with a conservative lower limit ˜1 per cent when interpreted with web models.

  3. Semantic Search of Web Services

    ERIC Educational Resources Information Center

    Hao, Ke

    2013-01-01

    This dissertation addresses semantic search of Web services using natural language processing. We first survey various existing approaches, focusing on the fact that the expensive costs of current semantic annotation frameworks result in limited use of semantic search for large scale applications. We then propose a vector space model based service…

  4. The cosmic spiderweb: equivalence of cosmic, architectural and origami tessellations.

    PubMed

    Neyrinck, Mark C; Hidding, Johan; Konstantatou, Marina; van de Weygaert, Rien

    2018-04-01

    For over 20 years, the term 'cosmic web' has guided our understanding of the large-scale arrangement of matter in the cosmos, accurately evoking the concept of a network of galaxies linked by filaments. But the physical correspondence between the cosmic web and structural engineering or textile 'spiderwebs' is even deeper than previously known, and also extends to origami tessellations. Here, we explain that in a good structure-formation approximation known as the adhesion model, threads of the cosmic web form a spiderweb, i.e. can be strung up to be entirely in tension. The correspondence is exact if nodes sampling voids are included, and if structure is excluded within collapsed regions (walls, filaments and haloes), where dark-matter multistreaming and baryonic physics affect the structure. We also suggest how concepts arising from this link might be used to test cosmological models: for example, to test for large-scale anisotropy and rotational flows in the cosmos.

  5. The cosmic spiderweb: equivalence of cosmic, architectural and origami tessellations

    NASA Astrophysics Data System (ADS)

    Neyrinck, Mark C.; Hidding, Johan; Konstantatou, Marina; van de Weygaert, Rien

    2018-04-01

    For over 20 years, the term `cosmic web' has guided our understanding of the large-scale arrangement of matter in the cosmos, accurately evoking the concept of a network of galaxies linked by filaments. But the physical correspondence between the cosmic web and structural engineering or textile `spiderwebs' is even deeper than previously known, and also extends to origami tessellations. Here, we explain that in a good structure-formation approximation known as the adhesion model, threads of the cosmic web form a spiderweb, i.e. can be strung up to be entirely in tension. The correspondence is exact if nodes sampling voids are included, and if structure is excluded within collapsed regions (walls, filaments and haloes), where dark-matter multistreaming and baryonic physics affect the structure. We also suggest how concepts arising from this link might be used to test cosmological models: for example, to test for large-scale anisotropy and rotational flows in the cosmos.

  6. Benchmark of Client and Server-Side Catchment Delineation Approaches on Web-Based Systems

    NASA Astrophysics Data System (ADS)

    Demir, I.; Sermet, M. Y.; Sit, M. A.

    2016-12-01

    Recent advances in internet and cyberinfrastructure technologies have provided the capability to acquire large scale spatial data from various gauges and sensor networks. The collection of environmental data increased demand for applications which are capable of managing and processing large-scale and high-resolution data sets. With the amount and resolution of data sets provided, one of the challenging tasks for organizing and customizing hydrological data sets is delineation of watersheds on demand. Watershed delineation is a process for creating a boundary that represents the contributing area for a specific control point or water outlet, with intent of characterization and analysis of portions of a study area. Although many GIS tools and software for watershed analysis are available on desktop systems, there is a need for web-based and client-side techniques for creating a dynamic and interactive environment for exploring hydrological data. In this project, we demonstrated several watershed delineation techniques on the web with various techniques implemented on the client-side using JavaScript and WebGL, and on the server-side using Python and C++. We also developed a client-side GPGPU (General Purpose Graphical Processing Unit) algorithm to analyze high-resolution terrain data for watershed delineation which allows parallelization using GPU. The web-based real-time analysis of watershed segmentation can be helpful for decision-makers and interested stakeholders while eliminating the need of installing complex software packages and dealing with large-scale data sets. Utilization of the client-side hardware resources also eliminates the need of servers due its crowdsourcing nature. Our goal for future work is to improve other hydrologic analysis methods such as rain flow tracking by adapting presented approaches.

  7. One dark matter mystery: halos in the cosmic web

    NASA Astrophysics Data System (ADS)

    Gaite, Jose

    2015-01-01

    The current cold dark matter cosmological model explains the large scale cosmic web structure but is challenged by the observation of a relatively smooth distribution of matter in galactic clusters. We consider various aspects of modeling the dark matter around galaxies as distributed in smooth halos and, especially, the smoothness of the dark matter halos seen in N-body cosmological simulations. We conclude that the problems of the cold dark matter cosmology on small scales are more serious than normally admitted.

  8. A Data Management System Integrating Web-Based Training and Randomized Trials

    ERIC Educational Resources Information Center

    Muroff, Jordana; Amodeo, Maryann; Larson, Mary Jo; Carey, Margaret; Loftin, Ralph D.

    2011-01-01

    This article describes a data management system (DMS) developed to support a large-scale randomized study of an innovative web-course that was designed to improve substance abuse counselors' knowledge and skills in applying a substance abuse treatment method (i.e., cognitive behavioral therapy; CBT). The randomized trial compared the performance…

  9. 78 FR 7464 - Large Scale Networking (LSN)-Middleware And Grid Interagency Coordination (MAGIC) Team

    Federal Register 2010, 2011, 2012, 2013, 2014

    2013-02-01

    ... Coordination (MAGIC) Team AGENCY: The Networking and Information Technology Research and Development (NITRD... (703) 292-4873. Date/Location: The MAGIC Team meetings are held on the first Wednesday of each month, 2... basis. WebEx participation is available for each meeting. Please reference the MAGIC Team Web site for...

  10. Panoptes: web-based exploration of large scale genome variation data.

    PubMed

    Vauterin, Paul; Jeffery, Ben; Miles, Alistair; Amato, Roberto; Hart, Lee; Wright, Ian; Kwiatkowski, Dominic

    2017-10-15

    The size and complexity of modern large-scale genome variation studies demand novel approaches for exploring and sharing the data. In order to unlock the potential of these data for a broad audience of scientists with various areas of expertise, a unified exploration framework is required that is accessible, coherent and user-friendly. Panoptes is an open-source software framework for collaborative visual exploration of large-scale genome variation data and associated metadata in a web browser. It relies on technology choices that allow it to operate in near real-time on very large datasets. It can be used to browse rich, hybrid content in a coherent way, and offers interactive visual analytics approaches to assist the exploration. We illustrate its application using genome variation data of Anopheles gambiae, Plasmodium falciparum and Plasmodium vivax. Freely available at https://github.com/cggh/panoptes, under the GNU Affero General Public License. paul.vauterin@gmail.com. © The Author 2017. Published by Oxford University Press. All rights reserved. For Permissions, please e-mail: journals.permissions@oup.com

  11. Enhancing the Spectral Hardening of Cosmic TeV Photons by Mixing with Axionlike Particles in the Magnetized Cosmic Web.

    PubMed

    Montanino, Daniele; Vazza, Franco; Mirizzi, Alessandro; Viel, Matteo

    2017-09-08

    Large-scale extragalactic magnetic fields may induce conversions between very-high-energy photons and axionlike particles (ALPs), thereby shielding the photons from absorption on the extragalactic background light. However, in simplified "cell" models, used so far to represent extragalactic magnetic fields, this mechanism would be strongly suppressed by current astrophysical bounds. Here we consider a recent model of extragalactic magnetic fields obtained from large-scale cosmological simulations. Such simulated magnetic fields would have large enhancement in the filaments of matter. As a result, photon-ALP conversions would produce a significant spectral hardening for cosmic TeV photons. This effect would be probed with the upcoming Cherenkov Telescope Array detector. This possible detection would give a unique chance to perform a tomography of the magnetized cosmic web with ALPs.

  12. Cloud-based MOTIFSIM: Detecting Similarity in Large DNA Motif Data Sets.

    PubMed

    Tran, Ngoc Tam L; Huang, Chun-Hsi

    2017-05-01

    We developed the cloud-based MOTIFSIM on Amazon Web Services (AWS) cloud. The tool is an extended version from our web-based tool version 2.0, which was developed based on a novel algorithm for detecting similarity in multiple DNA motif data sets. This cloud-based version further allows researchers to exploit the computing resources available from AWS to detect similarity in multiple large-scale DNA motif data sets resulting from the next-generation sequencing technology. The tool is highly scalable with expandable AWS.

  13. An experimental test of a fundamental food web motif.

    PubMed

    Rip, Jason M K; McCann, Kevin S; Lynn, Denis H; Fawcett, Sonia

    2010-06-07

    Large-scale changes to the world's ecosystem are resulting in the deterioration of biostructure-the complex web of species interactions that make up ecological communities. A difficult, yet crucial task is to identify food web structures, or food web motifs, that are the building blocks of this baroque network of interactions. Once identified, these food web motifs can then be examined through experiments and theory to provide mechanistic explanations for how structure governs ecosystem stability. Here, we synthesize recent ecological research to show that generalist consumers coupling resources with different interaction strengths, is one such motif. This motif amazingly occurs across an enormous range of spatial scales, and so acts to distribute coupled weak and strong interactions throughout food webs. We then perform an experiment that illustrates the importance of this motif to ecological stability. We find that weak interactions coupled to strong interactions by generalist consumers dampen strong interaction strengths and increase community stability. This study takes a critical step by isolating a common food web motif and through clear, experimental manipulation, identifies the fundamental stabilizing consequences of this structure for ecological communities.

  14. Environmental controls on food web regimes: A fluvial perspective

    NASA Astrophysics Data System (ADS)

    Power, Mary E.

    2006-02-01

    Because food web regimes control the biomass of primary producers (e.g., plants or algae), intermediate consumers (e.g., invertebrates), and large top predators (tuna, killer whales), they are of societal as well as academic interest. Some controls over food web regimes may be internal, but many are mediated by conditions or fluxes over large spatial scales. To understand locally observed changes in food webs, we must learn more about how environmental gradients and boundaries affect the fluxes of energy, materials, or organisms through landscapes or seascapes that influence local species interactions. Marine biologists and oceanographers have overcome formidable challenges of fieldwork on the high seas to make remarkable progress towards this goal. In river drainage networks, we have opportunities to address similar questions at smaller spatial scales, in ecosystems with clear physical structure and organization. Despite these advantages, we still have much to learn about linkages between fluxes from watershed landscapes and local food webs in river networks. Longitudinal (downstream) gradients in productivity, disturbance regimes, and habitat structure exert strong effects on the organisms and energy sources of river food webs, but their effects on species interactions are just beginning to be explored. In fluid ecosystems with less obvious physical structure, like the open ocean, discerning features that control the movement of organisms and affect food web dynamics is even more challenging. In both habitats, new sensing, tracing and mapping technologies have revealed how landscape or seascape features (e.g., watershed divides, ocean fronts or circulation cells) channel, contain or concentrate organisms, energy and materials. Field experiments and direct in situ observations of basic natural history, however, remain as vital as ever in interpreting the responses of biota to these features. We need field data that quantify the many spatial and temporal scales of functional relationships that link environments, fluxes and food web interactions to understand how they will respond to intensifying anthropogenic forcing over the coming decades.

  15. Exploring Additional Determinants of Environmentally Responsible Behavior: The Influence of Environmental Literature and Environmental Attitudes

    ERIC Educational Resources Information Center

    Mobley, Catherine; Vagias, Wade M.; DeWard, Sarah L.

    2010-01-01

    It is often assumed that individuals who are knowledgeable and concerned about the environment will engage in environmentally responsible behavior (ERB). We use data from a large scale Web survey hosted on National Geographic's Web site in 2001-2002 to investigate this premise. We examine whether reading three classic environmental books…

  16. Information Tailoring Enhancements for Large Scale Social Data

    DTIC Science & Technology

    2016-03-15

    i.com) 1 Work Performed within This Reporting Period .................................................... 2 1.1 Implemented Temporal Analytics ...following tasks.  Implemented Temporal Analysis Algorithms for Advanced Analytics in Scraawl. We implemented our backend web service design for the...temporal analysis and we created a prototyope GUI web service of Scraawl analytics dashboard.  Upgraded Scraawl computational framework to increase

  17. A Multivariate Analysis of Secondary Students' Experience of Web-Based Language Acquisition

    ERIC Educational Resources Information Center

    Felix, Uschi

    2004-01-01

    This paper reports on a large-scale project designed to replicate an earlier investigation of tertiary students (Felix, 2001) in a secondary school environment. The new project was carried out in five settings, again investigating the potential of the Web as a medium of language instruction. Data was collected by questionnaires and observational…

  18. Extending the Capabilities of Internet-Based Research: Lessons from the Field.

    ERIC Educational Resources Information Center

    Tingling, Peter; Parent, Michael; Wade, Michael

    2003-01-01

    Summarizes the existing practices of Internet research and suggests extensions to them (e.g., consideration of new capabilities, such as adaptive questions and higher levels of flexibility and control) based on a large-scale, national Web survey. Lessons learned include the use of a modular design, management of Web traffic, and the higher level…

  19. 77 FR 58416 - Large Scale Networking (LSN); Middleware and Grid Interagency Coordination (MAGIC) Team

    Federal Register 2010, 2011, 2012, 2013, 2014

    2012-09-20

    ... Coordination (MAGIC) Team AGENCY: The Networking and Information Technology Research and Development (NITRD.... Dates/Location: The MAGIC Team meetings are held on the first Wednesday of each month, 2:00-4:00pm, at... participation is available for each meeting. Please reference the MAGIC Team Web site for updates. Magic Web...

  20. 78 FR 70076 - Large Scale Networking (LSN)-Middleware and Grid Interagency Coordination (MAGIC) Team

    Federal Register 2010, 2011, 2012, 2013, 2014

    2013-11-22

    ... Coordination (MAGIC) Team AGENCY: The Networking and Information Technology Research and Development (NITRD... MAGIC Team meetings are held on the first Wednesday of each month, 2:00-4:00 p.m., at the National... for each meeting. Please reference the MAGIC Team Web site for updates. Magic Web site: The agendas...

  1. Evaluation of a metal shear web selectively reinforced with filamentary composites for space shuttle application

    NASA Technical Reports Server (NTRS)

    Laakso, J. H.; Straayer, J. W.

    1974-01-01

    A final program summary is reported for test and evaluation activities that were conducted for space shuttle web selection. Large scale advanced composite shear web components were tested and analyzed to evaluate application of advanced composite shear web construction to a space shuttle orbiter thrust structure. The shear web design concept consisted of a titanium-clad + or - 45 deg boron/epoxy web laminate stiffened with vertical boron-epoxy reinforced aluminum stiffeners and logitudinal aluminum stiffening. The design concept was evaluated to be efficient and practical for the application that was studied. Because of the effects of buckling deflections, a requirement is identified for shear buckling resistant design to maximize the efficiency of highly-loaded advanced composite shear webs.

  2. River Food Web Response to Large-Scale Riparian Zone Manipulations

    PubMed Central

    Wootton, J. Timothy

    2012-01-01

    Conservation programs often focus on select species, leading to management plans based on the autecology of the focal species, but multiple ecosystem components can be affected both by the environmental factors impacting, and the management targeting, focal species. These broader effects can have indirect impacts on target species through the web of interactions within ecosystems. For example, human activity can strongly alter riparian vegetation, potentially impacting both economically-important salmonids and their associated river food web. In an Olympic Peninsula river, Washington state, USA, replicated large-scale riparian vegetation manipulations implemented with the long-term (>40 yr) goal of improving salmon habitat did not affect water temperature, nutrient limitation or habitat characteristics, but reduced canopy cover, causing reduced energy input via leaf litter, increased incident solar radiation (UV and PAR) and increased algal production compared to controls. In response, benthic algae, most insect taxa, and juvenile salmonids increased in manipulated areas. Stable isotope analysis revealed a predominant contribution of algal-derived energy to salmonid diets in manipulated reaches. The experiment demonstrates that riparian management targeting salmonids strongly affects river food webs via changes in the energy base, illustrates how species-based management strategies can have unanticipated indirect effects on the target species via the associated food web, and supports ecosystem-based management approaches for restoring depleted salmonid stocks. PMID:23284786

  3. ‘Natural experiment’ Demonstrates Top-Down Control of Spiders by Birds on a Landscape Level

    PubMed Central

    Rogers, Haldre; Hille Ris Lambers, Janneke; Miller, Ross; Tewksbury, Joshua J.

    2012-01-01

    The combination of small-scale manipulative experiments and large-scale natural experiments provides a powerful approach for demonstrating the importance of top-down trophic control on the ecosystem scale. The most compelling natural experiments have come from studies examining the landscape-scale loss of apex predators like sea otters, wolves, fish and land crabs. Birds are dominant apex predators in terrestrial systems around the world, yet all studies on their role as predators have come from small-scale experiments; the top-down impact of bird loss on their arthropod prey has yet to be examined at a landscape scale. Here, we use a unique natural experiment, the extirpation of insectivorous birds from nearly all forests on the island of Guam by the invasive brown tree snake, to produce the first assessment of the impacts of bird loss on their prey. We focused on spiders because experimental studies showed a consistent top-down effect of birds on spiders. We conducted spider web surveys in native forest on Guam and three nearby islands with healthy bird populations. Spider web densities on the island of Guam were 40 times greater than densities on islands with birds during the wet season, and 2.3 times greater during the dry season. These results confirm the general trend from manipulative experiments conducted in other systems however, the effect size was much greater in this natural experiment than in most manipulative experiments. In addition, bird loss appears to have removed the seasonality of spider webs and led to larger webs in at least one spider species in the forests of Guam than on nearby islands with birds. We discuss several possible mechanisms for the observed changes. Overall, our results suggest that effect sizes from smaller-scale experimental studies may significantly underestimate the impact of bird loss on spider density as demonstrated by this large-scale natural experiment. PMID:22970126

  4. Openwebglobe 2: Visualization of Complex 3D-GEODATA in the (mobile) Webbrowser

    NASA Astrophysics Data System (ADS)

    Christen, M.

    2016-06-01

    Providing worldwide high resolution data for virtual globes consists of compute and storage intense tasks for processing data. Furthermore, rendering complex 3D-Geodata, such as 3D-City models with an extremely high polygon count and a vast amount of textures at interactive framerates is still a very challenging task, especially on mobile devices. This paper presents an approach for processing, caching and serving massive geospatial data in a cloud-based environment for large scale, out-of-core, highly scalable 3D scene rendering on a web based virtual globe. Cloud computing is used for processing large amounts of geospatial data and also for providing 2D and 3D map data to a large amount of (mobile) web clients. In this paper the approach for processing, rendering and caching very large datasets in the currently developed virtual globe "OpenWebGlobe 2" is shown, which displays 3D-Geodata on nearly every device.

  5. The pepATTRACT web server for blind, large-scale peptide-protein docking.

    PubMed

    de Vries, Sjoerd J; Rey, Julien; Schindler, Christina E M; Zacharias, Martin; Tuffery, Pierre

    2017-07-03

    Peptide-protein interactions are ubiquitous in the cell and form an important part of the interactome. Computational docking methods can complement experimental characterization of these complexes, but current protocols are not applicable on the proteome scale. pepATTRACT is a novel docking protocol that is fully blind, i.e. it does not require any information about the binding site. In various stages of its development, pepATTRACT has participated in CAPRI, making successful predictions for five out of seven protein-peptide targets. Its performance is similar or better than state-of-the-art local docking protocols that do require binding site information. Here we present a novel web server that carries out the rigid-body stage of pepATTRACT. On the peptiDB benchmark, the web server generates a correct model in the top 50 in 34% of the cases. Compared to the full pepATTRACT protocol, this leads to some loss of performance, but the computation time is reduced from ∼18 h to ∼10 min. Combined with the fact that it is fully blind, this makes the web server well-suited for large-scale in silico protein-peptide docking experiments. The rigid-body pepATTRACT server is freely available at http://bioserv.rpbs.univ-paris-diderot.fr/services/pepATTRACT. © The Author(s) 2017. Published by Oxford University Press on behalf of Nucleic Acids Research.

  6. Host Immunity via Mutable Virtualized Large-Scale Network Containers

    DTIC Science & Technology

    2016-07-25

    and constrain the distributed persistent inside crawlers that have va.lid credentials to access the web services. The main idea is to add a marker...to each web page URL and use the URL path and user inforn1ation contained in the marker to help accurately detect crawlers at its earliest stage...more than half of all website traffic, and malicious bots contributes almost one third of the traffic. As one type of bots, web crawlers have been

  7. The Department of Defense and the Power of Cloud Computing: Weighing Acceptable Cost Versus Acceptable Risk

    DTIC Science & Technology

    2016-04-01

    the DOD will put DOD systems and data at a risk level comparable to that of their neighbors in the cloud. Just as a user browses a Web page on the...proxy servers for controlling user access to Web pages, and large-scale storage for data management. Each of these devices allows access to the...user to develop applications. Acunetics.com describes Web applications as “computer programs allowing Website visitors to submit and retrieve data

  8. ΛGR Centennial: Cosmic Web in Dark Energy Background

    NASA Astrophysics Data System (ADS)

    Chernin, A. D.

    The basic building blocks of the Cosmic Web are groups and clusters of galaxies, super-clusters (pancakes) and filaments embedded in the universal dark energy background. The background produces antigravity, and the antigravity effect is strong in groups, clusters and superclusters. Antigravity is very weak in filaments where matter (dark matter and baryons) produces gravity dominating in the filament internal dynamics. Gravity-antigravity interplay on the large scales is a grandiose phenomenon predicted by ΛGR theory and seen in modern observations of the Cosmic Web.

  9. Interactive Profiler: An Intuitive, Web-Based Statistical Application in Visualizing Educational and Marketing Databases

    ERIC Educational Resources Information Center

    Ip, Edward H.; Leung, Phillip; Johnson, Joseph

    2004-01-01

    We describe the design and implementation of a web-based statistical program--the Interactive Profiler (IP). The prototypical program, developed in Java, was motivated by the need for the general public to query against data collected from the National Assessment of Educational Progress (NAEP), a large-scale US survey of the academic state of…

  10. Observations of a nearby filament of galaxy clusters with the Sardinia Radio Telescope

    NASA Astrophysics Data System (ADS)

    Vacca, Valentina; Murgia, M.; Loi, F. Govoni F.; Vazza, F.; Finoguenov, A.; Carretti, E.; Feretti, L.; Giovannini, G.; Concu, R.; Melis, A.; Gheller, C.; Paladino, R.; Poppi, S.; Valente, G.; Bernardi, G.; Boschin, W.; Brienza, M.; Clarke, T. E.; Colafrancesco, S.; Enßlin, T.; Ferrari, C.; de Gasperin, F.; Gastaldello, F.; Girardi, M.; Gregorini, L.; Johnston-Hollitt, M.; Junklewitz, H.; Orrù, E.; Parma, P.; Perley, R.; Taylor, G. B.

    2018-05-01

    We report the detection of diffuse radio emission which might be connected to a large-scale filament of the cosmic web covering a 8° × 8° area in the sky, likely associated with a z≈0.1 over-density traced by nine massive galaxy clusters. In this work, we present radio observations of this region taken with the Sardinia Radio Telescope. Two of the clusters in the field host a powerful radio halo sustained by violent ongoing mergers and provide direct proof of intra-cluster magnetic fields. In order to investigate the presence of large-scale diffuse radio synchrotron emission in and beyond the galaxy clusters in this complex system, we combined the data taken at 1.4 GHz with the Sardinia Radio Telescope with higher resolution data taken with the NRAO VLA Sky Survey. We found 28 candidate new sources with a size larger and X-ray emission fainter than known diffuse large-scale synchrotron cluster sources for a given radio power. This new population is potentially the tip of the iceberg of a class of diffuse large-scale synchrotron sources associated with the filaments of the cosmic web. In addition, we found in the field a candidate new giant radio galaxy.

  11. Massive gravity wrapped in the cosmic web

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Shim, Junsup; Lee, Jounghun; Li, Baojiu, E-mail: jsshim@astro.snu.ac.kr, E-mail: jounghun@astro.snu.ac.kr

    We study how the filamentary pattern of the cosmic web changes if the true gravity deviates from general relativity (GR) on a large scale. The f(R) gravity, whose strength is controlled to satisfy the current observational constraints on the cluster scale, is adopted as our fiducial model and a large, high-resolution N-body simulation is utilized for this study. By applying the minimal spanning tree algorithm to the halo catalogs from the simulation at various epochs, we identify the main stems of the rich superclusters located in the most prominent filamentary section of the cosmic web and determine their spatial extentsmore » per member cluster to be the degree of their straightness. It is found that the f(R) gravity has the effect of significantly bending the superclusters and that the effect becomes stronger as the universe evolves. Even in the case where the deviation from GR is too small to be detectable by any other observables, the degree of the supercluster straightness exhibits a conspicuous difference between the f(R) and the GR models. Our results also imply that the supercluster straightness could be a useful discriminator of f(R) gravity from the coupled dark energy since it is shown to evolve differently between the two models. As a final conclusion, the degree of the straightness of the rich superclusters should provide a powerful cosmological test of large scale gravity.« less

  12. mySyntenyPortal: an application package to construct websites for synteny block analysis.

    PubMed

    Lee, Jongin; Lee, Daehwan; Sim, Mikang; Kwon, Daehong; Kim, Juyeon; Ko, Younhee; Kim, Jaebum

    2018-06-05

    Advances in sequencing technologies have facilitated large-scale comparative genomics based on whole genome sequencing. Constructing and investigating conserved genomic regions among multiple species (called synteny blocks) are essential in the comparative genomics. However, they require significant amounts of computational resources and time in addition to bioinformatics skills. Many web interfaces have been developed to make such tasks easier. However, these web interfaces cannot be customized for users who want to use their own set of genome sequences or definition of synteny blocks. To resolve this limitation, we present mySyntenyPortal, a stand-alone application package to construct websites for synteny block analyses by using users' own genome data. mySyntenyPortal provides both command line and web-based interfaces to build and manage websites for large-scale comparative genomic analyses. The websites can be also easily published and accessed by other users. To demonstrate the usability of mySyntenyPortal, we present an example study for building websites to compare genomes of three mammalian species (human, mouse, and cow) and show how they can be easily utilized to identify potential genes affected by genome rearrangements. mySyntenyPortal will contribute for extended comparative genomic analyses based on large-scale whole genome sequences by providing unique functionality to support the easy creation of interactive websites for synteny block analyses from user's own genome data.

  13. web cellHTS2: a web-application for the analysis of high-throughput screening data.

    PubMed

    Pelz, Oliver; Gilsdorf, Moritz; Boutros, Michael

    2010-04-12

    The analysis of high-throughput screening data sets is an expanding field in bioinformatics. High-throughput screens by RNAi generate large primary data sets which need to be analyzed and annotated to identify relevant phenotypic hits. Large-scale RNAi screens are frequently used to identify novel factors that influence a broad range of cellular processes, including signaling pathway activity, cell proliferation, and host cell infection. Here, we present a web-based application utility for the end-to-end analysis of large cell-based screening experiments by cellHTS2. The software guides the user through the configuration steps that are required for the analysis of single or multi-channel experiments. The web-application provides options for various standardization and normalization methods, annotation of data sets and a comprehensive HTML report of the screening data analysis, including a ranked hit list. Sessions can be saved and restored for later re-analysis. The web frontend for the cellHTS2 R/Bioconductor package interacts with it through an R-server implementation that enables highly parallel analysis of screening data sets. web cellHTS2 further provides a file import and configuration module for common file formats. The implemented web-application facilitates the analysis of high-throughput data sets and provides a user-friendly interface. web cellHTS2 is accessible online at http://web-cellHTS2.dkfz.de. A standalone version as a virtual appliance and source code for platforms supporting Java 1.5.0 can be downloaded from the web cellHTS2 page. web cellHTS2 is freely distributed under GPL.

  14. Changes in the High-Latitude Topside Ionospheric Vertical Electron-Density Profiles in Response to Solar-Wind Perturbations During Large Magnetic Storms

    NASA Technical Reports Server (NTRS)

    Benson, Robert F.; Fainberg, Joseph; Osherovich, Vladimir; Truhlik, Vladimir; Wang, Yongli; Arbacher, Becca

    2011-01-01

    The latest results from an investigation to establish links between solar-wind and topside-ionospheric parameters will be presented including a case where high-latitude topside electron-density Ne(h) profiles indicated dramatic rapid changes in the scale height during the main phase of a large magnetic storm (Dst < -200 nT). These scale-height changes suggest a large heat input to the topside ionosphere at this time. The topside profiles were derived from ISIS-1 digital ionograms obtained from the NASA Space Physics Data Facility (SPDF) Coordinated Data Analysis Web (CDA Web). Solar-wind data obtained from the NASA OMNIWeb database indicated that the magnetic storm was due to a magnetic cloud. This event is one of several large magnetic storms being investigated during the interval from 1965 to 1984 when both solar-wind and digital topside ionograms, from either Alouette-2, ISIS-1, or ISIS-2, are potentially available.

  15. Revealing the z ~ 2.5 Cosmic Web with 3D Lyα Forest Tomography: a Deformation Tensor Approach

    NASA Astrophysics Data System (ADS)

    Lee, Khee-Gan; White, Martin

    2016-11-01

    Studies of cosmological objects should take into account their positions within the cosmic web of large-scale structure. Unfortunately, the cosmic web has only been extensively mapped at low redshifts (z\\lt 1), using galaxy redshifts as tracers of the underlying density field. At z\\gt 1, the required galaxy densities are inaccessible for the foreseeable future, but 3D reconstructions of Lyα forest absorption in closely separated background QSOs and star-forming galaxies already offer a detailed window into z˜ 2-3 large-scale structure. We quantify the utility of such maps for studying the cosmic web by using realistic z = 2.5 Lyα forest simulations matched to observational properties of upcoming surveys. A deformation tensor-based analysis is used to classify voids, sheets, filaments, and nodes in the flux, which are compared to those determined from the underlying dark matter (DM) field. We find an extremely good correspondence, with 70% of the volume in the flux maps correctly classified relative to the DM web, and 99% classified to within one eigenvalue. This compares favorably to the performance of galaxy-based classifiers with even the highest galaxy densities from low-redshift surveys. We find that narrow survey geometries can degrade the recovery of the cosmic web unless the survey is ≳ 60 {h}-1 {Mpc} or ≳ 1 deg on the sky. We also examine halo abundances as a function of the cosmic web, and find a clear dependence as a function of flux overdensity, but little explicit dependence on the cosmic web. These methods will provide a new window on cosmological environments of galaxies at this very special time in galaxy formation, “high noon,” and on overall properties of cosmological structures at this epoch.

  16. GeoCENS: a geospatial cyberinfrastructure for the world-wide sensor web.

    PubMed

    Liang, Steve H L; Huang, Chih-Yuan

    2013-10-02

    The world-wide sensor web has become a very useful technique for monitoring the physical world at spatial and temporal scales that were previously impossible. Yet we believe that the full potential of sensor web has thus far not been revealed. In order to harvest the world-wide sensor web's full potential, a geospatial cyberinfrastructure is needed to store, process, and deliver large amount of sensor data collected worldwide. In this paper, we first define the issue of the sensor web long tail followed by our view of the world-wide sensor web architecture. Then, we introduce the Geospatial Cyberinfrastructure for Environmental Sensing (GeoCENS) architecture and explain each of its components. Finally, with demonstration of three real-world powered-by-GeoCENS sensor web applications, we believe that the GeoCENS architecture can successfully address the sensor web long tail issue and consequently realize the world-wide sensor web vision.

  17. GeoCENS: A Geospatial Cyberinfrastructure for the World-Wide Sensor Web

    PubMed Central

    Liang, Steve H.L.; Huang, Chih-Yuan

    2013-01-01

    The world-wide sensor web has become a very useful technique for monitoring the physical world at spatial and temporal scales that were previously impossible. Yet we believe that the full potential of sensor web has thus far not been revealed. In order to harvest the world-wide sensor web's full potential, a geospatial cyberinfrastructure is needed to store, process, and deliver large amount of sensor data collected worldwide. In this paper, we first define the issue of the sensor web long tail followed by our view of the world-wide sensor web architecture. Then, we introduce the Geospatial Cyberinfrastructure for Environmental Sensing (GeoCENS) architecture and explain each of its components. Finally, with demonstration of three real-world powered-by-GeoCENS sensor web applications, we believe that the GeoCENS architecture can successfully address the sensor web long tail issue and consequently realize the world-wide sensor web vision. PMID:24152921

  18. The cosmic web in CosmoGrid void regions

    NASA Astrophysics Data System (ADS)

    Rieder, Steven; van de Weygaert, Rien; Cautun, Marius; Beygu, Burcu; Portegies Zwart, Simon

    2016-10-01

    We study the formation and evolution of the cosmic web, using the high-resolution CosmoGrid ΛCDM simulation. In particular, we investigate the evolution of the large-scale structure around void halo groups, and compare this to observations of the VGS-31 galaxy group, which consists of three interacting galaxies inside a large void. The structure around such haloes shows a great deal of tenuous structure, with most of such systems being embedded in intra-void filaments and walls. We use the Nexus+} algorithm to detect walls and filaments in CosmoGrid, and find them to be present and detectable at every scale. The void regions embed tenuous walls, which in turn embed tenuous filaments. We hypothesize that the void galaxy group of VGS-31 formed in such an environment.

  19. THUIR at TREC 2009 Web Track: Finding Relevant and Diverse Results for Large Scale Web Search

    DTIC Science & Technology

    2009-11-01

    Porn words‟ filtering is also one of the anti-spam techniques in real world search engines. A list of porn words was found from the internet [2...When the numbers of the porn words in the page is larger than α, then the page is taken as the spam. In our experiments, the threshold is set to 16

  20. Post Graduations in Technologies and Computing Applied to Education: From F2F Classes to Multimedia Online Open Courses

    ERIC Educational Resources Information Center

    Marques, Bertil P.; Carvalho, Piedade; Escudeiro, Paula; Barata, Ana; Silva, Ana; Queiros, Sandra

    2017-01-01

    Promoted by the significant increase of large scale internet access, many audiences have turned to the web and to its resources for learning and inspiration, with diverse sets of skills and intents. In this context, Multimedia Online Open Courses (MOOC) consist in learning models supported on user-friendly web tools that allow anyone with minimum…

  1. Correlations, Trends and Potential Biases among Publicly Accessible Web-Based Student Evaluations of Teaching: A Large-Scale Study of RateMyProfessors.com Data

    ERIC Educational Resources Information Center

    Rosen, Andrew S.

    2018-01-01

    Student evaluations of teaching are widely adopted across academic institutions, but there are many underlying trends and biases that can influence their interpretation. Publicly accessible web-based student evaluations of teaching are of particular relevance, due to their widespread use by students in the course selection process and the quantity…

  2. BioPlex Display: An Interactive Suite for Large-Scale AP-MS Protein-Protein Interaction Data.

    PubMed

    Schweppe, Devin K; Huttlin, Edward L; Harper, J Wade; Gygi, Steven P

    2018-01-05

    The development of large-scale data sets requires a new means to display and disseminate research studies to large audiences. Knowledge of protein-protein interaction (PPI) networks has become a principle interest of many groups within the field of proteomics. At the confluence of technologies, such as cross-linking mass spectrometry, yeast two-hybrid, protein cofractionation, and affinity purification mass spectrometry (AP-MS), detection of PPIs can uncover novel biological inferences at a high-throughput. Thus new platforms to provide community access to large data sets are necessary. To this end, we have developed a web application that enables exploration and dissemination of the growing BioPlex interaction network. BioPlex is a large-scale interactome data set based on AP-MS of baits from the human ORFeome. The latest BioPlex data set release (BioPlex 2.0) contains 56 553 interactions from 5891 AP-MS experiments. To improve community access to this vast compendium of interactions, we developed BioPlex Display, which integrates individual protein querying, access to empirical data, and on-the-fly annotation of networks within an easy-to-use and mobile web application. BioPlex Display enables rapid acquisition of data from BioPlex and development of hypotheses based on protein interactions.

  3. A Ubiquitous Sensor Network Platform for Integrating Smart Devices into the Semantic Sensor Web

    PubMed Central

    de Vera, David Díaz Pardo; Izquierdo, Álvaro Sigüenza; Vercher, Jesús Bernat; Gómez, Luis Alfonso Hernández

    2014-01-01

    Ongoing Sensor Web developments make a growing amount of heterogeneous sensor data available to smart devices. This is generating an increasing demand for homogeneous mechanisms to access, publish and share real-world information. This paper discusses, first, an architectural solution based on Next Generation Networks: a pilot Telco Ubiquitous Sensor Network (USN) Platform that embeds several OGC® Sensor Web services. This platform has already been deployed in large scale projects. Second, the USN-Platform is extended to explore a first approach to Semantic Sensor Web principles and technologies, so that smart devices can access Sensor Web data, allowing them also to share richer (semantically interpreted) information. An experimental scenario is presented: a smart car that consumes and produces real-world information which is integrated into the Semantic Sensor Web through a Telco USN-Platform. Performance tests revealed that observation publishing times with our experimental system were well within limits compatible with the adequate operation of smart safety assistance systems in vehicles. On the other hand, response times for complex queries on large repositories may be inappropriate for rapid reaction needs. PMID:24945678

  4. A ubiquitous sensor network platform for integrating smart devices into the semantic sensor web.

    PubMed

    de Vera, David Díaz Pardo; Izquierdo, Alvaro Sigüenza; Vercher, Jesús Bernat; Hernández Gómez, Luis Alfonso

    2014-06-18

    Ongoing Sensor Web developments make a growing amount of heterogeneous sensor data available to smart devices. This is generating an increasing demand for homogeneous mechanisms to access, publish and share real-world information. This paper discusses, first, an architectural solution based on Next Generation Networks: a pilot Telco Ubiquitous Sensor Network (USN) Platform that embeds several OGC® Sensor Web services. This platform has already been deployed in large scale projects. Second, the USN-Platform is extended to explore a first approach to Semantic Sensor Web principles and technologies, so that smart devices can access Sensor Web data, allowing them also to share richer (semantically interpreted) information. An experimental scenario is presented: a smart car that consumes and produces real-world information which is integrated into the Semantic Sensor Web through a Telco USN-Platform. Performance tests revealed that observation publishing times with our experimental system were well within limits compatible with the adequate operation of smart safety assistance systems in vehicles. On the other hand, response times for complex queries on large repositories may be inappropriate for rapid reaction needs.

  5. miRNAFold: a web server for fast miRNA precursor prediction in genomes.

    PubMed

    Tav, Christophe; Tempel, Sébastien; Poligny, Laurent; Tahi, Fariza

    2016-07-08

    Computational methods are required for prediction of non-coding RNAs (ncRNAs), which are involved in many biological processes, especially at post-transcriptional level. Among these ncRNAs, miRNAs have been largely studied and biologists need efficient and fast tools for their identification. In particular, ab initio methods are usually required when predicting novel miRNAs. Here we present a web server dedicated for miRNA precursors identification at a large scale in genomes. It is based on an algorithm called miRNAFold that allows predicting miRNA hairpin structures quickly with high sensitivity. miRNAFold is implemented as a web server with an intuitive and user-friendly interface, as well as a standalone version. The web server is freely available at: http://EvryRNA.ibisc.univ-evry.fr/miRNAFold. © The Author(s) 2016. Published by Oxford University Press on behalf of Nucleic Acids Research.

  6. Augmented Reality 2.0

    NASA Astrophysics Data System (ADS)

    Schmalstieg, Dieter; Langlotz, Tobias; Billinghurst, Mark

    Augmented Reality (AR) was first demonstrated in the 1960s, but only recently have technologies emerged that can be used to easily deploy AR applications to many users. Camera-equipped cell phones with significant processing power and graphics abilities provide an inexpensive and versatile platform for AR applications, while the social networking technology of Web 2.0 provides a large-scale infrastructure for collaboratively producing and distributing geo-referenced AR content. This combination of widely used mobile hardware and Web 2.0 software allows the development of a new type of AR platform that can be used on a global scale. In this paper we describe the Augmented Reality 2.0 concept and present existing work on mobile AR and web technologies that could be used to create AR 2.0 applications.

  7. Nexus of the Cosmic Web

    NASA Astrophysics Data System (ADS)

    Cautun, Marius; van de Weygaert, Rien; Jones, Bernard J. T.; Frenk, Carlos S.; Hellwing, Wojciech A.

    2015-01-01

    One of the important unknowns of current cosmology concerns the effects of the large scale distribution of matter on the formation and evolution of dark matter haloes and galaxies. One main difficulty in answering this question lies in the absence of a robust and natural way of identifying the large scale environments and their characteristics. This work summarizes the NEXUS+ formalism which extends and improves our multiscale scale-space MMF method. The new algorithm is very successful in tracing the Cosmic Web components, mainly due to its novel filtering of the density in logarithmic space. The method, due to its multiscale and hierarchical character, has the advantage of detecting all the cosmic structures, either prominent or tenuous, without preference for a certain size or shape. The resulting filamentary and wall networks can easily be characterized by their direction, thickness, mass density and density profile. These additional environmental properties allows to us to investigate not only the effect of environment on haloes, but also how it correlates with the environment characteristics.

  8. Mapping Dark Matter in Simulated Galaxy Clusters

    NASA Astrophysics Data System (ADS)

    Bowyer, Rachel

    2018-01-01

    Galaxy clusters are the most massive bound objects in the Universe with most of their mass being dark matter. Cosmological simulations of structure formation show that clusters are embedded in a cosmic web of dark matter filaments and large scale structure. It is thought that these filaments are found preferentially close to the long axes of clusters. We extract galaxy clusters from the simulations "cosmo-OWLS" in order to study their properties directly and also to infer their properties from weak gravitational lensing signatures. We investigate various stacking procedures to enhance the signal of the filaments and large scale structure surrounding the clusters to better understand how the filaments of the cosmic web connect with galaxy clusters. This project was supported in part by the NSF REU grant AST-1358980 and by the Nantucket Maria Mitchell Association.

  9. Web Archiving for the Rest of Us: How to Collect and Manage Websites Using Free and Easy Software

    ERIC Educational Resources Information Center

    Dunn, Katharine; Szydlowski, Nick

    2009-01-01

    Large-scale projects such as the Internet Archive (www.archive.org) send out crawlers to gather snapshots of much of the web. This massive collection of archived websites may include content of interest to one's patrons. But if librarians want to control exactly when and what is archived, relying on someone else to do the archiving is not ideal.…

  10. The cosmic web and the orientation of angular momenta

    NASA Astrophysics Data System (ADS)

    Libeskind, Noam I.; Hoffman, Yehuda; Knebe, Alexander; Steinmetz, Matthias; Gottlöber, Stefan; Metuki, Ofer; Yepes, Gustavo

    2012-03-01

    We use a 64 h-1 Mpc dark-matter-only cosmological simulation to examine the large-scale orientation of haloes and substructures with respect to the cosmic web. A web classification scheme based on the velocity shear tensor is used to assign to each halo in the simulation a web type: knot, filament, sheet or void. Using ˜106 haloes that span ˜3 orders of magnitude in mass, the orientation of the halo's spin and the orbital angular momentum of subhaloes with respect to the eigenvectors of the shear tensor is examined. We find that the orbital angular momentum of subhaloes tends to align with the intermediate eigenvector of the velocity shear tensor for all haloes in knots, filaments and sheets. This result indicates that the kinematics of substructures located deep within the virialized regions of a halo is determined by its infall which in turn is determined by the large-scale velocity shear, a surprising result given the virialized nature of haloes. The non-random nature of subhalo accretion is thus imprinted on the angular momentum measured at z= 0. We also find that the haloes' spin axis is aligned with the third eigenvector of the velocity shear tensor in filaments and sheets: the halo spin axis points along filaments and lies in the plane of cosmic sheets.

  11. Is the S-Web the Secret to Observed Heliospheric Particle Distributions?

    NASA Astrophysics Data System (ADS)

    Higginson, A. K.; Antiochos, S. K.; DeVore, C. R.; Daldorff, L. K. S.; Wyper, P. F.; Ukhorskiy, A. Y.; Sorathia, K.

    2017-12-01

    Particle transport in the heliosphere remains an unsolved problem across energy regimes. Observations of slow solar wind show that plasma escapes from the closed-field corona, but ends up far away from the heliospheric current sheet, even though the release mechanisms are expected to occur at the HCS. Similarly, some impulsive SEP events have extreme longitudinal extents of 100 degrees or more. Recent theoretical and numerical work has shown that interchange reconnection near a coronal-hole corridor can release plasma from originally closed magnetic field lines into a large swath spread across the heliosphere, forming what is known as an S-Web arc. This is a promising mechanism for explaining both the slow solar wind, with its large latitudinal extent, and impulsive SEP particles, with their large longitudinal extent. Here we compute, for the first time, the dynamics of the S-Web when the photospheric driver is applied over a large portion of the solar surface compared to the scale of the driving. We examine the time scales for the interchange reconnection and compute the angular extent of the plasma released, in the context of understanding both the slow solar wind and flare-accelerated SEPs. We will make predictions for Solar Orbiter and Parker Solar Probe and discuss how these new measurements will help to both pinpoint the source of the slow solar wind and illuminate the transport mechanisms of wide-spread impulsive SEP events.

  12. HEPCloud, a New Paradigm for HEP Facilities: CMS Amazon Web Services Investigation

    DOE PAGES

    Holzman, Burt; Bauerdick, Lothar A. T.; Bockelman, Brian; ...

    2017-09-29

    Historically, high energy physics computing has been performed on large purpose-built computing systems. These began as single-site compute facilities, but have evolved into the distributed computing grids used today. Recently, there has been an exponential increase in the capacity and capability of commercial clouds. Cloud resources are highly virtualized and intended to be able to be flexibly deployed for a variety of computing tasks. There is a growing interest among the cloud providers to demonstrate the capability to perform large-scale scientific computing. In this paper, we discuss results from the CMS experiment using the Fermilab HEPCloud facility, which utilized bothmore » local Fermilab resources and virtual machines in the Amazon Web Services Elastic Compute Cloud. We discuss the planning, technical challenges, and lessons learned involved in performing physics workflows on a large-scale set of virtualized resources. Additionally, we will discuss the economics and operational efficiencies when executing workflows both in the cloud and on dedicated resources.« less

  13. HEPCloud, a New Paradigm for HEP Facilities: CMS Amazon Web Services Investigation

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Holzman, Burt; Bauerdick, Lothar A. T.; Bockelman, Brian

    Historically, high energy physics computing has been performed on large purpose-built computing systems. These began as single-site compute facilities, but have evolved into the distributed computing grids used today. Recently, there has been an exponential increase in the capacity and capability of commercial clouds. Cloud resources are highly virtualized and intended to be able to be flexibly deployed for a variety of computing tasks. There is a growing interest among the cloud providers to demonstrate the capability to perform large-scale scientific computing. In this paper, we discuss results from the CMS experiment using the Fermilab HEPCloud facility, which utilized bothmore » local Fermilab resources and virtual machines in the Amazon Web Services Elastic Compute Cloud. We discuss the planning, technical challenges, and lessons learned involved in performing physics workflows on a large-scale set of virtualized resources. Additionally, we will discuss the economics and operational efficiencies when executing workflows both in the cloud and on dedicated resources.« less

  14. Using Syntactic Patterns to Enhance Text Analytics

    ERIC Educational Resources Information Center

    Meyer, Bradley B.

    2017-01-01

    Large scale product and service reviews proliferate and are commonly found across the web. The ability to harvest, digest and analyze a large corpus of reviews from online websites is still however a difficult problem. This problem is referred to as "opinion mining." Opinion mining is an important area of research as advances in the…

  15. IMAGE EXPLORER: Astronomical Image Analysis on an HTML5-based Web Application

    NASA Astrophysics Data System (ADS)

    Gopu, A.; Hayashi, S.; Young, M. D.

    2014-05-01

    Large datasets produced by recent astronomical imagers cause the traditional paradigm for basic visual analysis - typically downloading one's entire image dataset and using desktop clients like DS9, Aladin, etc. - to not scale, despite advances in desktop computing power and storage. This paper describes Image Explorer, a web framework that offers several of the basic visualization and analysis functionality commonly provided by tools like DS9, on any HTML5 capable web browser on various platforms. It uses a combination of the modern HTML5 canvas, JavaScript, and several layers of lossless PNG tiles producted from the FITS image data. Astronomers are able to rapidly and simultaneously open up several images on their web-browser, adjust the intensity min/max cutoff or its scaling function, and zoom level, apply color-maps, view position and FITS header information, execute typically used data reduction codes on the corresponding FITS data using the FRIAA framework, and overlay tiles for source catalog objects, etc.

  16. Improving data workflow systems with cloud services and use of open data for bioinformatics research.

    PubMed

    Karim, Md Rezaul; Michel, Audrey; Zappa, Achille; Baranov, Pavel; Sahay, Ratnesh; Rebholz-Schuhmann, Dietrich

    2017-04-16

    Data workflow systems (DWFSs) enable bioinformatics researchers to combine components for data access and data analytics, and to share the final data analytics approach with their collaborators. Increasingly, such systems have to cope with large-scale data, such as full genomes (about 200 GB each), public fact repositories (about 100 TB of data) and 3D imaging data at even larger scales. As moving the data becomes cumbersome, the DWFS needs to embed its processes into a cloud infrastructure, where the data are already hosted. As the standardized public data play an increasingly important role, the DWFS needs to comply with Semantic Web technologies. This advancement to DWFS would reduce overhead costs and accelerate the progress in bioinformatics research based on large-scale data and public resources, as researchers would require less specialized IT knowledge for the implementation. Furthermore, the high data growth rates in bioinformatics research drive the demand for parallel and distributed computing, which then imposes a need for scalability and high-throughput capabilities onto the DWFS. As a result, requirements for data sharing and access to public knowledge bases suggest that compliance of the DWFS with Semantic Web standards is necessary. In this article, we will analyze the existing DWFS with regard to their capabilities toward public open data use as well as large-scale computational and human interface requirements. We untangle the parameters for selecting a preferable solution for bioinformatics research with particular consideration to using cloud services and Semantic Web technologies. Our analysis leads to research guidelines and recommendations toward the development of future DWFS for the bioinformatics research community. © The Author 2017. Published by Oxford University Press.

  17. Is the universe a sponge?

    NASA Astrophysics Data System (ADS)

    Bucher, Martin

    2016-11-01

    Does the large-scale universe look more like meatballs, like Swiss cheese or like a sponge? The differences between these types of universe are described in J Richard Gott's The Cosmic Web: Mysterious Architecture of the Universe.

  18. Ontology-guided organ detection to retrieve web images of disease manifestation: towards the construction of a consumer-based health image library.

    PubMed

    Chen, Yang; Ren, Xiaofeng; Zhang, Guo-Qiang; Xu, Rong

    2013-01-01

    Visual information is a crucial aspect of medical knowledge. Building a comprehensive medical image base, in the spirit of the Unified Medical Language System (UMLS), would greatly benefit patient education and self-care. However, collection and annotation of such a large-scale image base is challenging. To combine visual object detection techniques with medical ontology to automatically mine web photos and retrieve a large number of disease manifestation images with minimal manual labeling effort. As a proof of concept, we first learnt five organ detectors on three detection scales for eyes, ears, lips, hands, and feet. Given a disease, we used information from the UMLS to select affected body parts, ran the pretrained organ detectors on web images, and combined the detection outputs to retrieve disease images. Compared with a supervised image retrieval approach that requires training images for every disease, our ontology-guided approach exploits shared visual information of body parts across diseases. In retrieving 2220 web images of 32 diseases, we reduced manual labeling effort to 15.6% while improving the average precision by 3.9% from 77.7% to 81.6%. For 40.6% of the diseases, we improved the precision by 10%. The results confirm the concept that the web is a feasible source for automatic disease image retrieval for health image database construction. Our approach requires a small amount of manual effort to collect complex disease images, and to annotate them by standard medical ontology terms.

  19. Large-scale fabrication of bioinspired fibers for directional water collection.

    PubMed

    Bai, Hao; Sun, Ruize; Ju, Jie; Yao, Xi; Zheng, Yongmei; Jiang, Lei

    2011-12-16

    Spider-silk inspired functional fibers with periodic spindle-knots and the ability to collect water in a directional manner are fabricated on a large scale using a fluid coating method. The fabrication process is investigated in detail, considering factors like the fiber-drawing velocity, solution viscosity, and surface tension. These bioinspired fibers are inexpensive and durable, which makes it possible to collect water from fog in a similar manner to a spider's web. Copyright © 2011 WILEY-VCH Verlag GmbH & Co. KGaA, Weinheim.

  20. Strategic Alliances in Education: The Knowledge Engineering Web

    ERIC Educational Resources Information Center

    Westera, Wim; van den Herik, Jaap; van de Vrie, Evert

    2004-01-01

    The field of higher education shows a jumble of alliances between fellow institutes. The alliances are strategic in kind and serve an economy-of-scales concept. A large scale is a prerequisite for allocating the budgets for new educational methods and technologies in order to keep the educational services up-to-date. All too often, however,…

  1. Novel Web-based Education Platforms for Information Communication utilizing Gamification, Virtual and Immersive Reality

    NASA Astrophysics Data System (ADS)

    Demir, I.

    2015-12-01

    Recent developments in internet technologies make it possible to manage and visualize large data on the web. Novel visualization techniques and interactive user interfaces allow users to create realistic environments, and interact with data to gain insight from simulations and environmental observations. This presentation showcase information communication interfaces, games, and virtual and immersive reality applications for supporting teaching and learning of concepts in atmospheric and hydrological sciences. The information communication platforms utilizes latest web technologies and allow accessing and visualizing large scale data on the web. The simulation system is a web-based 3D interactive learning environment for teaching hydrological and atmospheric processes and concepts. The simulation systems provides a visually striking platform with realistic terrain and weather information, and water simulation. The web-based simulation system provides an environment for students to learn about the earth science processes, and effects of development and human activity on the terrain. Users can access the system in three visualization modes including virtual reality, augmented reality, and immersive reality using heads-up display. The system provides various scenarios customized to fit the age and education level of various users.

  2. An interactive web-based system using cloud for large-scale visual analytics

    NASA Astrophysics Data System (ADS)

    Kaseb, Ahmed S.; Berry, Everett; Rozolis, Erik; McNulty, Kyle; Bontrager, Seth; Koh, Youngsol; Lu, Yung-Hsiang; Delp, Edward J.

    2015-03-01

    Network cameras have been growing rapidly in recent years. Thousands of public network cameras provide tremendous amount of visual information about the environment. There is a need to analyze this valuable information for a better understanding of the world around us. This paper presents an interactive web-based system that enables users to execute image analysis and computer vision techniques on a large scale to analyze the data from more than 65,000 worldwide cameras. This paper focuses on how to use both the system's website and Application Programming Interface (API). Given a computer program that analyzes a single frame, the user needs to make only slight changes to the existing program and choose the cameras to analyze. The system handles the heterogeneity of the geographically distributed cameras, e.g. different brands, resolutions. The system allocates and manages Amazon EC2 and Windows Azure cloud resources to meet the analysis requirements.

  3. Cloud-based solution to identify statistically significant MS peaks differentiating sample categories.

    PubMed

    Ji, Jun; Ling, Jeffrey; Jiang, Helen; Wen, Qiaojun; Whitin, John C; Tian, Lu; Cohen, Harvey J; Ling, Xuefeng B

    2013-03-23

    Mass spectrometry (MS) has evolved to become the primary high throughput tool for proteomics based biomarker discovery. Until now, multiple challenges in protein MS data analysis remain: large-scale and complex data set management; MS peak identification, indexing; and high dimensional peak differential analysis with the concurrent statistical tests based false discovery rate (FDR). "Turnkey" solutions are needed for biomarker investigations to rapidly process MS data sets to identify statistically significant peaks for subsequent validation. Here we present an efficient and effective solution, which provides experimental biologists easy access to "cloud" computing capabilities to analyze MS data. The web portal can be accessed at http://transmed.stanford.edu/ssa/. Presented web application supplies large scale MS data online uploading and analysis with a simple user interface. This bioinformatic tool will facilitate the discovery of the potential protein biomarkers using MS.

  4. Graph-Based Semantic Web Service Composition for Healthcare Data Integration.

    PubMed

    Arch-Int, Ngamnij; Arch-Int, Somjit; Sonsilphong, Suphachoke; Wanchai, Paweena

    2017-01-01

    Within the numerous and heterogeneous web services offered through different sources, automatic web services composition is the most convenient method for building complex business processes that permit invocation of multiple existing atomic services. The current solutions in functional web services composition lack autonomous queries of semantic matches within the parameters of web services, which are necessary in the composition of large-scale related services. In this paper, we propose a graph-based Semantic Web Services composition system consisting of two subsystems: management time and run time. The management-time subsystem is responsible for dependency graph preparation in which a dependency graph of related services is generated automatically according to the proposed semantic matchmaking rules. The run-time subsystem is responsible for discovering the potential web services and nonredundant web services composition of a user's query using a graph-based searching algorithm. The proposed approach was applied to healthcare data integration in different health organizations and was evaluated according to two aspects: execution time measurement and correctness measurement.

  5. Graph-Based Semantic Web Service Composition for Healthcare Data Integration

    PubMed Central

    2017-01-01

    Within the numerous and heterogeneous web services offered through different sources, automatic web services composition is the most convenient method for building complex business processes that permit invocation of multiple existing atomic services. The current solutions in functional web services composition lack autonomous queries of semantic matches within the parameters of web services, which are necessary in the composition of large-scale related services. In this paper, we propose a graph-based Semantic Web Services composition system consisting of two subsystems: management time and run time. The management-time subsystem is responsible for dependency graph preparation in which a dependency graph of related services is generated automatically according to the proposed semantic matchmaking rules. The run-time subsystem is responsible for discovering the potential web services and nonredundant web services composition of a user's query using a graph-based searching algorithm. The proposed approach was applied to healthcare data integration in different health organizations and was evaluated according to two aspects: execution time measurement and correctness measurement. PMID:29065602

  6. The Ins and Outs of Evaluating Web-Scale Discovery Services

    ERIC Educational Resources Information Center

    Hoeppner, Athena

    2012-01-01

    Librarians are familiar with the single-line form, the consolidated index, which represents a very large portion of a library's print and online collection. Their end users are familiar with the idea of a single search across a comprehensive index that produces a large, relevancy-ranked results list. Even though most patrons would not recognize…

  7. Processing Shotgun Proteomics Data on the Amazon Cloud with the Trans-Proteomic Pipeline*

    PubMed Central

    Slagel, Joseph; Mendoza, Luis; Shteynberg, David; Deutsch, Eric W.; Moritz, Robert L.

    2015-01-01

    Cloud computing, where scalable, on-demand compute cycles and storage are available as a service, has the potential to accelerate mass spectrometry-based proteomics research by providing simple, expandable, and affordable large-scale computing to all laboratories regardless of location or information technology expertise. We present new cloud computing functionality for the Trans-Proteomic Pipeline, a free and open-source suite of tools for the processing and analysis of tandem mass spectrometry datasets. Enabled with Amazon Web Services cloud computing, the Trans-Proteomic Pipeline now accesses large scale computing resources, limited only by the available Amazon Web Services infrastructure, for all users. The Trans-Proteomic Pipeline runs in an environment fully hosted on Amazon Web Services, where all software and data reside on cloud resources to tackle large search studies. In addition, it can also be run on a local computer with computationally intensive tasks launched onto the Amazon Elastic Compute Cloud service to greatly decrease analysis times. We describe the new Trans-Proteomic Pipeline cloud service components, compare the relative performance and costs of various Elastic Compute Cloud service instance types, and present on-line tutorials that enable users to learn how to deploy cloud computing technology rapidly with the Trans-Proteomic Pipeline. We provide tools for estimating the necessary computing resources and costs given the scale of a job and demonstrate the use of cloud enabled Trans-Proteomic Pipeline by performing over 1100 tandem mass spectrometry files through four proteomic search engines in 9 h and at a very low cost. PMID:25418363

  8. Processing shotgun proteomics data on the Amazon cloud with the trans-proteomic pipeline.

    PubMed

    Slagel, Joseph; Mendoza, Luis; Shteynberg, David; Deutsch, Eric W; Moritz, Robert L

    2015-02-01

    Cloud computing, where scalable, on-demand compute cycles and storage are available as a service, has the potential to accelerate mass spectrometry-based proteomics research by providing simple, expandable, and affordable large-scale computing to all laboratories regardless of location or information technology expertise. We present new cloud computing functionality for the Trans-Proteomic Pipeline, a free and open-source suite of tools for the processing and analysis of tandem mass spectrometry datasets. Enabled with Amazon Web Services cloud computing, the Trans-Proteomic Pipeline now accesses large scale computing resources, limited only by the available Amazon Web Services infrastructure, for all users. The Trans-Proteomic Pipeline runs in an environment fully hosted on Amazon Web Services, where all software and data reside on cloud resources to tackle large search studies. In addition, it can also be run on a local computer with computationally intensive tasks launched onto the Amazon Elastic Compute Cloud service to greatly decrease analysis times. We describe the new Trans-Proteomic Pipeline cloud service components, compare the relative performance and costs of various Elastic Compute Cloud service instance types, and present on-line tutorials that enable users to learn how to deploy cloud computing technology rapidly with the Trans-Proteomic Pipeline. We provide tools for estimating the necessary computing resources and costs given the scale of a job and demonstrate the use of cloud enabled Trans-Proteomic Pipeline by performing over 1100 tandem mass spectrometry files through four proteomic search engines in 9 h and at a very low cost. © 2015 by The American Society for Biochemistry and Molecular Biology, Inc.

  9. Imprints of the large-scale structure on AGN formation and evolution

    NASA Astrophysics Data System (ADS)

    Porqueres, Natàlia; Jasche, Jens; Enßlin, Torsten A.; Lavaux, Guilhem

    2018-04-01

    Black hole masses are found to correlate with several global properties of their host galaxies, suggesting that black holes and galaxies have an intertwined evolution and that active galactic nuclei (AGN) have a significant impact on galaxy evolution. Since the large-scale environment can also affect AGN, this work studies how their formation and properties depend on the environment. We have used a reconstructed three-dimensional high-resolution density field obtained from a Bayesian large-scale structure reconstruction method applied to the 2M++ galaxy sample. A web-type classification relying on the shear tensor is used to identify different structures on the cosmic web, defining voids, sheets, filaments, and clusters. We confirm that the environmental density affects the AGN formation and their properties. We found that the AGN abundance is equivalent to the galaxy abundance, indicating that active and inactive galaxies reside in similar dark matter halos. However, occurrence rates are different for each spectral type and accretion rate. These differences are consistent with the AGN evolutionary sequence suggested by previous authors, Seyferts and Transition objects transforming into low-ionization nuclear emission line regions (LINERs), the weaker counterpart of Seyferts. We conclude that AGN properties depend on the environmental density more than on the web-type. More powerful starbursts and younger stellar populations are found in high densities, where interactions and mergers are more likely. AGN hosts show smaller masses in clusters for Seyferts and Transition objects, which might be due to gas stripping. In voids, the AGN population is dominated by the most massive galaxy hosts.

  10. The large-scale environment from cosmological simulations - I. The baryonic cosmic web

    NASA Astrophysics Data System (ADS)

    Cui, Weiguang; Knebe, Alexander; Yepes, Gustavo; Yang, Xiaohu; Borgani, Stefano; Kang, Xi; Power, Chris; Staveley-Smith, Lister

    2018-01-01

    Using a series of cosmological simulations that includes one dark-matter-only (DM-only) run, one gas cooling-star formation-supernova feedback (CSF) run and one that additionally includes feedback from active galactic nuclei (AGNs), we classify the large-scale structures with both a velocity-shear-tensor code (VWEB) and a tidal-tensor code (PWEB). We find that the baryonic processes have almost no impact on large-scale structures - at least not when classified using aforementioned techniques. More importantly, our results confirm that the gas component alone can be used to infer the filamentary structure of the universe practically un-biased, which could be applied to cosmology constraints. In addition, the gas filaments are classified with its velocity (VWEB) and density (PWEB) fields, which can theoretically connect to the radio observations, such as H I surveys. This will help us to bias-freely link the radio observations with dark matter distributions at large scale.

  11. Genome Partitioner: A web tool for multi-level partitioning of large-scale DNA constructs for synthetic biology applications.

    PubMed

    Christen, Matthias; Del Medico, Luca; Christen, Heinz; Christen, Beat

    2017-01-01

    Recent advances in lower-cost DNA synthesis techniques have enabled new innovations in the field of synthetic biology. Still, efficient design and higher-order assembly of genome-scale DNA constructs remains a labor-intensive process. Given the complexity, computer assisted design tools that fragment large DNA sequences into fabricable DNA blocks are needed to pave the way towards streamlined assembly of biological systems. Here, we present the Genome Partitioner software implemented as a web-based interface that permits multi-level partitioning of genome-scale DNA designs. Without the need for specialized computing skills, biologists can submit their DNA designs to a fully automated pipeline that generates the optimal retrosynthetic route for higher-order DNA assembly. To test the algorithm, we partitioned a 783 kb Caulobacter crescentus genome design. We validated the partitioning strategy by assembling a 20 kb test segment encompassing a difficult to synthesize DNA sequence. Successful assembly from 1 kb subblocks into the 20 kb segment highlights the effectiveness of the Genome Partitioner for reducing synthesis costs and timelines for higher-order DNA assembly. The Genome Partitioner is broadly applicable to translate DNA designs into ready to order sequences that can be assembled with standardized protocols, thus offering new opportunities to harness the diversity of microbial genomes for synthetic biology applications. The Genome Partitioner web tool can be accessed at https://christenlab.ethz.ch/GenomePartitioner.

  12. A Follow-Up Web-Based Survey: Test and Measurement Expert Opinions on the Psychometric Properties of Out-of-Level Tests. Out-of-Level Testing Report.

    ERIC Educational Resources Information Center

    Bielinski, John; Minnema, Jane; Thurlow, Martha

    A Web-based survey of 25 experts in testing theory and large-scale assessment examined the utility of out-of-level testing for making decisions about students and schools. Survey respondents were given a series of scenarios and asked to judge the degree to which out-of-level testing would affect the reliability and validity of test scores within…

  13. Twitter web-service for soft agent reporting in persistent surveillance systems

    NASA Astrophysics Data System (ADS)

    Rababaah, Haroun; Shirkhodaie, Amir

    2010-04-01

    Persistent surveillance is an intricate process requiring monitoring, gathering, processing, tracking, and characterization of many spatiotemporal events occurring concurrently. Data associated with events can be readily attained by networking of hard (physical) sensors. Sensors may have homogeneous or heterogeneous (hybrid) sensing modalities with different communication bandwidth requirements. Complimentary to hard sensors are human observers or "soft sensors" that can report occurrences of evolving events via different communication devices (e.g., texting, cell phones, emails, instant messaging, etc.) to the command control center. However, networking of human observers in ad-hoc way is rather a difficult task. In this paper, we present a Twitter web-service for soft agent reporting in persistent surveillance systems (called Web-STARS). The objective of this web-service is to aggregate multi-source human observations in hybrid sensor networks rapidly. With availability of Twitter social network, such a human networking concept can not only be realized for large scale persistent surveillance systems (PSS), but also, it can be employed with proper interfaces to expedite rapid events reporting by human observers. The proposed technique is particularly suitable for large-scale persistent surveillance systems with distributed soft and hard sensor networks. The efficiency and effectiveness of the proposed technique is measured experimentally by conducting several simulated persistent surveillance scenarios. It is demonstrated that by fusion of information from hard and soft agents improves understanding of common operating picture and enhances situational awareness.

  14. Characterizing stroke lesions using digital templates and lesion quantification tools in a web-based imaging informatics system for a large-scale stroke rehabilitation clinical trial

    NASA Astrophysics Data System (ADS)

    Wang, Ximing; Edwardson, Matthew; Dromerick, Alexander; Winstein, Carolee; Wang, Jing; Liu, Brent

    2015-03-01

    Previously, we presented an Interdisciplinary Comprehensive Arm Rehabilitation Evaluation (ICARE) imaging informatics system that supports a large-scale phase III stroke rehabilitation trial. The ePR system is capable of displaying anonymized patient imaging studies and reports, and the system is accessible to multiple clinical trial sites and users across the United States via the web. However, the prior multicenter stroke rehabilitation trials lack any significant neuroimaging analysis infrastructure. In stroke related clinical trials, identification of the stroke lesion characteristics can be meaningful as recent research shows that lesion characteristics are related to stroke scale and functional recovery after stroke. To facilitate the stroke clinical trials, we hope to gain insight into specific lesion characteristics, such as vascular territory, for patients enrolled into large stroke rehabilitation trials. To enhance the system's capability for data analysis and data reporting, we have integrated new features with the system: a digital brain template display, a lesion quantification tool and a digital case report form. The digital brain templates are compiled from published vascular territory templates at each of 5 angles of incidence. These templates were updated to include territories in the brainstem using a vascular territory atlas and the Medical Image Processing, Analysis and Visualization (MIPAV) tool. The digital templates are displayed for side-by-side comparisons and transparent template overlay onto patients' images in the image viewer. The lesion quantification tool quantifies planimetric lesion area from user-defined contour. The digital case report form stores user input into a database, then displays contents in the interface to allow for reviewing, editing, and new inputs. In sum, the newly integrated system features provide the user with readily-accessible web-based tools to identify the vascular territory involved, estimate lesion area, and store these results in a web-based digital format.

  15. WImpiBLAST: web interface for mpiBLAST to help biologists perform large-scale annotation using high performance computing.

    PubMed

    Sharma, Parichit; Mantri, Shrikant S

    2014-01-01

    The function of a newly sequenced gene can be discovered by determining its sequence homology with known proteins. BLAST is the most extensively used sequence analysis program for sequence similarity search in large databases of sequences. With the advent of next generation sequencing technologies it has now become possible to study genes and their expression at a genome-wide scale through RNA-seq and metagenome sequencing experiments. Functional annotation of all the genes is done by sequence similarity search against multiple protein databases. This annotation task is computationally very intensive and can take days to obtain complete results. The program mpiBLAST, an open-source parallelization of BLAST that achieves superlinear speedup, can be used to accelerate large-scale annotation by using supercomputers and high performance computing (HPC) clusters. Although many parallel bioinformatics applications using the Message Passing Interface (MPI) are available in the public domain, researchers are reluctant to use them due to lack of expertise in the Linux command line and relevant programming experience. With these limitations, it becomes difficult for biologists to use mpiBLAST for accelerating annotation. No web interface is available in the open-source domain for mpiBLAST. We have developed WImpiBLAST, a user-friendly open-source web interface for parallel BLAST searches. It is implemented in Struts 1.3 using a Java backbone and runs atop the open-source Apache Tomcat Server. WImpiBLAST supports script creation and job submission features and also provides a robust job management interface for system administrators. It combines script creation and modification features with job monitoring and management through the Torque resource manager on a Linux-based HPC cluster. Use case information highlights the acceleration of annotation analysis achieved by using WImpiBLAST. Here, we describe the WImpiBLAST web interface features and architecture, explain design decisions, describe workflows and provide a detailed analysis.

  16. WImpiBLAST: Web Interface for mpiBLAST to Help Biologists Perform Large-Scale Annotation Using High Performance Computing

    PubMed Central

    Sharma, Parichit; Mantri, Shrikant S.

    2014-01-01

    The function of a newly sequenced gene can be discovered by determining its sequence homology with known proteins. BLAST is the most extensively used sequence analysis program for sequence similarity search in large databases of sequences. With the advent of next generation sequencing technologies it has now become possible to study genes and their expression at a genome-wide scale through RNA-seq and metagenome sequencing experiments. Functional annotation of all the genes is done by sequence similarity search against multiple protein databases. This annotation task is computationally very intensive and can take days to obtain complete results. The program mpiBLAST, an open-source parallelization of BLAST that achieves superlinear speedup, can be used to accelerate large-scale annotation by using supercomputers and high performance computing (HPC) clusters. Although many parallel bioinformatics applications using the Message Passing Interface (MPI) are available in the public domain, researchers are reluctant to use them due to lack of expertise in the Linux command line and relevant programming experience. With these limitations, it becomes difficult for biologists to use mpiBLAST for accelerating annotation. No web interface is available in the open-source domain for mpiBLAST. We have developed WImpiBLAST, a user-friendly open-source web interface for parallel BLAST searches. It is implemented in Struts 1.3 using a Java backbone and runs atop the open-source Apache Tomcat Server. WImpiBLAST supports script creation and job submission features and also provides a robust job management interface for system administrators. It combines script creation and modification features with job monitoring and management through the Torque resource manager on a Linux-based HPC cluster. Use case information highlights the acceleration of annotation analysis achieved by using WImpiBLAST. Here, we describe the WImpiBLAST web interface features and architecture, explain design decisions, describe workflows and provide a detailed analysis. PMID:24979410

  17. Empirical analysis of web-based user-object bipartite networks

    NASA Astrophysics Data System (ADS)

    Shang, Ming-Sheng; Lü, Linyuan; Zhang, Yi-Cheng; Zhou, Tao

    2010-05-01

    Understanding the structure and evolution of web-based user-object networks is a significant task since they play a crucial role in e-commerce nowadays. This letter reports the empirical analysis on two large-scale web sites, audioscrobbler.com and del.icio.us, where users are connected with music groups and bookmarks, respectively. The degree distributions and degree-degree correlations for both users and objects are reported. We propose a new index, named collaborative similarity, to quantify the diversity of tastes based on the collaborative selection. Accordingly, the correlation between degree and selection diversity is investigated. We report some novel phenomena well characterizing the selection mechanism of web users and outline the relevance of these phenomena to the information recommendation problem.

  18. Increasing Scalability of Researcher Network Extraction from the Web

    NASA Astrophysics Data System (ADS)

    Asada, Yohei; Matsuo, Yutaka; Ishizuka, Mitsuru

    Social networks, which describe relations among people or organizations as a network, have recently attracted attention. With the help of a social network, we can analyze the structure of a community and thereby promote efficient communications within it. We investigate the problem of extracting a network of researchers from the Web, to assist efficient cooperation among researchers. Our method uses a search engine to get the cooccurences of names of two researchers and calculates the streangth of the relation between them. Then we label the relation by analyzing the Web pages in which these two names cooccur. Research on social network extraction using search engines as ours, is attracting attention in Japan as well as abroad. However, the former approaches issue too many queries to search engines to extract a large-scale network. In this paper, we propose a method to filter superfluous queries and facilitates the extraction of large-scale networks. By this method we are able to extract a network of around 3000-nodes. Our experimental results show that the proposed method reduces the number of queries significantly while preserving the quality of the network as compared to former methods.

  19. Tropical Cyclone Information System

    NASA Technical Reports Server (NTRS)

    Li, P. Peggy; Knosp, Brian W.; Vu, Quoc A.; Yi, Chao; Hristova-Veleva, Svetla M.

    2009-01-01

    The JPL Tropical Cyclone Infor ma tion System (TCIS) is a Web portal (http://tropicalcyclone.jpl.nasa.gov) that provides researchers with an extensive set of observed hurricane parameters together with large-scale and convection resolving model outputs. It provides a comprehensive set of high-resolution satellite (see figure), airborne, and in-situ observations in both image and data formats. Large-scale datasets depict the surrounding environmental parameters such as SST (Sea Surface Temperature) and aerosol loading. Model outputs and analysis tools are provided to evaluate model performance and compare observations from different platforms. The system pertains to the thermodynamic and microphysical structure of the storm, the air-sea interaction processes, and the larger-scale environment as depicted by ocean heat content and the aerosol loading of the environment. Currently, the TCIS is populated with satellite observations of all tropical cyclones observed globally during 2005. There is a plan to extend the database both forward in time till present as well as backward to 1998. The portal is powered by a MySQL database and an Apache/Tomcat Web server on a Linux system. The interactive graphic user interface is provided by Google Map.

  20. The cosmic web and microwave background fossilize the first turbulent combustion

    NASA Astrophysics Data System (ADS)

    Gibson, Carl H.; Keeler, R. Norris

    2016-10-01

    Collisional fluid mechanics theory predicts a turbulent hot big bang at Planck conditions from large, negative, turbulence stresses below the Fortov-Kerr limit (< -10113 Pa). Big bang turbulence fossilized when quarks formed, extracting the mass energy of the universe by extreme negative viscous stresses of inflation, expanding to length scales larger than the horizon scale ct. Viscous-gravitational structure formation by fragmentation was triggered at big bang fossil vorticity turbulence vortex lines during the plasma epoch, as observed by the Planck space telescope. A cosmic web of protogalaxies, protogalaxyclusters, and protogalaxysuperclusters that formed in turbulent boundary layers of the spinning voids are hereby identified as expanding turbulence fossils that falsify CDMHC cosmology.

  1. Wilber 3: A Python-Django Web Application For Acquiring Large-scale Event-oriented Seismic Data

    NASA Astrophysics Data System (ADS)

    Newman, R. L.; Clark, A.; Trabant, C. M.; Karstens, R.; Hutko, A. R.; Casey, R. E.; Ahern, T. K.

    2013-12-01

    Since 2001, the IRIS Data Management Center (DMC) WILBER II system has provided a convenient web-based interface for locating seismic data related to a particular event, and requesting a subset of that data for download. Since its launch, both the scale of available data and the technology of web-based applications have developed significantly. Wilber 3 is a ground-up redesign that leverages a number of public and open-source projects to provide an event-oriented data request interface with a high level of interactivity and scalability for multiple data types. Wilber 3 uses the IRIS/Federation of Digital Seismic Networks (FDSN) web services for event data, metadata, and time-series data. Combining a carefully optimized Google Map with the highly scalable SlickGrid data API, the Wilber 3 client-side interface can load tens of thousands of events or networks/stations in a single request, and provide instantly responsive browsing, sorting, and filtering of event and meta data in the web browser, without further reliance on the data service. The server-side of Wilber 3 is a Python-Django application, one of over a dozen developed in the last year at IRIS, whose common framework, components, and administrative overhead represent a massive savings in developer resources. Requests for assembled datasets, which may include thousands of data channels and gigabytes of data, are queued and executed using the Celery distributed Python task scheduler, giving Wilber 3 the ability to operate in parallel across a large number of nodes.

  2. Food Web Structure in a Harsh Glacier-Fed River

    PubMed Central

    Clitherow, Leonie R.; Carrivick, Jonathan L.; Brown, Lee E.

    2013-01-01

    Glacier retreat is occurring across the world, and associated river ecosystems are expected to respond more rapidly than those in flowing waters in other regions. The river environment directly downstream of a glacier snout is characterised by extreme low water temperature and unstable channel sediments but these habitats may become rarer with widespread glacier retreat. In these extreme environments food web dynamics have been little studied, yet they could offer opportunities to test food web theories using highly resolved food webs owing to their low taxonomic richness. This study examined the interactions of macroinvertebrate and diatom taxa in the Ödenwinkelkees river, Austrian central Alps between 2006 and 2011. The webs were characterised by low taxon richness (13–22), highly connected individuals (directed connectance up to 0.19) and short mean food chain length (2.00–2.36). The dominant macroinvertebrates were members of the Chironomidae genus Diamesa and had an omnivorous diet rich in detritus and diatoms as well as other Chironomidae. Simuliidae (typically detritivorous filterers) had a diet rich in diatoms but also showed evidence of predation on Chironomidae larvae. Food webs showed strong species-averaged and individual size structuring but mass-abundance scaling coefficients were larger than those predicted by metabolic theory, perhaps due to a combination of spatial averaging effects of patchily distributed consumers and resources, and/or consumers deriving unquantified resources from microorganisms attached to the large amounts of ingested rock fragments. Comparison of food web structural metrics with those from 62 published river webs suggest these glacier-fed river food web properties were extreme but in line with general food web scaling predictions, a finding which could prove useful to forecast the effects of anticipated future glacier retreat on the structure of aquatic food webs. PMID:23613751

  3. The Comprehensive Microbial Resource.

    PubMed

    Peterson, J D; Umayam, L A; Dickinson, T; Hickey, E K; White, O

    2001-01-01

    One challenge presented by large-scale genome sequencing efforts is effective display of uniform information to the scientific community. The Comprehensive Microbial Resource (CMR) contains robust annotation of all complete microbial genomes and allows for a wide variety of data retrievals. The bacterial information has been placed on the Web at http://www.tigr.org/CMR for retrieval using standard web browsing technology. Retrievals can be based on protein properties such as molecular weight or hydrophobicity, GC-content, functional role assignments and taxonomy. The CMR also has special web-based tools to allow data mining using pre-run homology searches, whole genome dot-plots, batch downloading and traversal across genomes using a variety of datatypes.

  4. ePlant and the 3D data display initiative: integrative systems biology on the world wide web.

    PubMed

    Fucile, Geoffrey; Di Biase, David; Nahal, Hardeep; La, Garon; Khodabandeh, Shokoufeh; Chen, Yani; Easley, Kante; Christendat, Dinesh; Kelley, Lawrence; Provart, Nicholas J

    2011-01-10

    Visualization tools for biological data are often limited in their ability to interactively integrate data at multiple scales. These computational tools are also typically limited by two-dimensional displays and programmatic implementations that require separate configurations for each of the user's computing devices and recompilation for functional expansion. Towards overcoming these limitations we have developed "ePlant" (http://bar.utoronto.ca/eplant) - a suite of open-source world wide web-based tools for the visualization of large-scale data sets from the model organism Arabidopsis thaliana. These tools display data spanning multiple biological scales on interactive three-dimensional models. Currently, ePlant consists of the following modules: a sequence conservation explorer that includes homology relationships and single nucleotide polymorphism data, a protein structure model explorer, a molecular interaction network explorer, a gene product subcellular localization explorer, and a gene expression pattern explorer. The ePlant's protein structure explorer module represents experimentally determined and theoretical structures covering >70% of the Arabidopsis proteome. The ePlant framework is accessed entirely through a web browser, and is therefore platform-independent. It can be applied to any model organism. To facilitate the development of three-dimensional displays of biological data on the world wide web we have established the "3D Data Display Initiative" (http://3ddi.org).

  5. Globus | Informatics Technology for Cancer Research (ITCR)

    Cancer.gov

    Globus software services provide secure cancer research data transfer, synchronization, and sharing in distributed environments at large scale. These services can be integrated into applications and research data gateways, leveraging Globus identity management, single sign-on, search, and authorization capabilities. Globus Genomics integrates Globus with the Galaxy genomics workflow engine and Amazon Web Services to enable cancer genomics analysis that can elastically scale compute resources with demand.

  6. Comparative Initial and Sustained Engagement in Web-based Training by Behavioral Healthcare Providers in New York State.

    PubMed

    Talley, Rachel; Chiang, I-Chin; Covell, Nancy H; Dixon, Lisa

    2018-06-01

    Improved dissemination is critical to implementation of evidence-based practice in community behavioral healthcare settings. Web-based training modalities are a promising strategy for dissemination of evidence-based practice in community behavioral health settings. Initial and sustained engagement of these modalities in large, multidisciplinary community provider samples is not well understood. This study evaluates comparative engagement and user preferences by provider type in a web-based training platform in a large, multidisciplinary community sample of behavioral health staff in New York State. Workforce make-up among platform registrants was compared to the general NYS behavioral health workforce. Training completion by functional job type was compared to characterize user engagement and preferences. Frequently completed modules were classified by credit and requirement incentives. High initial training engagement across professional role was demonstrated, with significant differences in initial and sustained engagement by professional role. The most frequently completed modules across functional job types contained credit or requirement incentives. The analysis demonstrated that high engagement of a web-based training in a multidisciplinary provider audience can be achieved without tailoring content to specific professional roles. Overlap between frequently completed modules and incentives suggests a role for incentives in promoting engagement of web-based training. These findings further the understanding of strategies to promote large-scale dissemination of evidence-based practice in community behavioral health settings.

  7. EuroPhenome and EMPReSS: online mouse phenotyping resource

    PubMed Central

    Mallon, Ann-Marie; Hancock, John M.

    2008-01-01

    EuroPhenome (http://www.europhenome.org) and EMPReSS (http://empress.har.mrc.ac.uk/) form an integrated resource to provide access to data and procedures for mouse phenotyping. EMPReSS describes 96 Standard Operating Procedures for mouse phenotyping. EuroPhenome contains data resulting from carrying out EMPReSS protocols on four inbred laboratory mouse strains. As well as web interfaces, both resources support web services to enable integration with other mouse phenotyping and functional genetics resources, and are committed to initiatives to improve integration of mouse phenotype databases. EuroPhenome will be the repository for a recently initiated effort to carry out large-scale phenotyping on a large number of knockout mouse lines (EUMODIC). PMID:17905814

  8. EuroPhenome and EMPReSS: online mouse phenotyping resource.

    PubMed

    Mallon, Ann-Marie; Blake, Andrew; Hancock, John M

    2008-01-01

    EuroPhenome (http://www.europhenome.org) and EMPReSS (http://empress.har.mrc.ac.uk/) form an integrated resource to provide access to data and procedures for mouse phenotyping. EMPReSS describes 96 Standard Operating Procedures for mouse phenotyping. EuroPhenome contains data resulting from carrying out EMPReSS protocols on four inbred laboratory mouse strains. As well as web interfaces, both resources support web services to enable integration with other mouse phenotyping and functional genetics resources, and are committed to initiatives to improve integration of mouse phenotype databases. EuroPhenome will be the repository for a recently initiated effort to carry out large-scale phenotyping on a large number of knockout mouse lines (EUMODIC).

  9. Satisfaction with web-based training in an integrated healthcare delivery network: do age, education, computer skills and attitudes matter?

    PubMed Central

    Atreja, Ashish; Mehta, Neil B; Jain, Anil K; Harris, CM; Ishwaran, Hemant; Avital, Michel; Fishleder, Andrew J

    2008-01-01

    Background Healthcare institutions spend enormous time and effort to train their workforce. Web-based training can potentially streamline this process. However the deployment of web-based training in a large-scale setting with a diverse healthcare workforce has not been evaluated. The aim of this study was to evaluate the satisfaction of healthcare professionals with web-based training and to determine the predictors of such satisfaction including age, education status and computer proficiency. Methods Observational, cross-sectional survey of healthcare professionals from six hospital systems in an integrated delivery network. We measured overall satisfaction to web-based training and response to survey items measuring Website Usability, Course Usefulness, Instructional Design Effectiveness, Computer Proficiency and Self-learning Attitude. Results A total of 17,891 healthcare professionals completed the web-based training on HIPAA Privacy Rule; and of these, 13,537 completed the survey (response rate 75.6%). Overall course satisfaction was good (median, 4; scale, 1 to 5) with more than 75% of the respondents satisfied with the training (rating 4 or 5) and 65% preferring web-based training over traditional instructor-led training (rating 4 or 5). Multivariable ordinal regression revealed 3 key predictors of satisfaction with web-based training: Instructional Design Effectiveness, Website Usability and Course Usefulness. Demographic predictors such as gender, age and education did not have an effect on satisfaction. Conclusion The study shows that web-based training when tailored to learners' background, is perceived as a satisfactory mode of learning by an interdisciplinary group of healthcare professionals, irrespective of age, education level or prior computer experience. Future studies should aim to measure the long-term outcomes of web-based training. PMID:18922178

  10. Using Web-Based Knowledge Extraction Techniques to Support Cultural Modeling

    NASA Astrophysics Data System (ADS)

    Smart, Paul R.; Sieck, Winston R.; Shadbolt, Nigel R.

    The World Wide Web is a potentially valuable source of information about the cognitive characteristics of cultural groups. However, attempts to use the Web in the context of cultural modeling activities are hampered by the large-scale nature of the Web and the current dominance of natural language formats. In this paper, we outline an approach to support the exploitation of the Web for cultural modeling activities. The approach begins with the development of qualitative cultural models (which describe the beliefs, concepts and values of cultural groups), and these models are subsequently used to develop an ontology-based information extraction capability. Our approach represents an attempt to combine conventional approaches to information extraction with epidemiological perspectives of culture and network-based approaches to cultural analysis. The approach can be used, we suggest, to support the development of models providing a better understanding of the cognitive characteristics of particular cultural groups.

  11. Genome Partitioner: A web tool for multi-level partitioning of large-scale DNA constructs for synthetic biology applications

    PubMed Central

    Del Medico, Luca; Christen, Heinz; Christen, Beat

    2017-01-01

    Recent advances in lower-cost DNA synthesis techniques have enabled new innovations in the field of synthetic biology. Still, efficient design and higher-order assembly of genome-scale DNA constructs remains a labor-intensive process. Given the complexity, computer assisted design tools that fragment large DNA sequences into fabricable DNA blocks are needed to pave the way towards streamlined assembly of biological systems. Here, we present the Genome Partitioner software implemented as a web-based interface that permits multi-level partitioning of genome-scale DNA designs. Without the need for specialized computing skills, biologists can submit their DNA designs to a fully automated pipeline that generates the optimal retrosynthetic route for higher-order DNA assembly. To test the algorithm, we partitioned a 783 kb Caulobacter crescentus genome design. We validated the partitioning strategy by assembling a 20 kb test segment encompassing a difficult to synthesize DNA sequence. Successful assembly from 1 kb subblocks into the 20 kb segment highlights the effectiveness of the Genome Partitioner for reducing synthesis costs and timelines for higher-order DNA assembly. The Genome Partitioner is broadly applicable to translate DNA designs into ready to order sequences that can be assembled with standardized protocols, thus offering new opportunities to harness the diversity of microbial genomes for synthetic biology applications. The Genome Partitioner web tool can be accessed at https://christenlab.ethz.ch/GenomePartitioner. PMID:28531174

  12. To Make Archives Available Online: Transcending Boundaries or Building Walls?

    ERIC Educational Resources Information Center

    Hansen, Lars-Erik; Sundqvist, Anneli

    2012-01-01

    The development of information technology and the rise of the Internet have rendered a large-scale digitization and dissemination of originally analog information objects. On the Web sites "Lararnas Historia" ("History of Teachers" www.lararhistoria.se) and "Ingenjorshistoria" ("History of Engineers"…

  13. The f ( R ) halo mass function in the cosmic web

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Braun-Bates, F. von; Winther, H.A.; Alonso, D.

    An important indicator of modified gravity is the effect of the local environment on halo properties. This paper examines the influence of the local tidal structure on the halo mass function, the halo orientation, spin and the concentration-mass relation. We use the excursion set formalism to produce a halo mass function conditional on large-scale structure. Our simple model agrees well with simulations on large scales at which the density field is linear or weakly non-linear. Beyond this, our principal result is that f ( R ) does affect halo abundances, the halo spin parameter and the concentration-mass relationship in anmore » environment-independent way, whereas we find no appreciable deviation from \\text(ΛCDM) for the mass function with fixed environment density, nor the alignment of the orientation and spin vectors of the halo to the eigenvectors of the local cosmic web. There is a general trend for greater deviation from \\text(ΛCDM) in underdense environments and for high-mass haloes, as expected from chameleon screening.« less

  14. The cosmic spiderweb: equivalence of cosmic, architectural and origami tessellations

    PubMed Central

    Hidding, Johan; Konstantatou, Marina; van de Weygaert, Rien

    2018-01-01

    For over 20 years, the term ‘cosmic web’ has guided our understanding of the large-scale arrangement of matter in the cosmos, accurately evoking the concept of a network of galaxies linked by filaments. But the physical correspondence between the cosmic web and structural engineering or textile ‘spiderwebs’ is even deeper than previously known, and also extends to origami tessellations. Here, we explain that in a good structure-formation approximation known as the adhesion model, threads of the cosmic web form a spiderweb, i.e. can be strung up to be entirely in tension. The correspondence is exact if nodes sampling voids are included, and if structure is excluded within collapsed regions (walls, filaments and haloes), where dark-matter multistreaming and baryonic physics affect the structure. We also suggest how concepts arising from this link might be used to test cosmological models: for example, to test for large-scale anisotropy and rotational flows in the cosmos. PMID:29765637

  15. Evaluation of a metal shear web selectively reinforced with filamentary composites for space shuttle application. Phase 3 Summary report: Shear web component testing and analysis

    NASA Technical Reports Server (NTRS)

    Laakso, J. H.; Straayer, J. W.

    1973-01-01

    Three large scale advanced composite shear web components were tested and analyzed to evaluate application of the design concept to a space shuttle orbiter thrust structure. The shear web design concept consisted of a titanium-clad + or - 45 deg boron/epoxy web laminate stiffened with vertical boron/epoxy reinforced aluminum stiffeners. The design concept was evaluated to be efficient and practical for the application that was studied. Because of the effects of buckling deflections, a requirement is identified for shear buckling resistant design to maximize the efficiency of highly-loaded advanced composite shear webs. An approximate analysis of prebuckling deflections is presented and computer-aided design results, which consider prebuckling deformations, indicate that the design concept offers a theoretical weight saving of 31 percent relative to all metal construction. Recommendations are made for design concept options and analytical methods that are appropriate for production hardware.

  16. The Comprehensive Microbial Resource

    PubMed Central

    Peterson, Jeremy D.; Umayam, Lowell A.; Dickinson, Tanja; Hickey, Erin K.; White, Owen

    2001-01-01

    One challenge presented by large-scale genome sequencing efforts is effective display of uniform information to the scientific community. The Comprehensive Microbial Resource (CMR) contains robust annotation of all complete microbial genomes and allows for a wide variety of data retrievals. The bacterial information has been placed on the Web at http://www.tigr.org/CMR for retrieval using standard web browsing technology. Retrievals can be based on protein properties such as molecular weight or hydrophobicity, GC-content, functional role assignments and taxonomy. The CMR also has special web-based tools to allow data mining using pre-run homology searches, whole genome dot-plots, batch downloading and traversal across genomes using a variety of datatypes. PMID:11125067

  17. Providing Multi-Page Data Extraction Services with XWRAPComposer

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Liu, Ling; Zhang, Jianjun; Han, Wei

    2008-04-30

    Dynamic Web data sources – sometimes known collectively as the Deep Web – increase the utility of the Web by providing intuitive access to data repositories anywhere that Web access is available. Deep Web services provide access to real-time information, like entertainment event listings, or present a Web interface to large databases or other data repositories. Recent studies suggest that the size and growth rate of the dynamic Web greatly exceed that of the static Web, yet dynamic content is often ignored by existing search engine indexers owing to the technical challenges that arise when attempting to search the Deepmore » Web. To address these challenges, we present DYNABOT, a service-centric crawler for discovering and clustering Deep Web sources offering dynamic content. DYNABOT has three unique characteristics. First, DYNABOT utilizes a service class model of the Web implemented through the construction of service class descriptions (SCDs). Second, DYNABOT employs a modular, self-tuning system architecture for focused crawling of the Deep Web using service class descriptions. Third, DYNABOT incorporates methods and algorithms for efficient probing of the Deep Web and for discovering and clustering Deep Web sources and services through SCD-based service matching analysis. Our experimental results demonstrate the effectiveness of the service class discovery, probing, and matching algorithms and suggest techniques for efficiently managing service discovery in the face of the immense scale of the Deep Web.« less

  18. Web-based visualization of very large scientific astronomy imagery

    NASA Astrophysics Data System (ADS)

    Bertin, E.; Pillay, R.; Marmo, C.

    2015-04-01

    Visualizing and navigating through large astronomy images from a remote location with current astronomy display tools can be a frustrating experience in terms of speed and ergonomics, especially on mobile devices. In this paper, we present a high performance, versatile and robust client-server system for remote visualization and analysis of extremely large scientific images. Applications of this work include survey image quality control, interactive data query and exploration, citizen science, as well as public outreach. The proposed software is entirely open source and is designed to be generic and applicable to a variety of datasets. It provides access to floating point data at terabyte scales, with the ability to precisely adjust image settings in real-time. The proposed clients are light-weight, platform-independent web applications built on standard HTML5 web technologies and compatible with both touch and mouse-based devices. We put the system to the test and assess the performance of the system and show that a single server can comfortably handle more than a hundred simultaneous users accessing full precision 32 bit astronomy data.

  19. BRepertoire: a user-friendly web server for analysing antibody repertoire data.

    PubMed

    Margreitter, Christian; Lu, Hui-Chun; Townsend, Catherine; Stewart, Alexander; Dunn-Walters, Deborah K; Fraternali, Franca

    2018-04-14

    Antibody repertoire analysis by high throughput sequencing is now widely used, but a persisting challenge is enabling immunologists to explore their data to discover discriminating repertoire features for their own particular investigations. Computational methods are necessary for large-scale evaluation of antibody properties. We have developed BRepertoire, a suite of user-friendly web-based software tools for large-scale statistical analyses of repertoire data. The software is able to use data preprocessed by IMGT, and performs statistical and comparative analyses with versatile plotting options. BRepertoire has been designed to operate in various modes, for example analysing sequence-specific V(D)J gene usage, discerning physico-chemical properties of the CDR regions and clustering of clonotypes. Those analyses are performed on the fly by a number of R packages and are deployed by a shiny web platform. The user can download the analysed data in different table formats and save the generated plots as image files ready for publication. We believe BRepertoire to be a versatile analytical tool that complements experimental studies of immune repertoires. To illustrate the server's functionality, we show use cases including differential gene usage in a vaccination dataset and analysis of CDR3H properties in old and young individuals. The server is accessible under http://mabra.biomed.kcl.ac.uk/BRepertoire.

  20. Data integration in the era of omics: current and future challenges

    PubMed Central

    2014-01-01

    To integrate heterogeneous and large omics data constitutes not only a conceptual challenge but a practical hurdle in the daily analysis of omics data. With the rise of novel omics technologies and through large-scale consortia projects, biological systems are being further investigated at an unprecedented scale generating heterogeneous and often large data sets. These data-sets encourage researchers to develop novel data integration methodologies. In this introduction we review the definition and characterize current efforts on data integration in the life sciences. We have used a web-survey to assess current research projects on data-integration to tap into the views, needs and challenges as currently perceived by parts of the research community. PMID:25032990

  1. Leveraging Web-Based Environments for Mass Atrocity Prevention

    ERIC Educational Resources Information Center

    Harding, Tucker B.; Whitlock, Mark A.

    2013-01-01

    A growing literature exploring large-scale, identity-based political violence, including mass killing and genocide, debates the plausibility of, and prospects for, early warning and prevention. An extension of the debate involves the prospects for creating educational experiences that result in more sophisticated analytical products that enhance…

  2. Web based visualization of large climate data sets

    USGS Publications Warehouse

    Alder, Jay R.; Hostetler, Steven W.

    2015-01-01

    We have implemented the USGS National Climate Change Viewer (NCCV), which is an easy-to-use web application that displays future projections from global climate models over the United States at the state, county and watershed scales. We incorporate the NASA NEX-DCP30 statistically downscaled temperature and precipitation for 30 global climate models being used in the Fifth Assessment Report (AR5) of the Intergovernmental Panel on Climate Change (IPCC), and hydrologic variables we simulated using a simple water-balance model. Our application summarizes very large, complex data sets at scales relevant to resource managers and citizens and makes climate-change projection information accessible to users of varying skill levels. Tens of terabytes of high-resolution climate and water-balance data are distilled to compact binary format summary files that are used in the application. To alleviate slow response times under high loads, we developed a map caching technique that reduces the time it takes to generate maps by several orders of magnitude. The reduced access time scales to >500 concurrent users. We provide code examples that demonstrate key aspects of data processing, data exporting/importing and the caching technique used in the NCCV.

  3. An Extraction Method of an Informative DOM Node from a Web Page by Using Layout Information

    NASA Astrophysics Data System (ADS)

    Tsuruta, Masanobu; Masuyama, Shigeru

    We propose an informative DOM node extraction method from a Web page for preprocessing of Web content mining. Our proposed method LM uses layout data of DOM nodes generated by a generic Web browser, and the learning set consists of hundreds of Web pages and the annotations of informative DOM nodes of those Web pages. Our method does not require large scale crawling of the whole Web site to which the target Web page belongs. We design LM so that it uses the information of the learning set more efficiently in comparison to the existing method that uses the same learning set. By experiments, we evaluate the methods obtained by combining one that consists of the method for extracting the informative DOM node both the proposed method and the existing methods, and the existing noise elimination methods: Heur removes advertisements and link-lists by some heuristics and CE removes the DOM nodes existing in the Web pages in the same Web site to which the target Web page belongs. Experimental results show that 1) LM outperforms other methods for extracting the informative DOM node, 2) the combination method (LM, {CE(10), Heur}) based on LM (precision: 0.755, recall: 0.826, F-measure: 0.746) outperforms other combination methods.

  4. MetaRanker 2.0: a web server for prioritization of genetic variation data

    PubMed Central

    Pers, Tune H.; Dworzyński, Piotr; Thomas, Cecilia Engel; Lage, Kasper; Brunak, Søren

    2013-01-01

    MetaRanker 2.0 is a web server for prioritization of common and rare frequency genetic variation data. Based on heterogeneous data sets including genetic association data, protein–protein interactions, large-scale text-mining data, copy number variation data and gene expression experiments, MetaRanker 2.0 prioritizes the protein-coding part of the human genome to shortlist candidate genes for targeted follow-up studies. MetaRanker 2.0 is made freely available at www.cbs.dtu.dk/services/MetaRanker-2.0. PMID:23703204

  5. MetaRanker 2.0: a web server for prioritization of genetic variation data.

    PubMed

    Pers, Tune H; Dworzyński, Piotr; Thomas, Cecilia Engel; Lage, Kasper; Brunak, Søren

    2013-07-01

    MetaRanker 2.0 is a web server for prioritization of common and rare frequency genetic variation data. Based on heterogeneous data sets including genetic association data, protein-protein interactions, large-scale text-mining data, copy number variation data and gene expression experiments, MetaRanker 2.0 prioritizes the protein-coding part of the human genome to shortlist candidate genes for targeted follow-up studies. MetaRanker 2.0 is made freely available at www.cbs.dtu.dk/services/MetaRanker-2.0.

  6. Implementing WebGL and HTML5 in Macromolecular Visualization and Modern Computer-Aided Drug Design.

    PubMed

    Yuan, Shuguang; Chan, H C Stephen; Hu, Zhenquan

    2017-06-01

    Web browsers have long been recognized as potential platforms for remote macromolecule visualization. However, the difficulty in transferring large-scale data to clients and the lack of native support for hardware-accelerated applications in the local browser undermine the feasibility of such utilities. With the introduction of WebGL and HTML5 technologies in recent years, it is now possible to exploit the power of a graphics-processing unit (GPU) from a browser without any third-party plugin. Many new tools have been developed for biological molecule visualization and modern drug discovery. In contrast to traditional offline tools, real-time computing, interactive data analysis, and cross-platform analyses feature WebGL- and HTML5-based tools, facilitating biological research in a more efficient and user-friendly way. Copyright © 2017 Elsevier Ltd. All rights reserved.

  7. Text categorization models for identifying unproven cancer treatments on the web.

    PubMed

    Aphinyanaphongs, Yin; Aliferis, Constantin

    2007-01-01

    The nature of the internet as a non-peer-reviewed (and largely unregulated) publication medium has allowed wide-spread promotion of inaccurate and unproven medical claims in unprecedented scale. Patients with conditions that are not currently fully treatable are particularly susceptible to unproven and dangerous promises about miracle treatments. In extreme cases, fatal adverse outcomes have been documented. Most commonly, the cost is financial, psychological, and delayed application of imperfect but proven scientific modalities. To help protect patients, who may be desperately ill and thus prone to exploitation, we explored the use of machine learning techniques to identify web pages that make unproven claims. This feasibility study shows that the resulting models can identify web pages that make unproven claims in a fully automatic manner, and substantially better than previous web tools and state-of-the-art search engine technology.

  8. RBscore&NBench: a high-level web server for nucleic acid binding residues prediction with a large-scale benchmarking database.

    PubMed

    Miao, Zhichao; Westhof, Eric

    2016-07-08

    RBscore&NBench combines a web server, RBscore and a database, NBench. RBscore predicts RNA-/DNA-binding residues in proteins and visualizes the prediction scores and features on protein structures. The scoring scheme of RBscore directly links feature values to nucleic acid binding probabilities and illustrates the nucleic acid binding energy funnel on the protein surface. To avoid dataset, binding site definition and assessment metric biases, we compared RBscore with 18 web servers and 3 stand-alone programs on 41 datasets, which demonstrated the high and stable accuracy of RBscore. A comprehensive comparison led us to develop a benchmark database named NBench. The web server is available on: http://ahsoka.u-strasbg.fr/rbscorenbench/. © The Author(s) 2016. Published by Oxford University Press on behalf of Nucleic Acids Research.

  9. Geospatial Optimization of Siting Large-Scale Solar Projects

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Macknick, Jordan; Quinby, Ted; Caulfield, Emmet

    2014-03-01

    Recent policy and economic conditions have encouraged a renewed interest in developing large-scale solar projects in the U.S. Southwest. However, siting large-scale solar projects is complex. In addition to the quality of the solar resource, solar developers must take into consideration many environmental, social, and economic factors when evaluating a potential site. This report describes a proof-of-concept, Web-based Geographical Information Systems (GIS) tool that evaluates multiple user-defined criteria in an optimization algorithm to inform discussions and decisions regarding the locations of utility-scale solar projects. Existing siting recommendations for large-scale solar projects from governmental and non-governmental organizations are not consistent withmore » each other, are often not transparent in methods, and do not take into consideration the differing priorities of stakeholders. The siting assistance GIS tool we have developed improves upon the existing siting guidelines by being user-driven, transparent, interactive, capable of incorporating multiple criteria, and flexible. This work provides the foundation for a dynamic siting assistance tool that can greatly facilitate siting decisions among multiple stakeholders.« less

  10. Global change in the trophic functioning of marine food webs.

    PubMed

    Maureaud, Aurore; Gascuel, Didier; Colléter, Mathieu; Palomares, Maria L D; Du Pontavice, Hubert; Pauly, Daniel; Cheung, William W L

    2017-01-01

    The development of fisheries in the oceans, and other human drivers such as climate warming, have led to changes in species abundance, assemblages, trophic interactions, and ultimately in the functioning of marine food webs. Here, using a trophodynamic approach and global databases of catches and life history traits of marine species, we tested the hypothesis that anthropogenic ecological impacts may have led to changes in the global parameters defining the transfers of biomass within the food web. First, we developed two indicators to assess such changes: the Time Cumulated Indicator (TCI) measuring the residence time of biomass within the food web, and the Efficiency Cumulated Indicator (ECI) quantifying the fraction of secondary production reaching the top of the trophic chain. Then, we assessed, at the large marine ecosystem scale, the worldwide change of these two indicators over the 1950-2010 time-periods. Global trends were identified and cluster analyses were used to characterize the variability of trends between ecosystems. Results showed that the most common pattern over the study period is a global decrease in TCI, while the ECI indicator tends to increase. Thus, changes in species assemblages would induce faster and apparently more efficient biomass transfers in marine food webs. Results also suggested that the main driver of change over that period had been the large increase in fishing pressure. The largest changes occurred in ecosystems where 'fishing down the marine food web' are most intensive.

  11. The BCube Crawler: Web Scale Data and Service Discovery for EarthCube.

    NASA Astrophysics Data System (ADS)

    Lopez, L. A.; Khalsa, S. J. S.; Duerr, R.; Tayachow, A.; Mingo, E.

    2014-12-01

    Web-crawling, a core component of the NSF-funded BCube project, is researching and applying the use of big data technologies to find and characterize different types of web services, catalog interfaces, and data feeds such as the ESIP OpenSearch, OGC W*S, THREDDS, and OAI-PMH that describe or provide access to scientific datasets. Given the scale of the Internet, which challenges even large search providers such as Google, the BCube plan for discovering these web accessible services is to subdivide the problem into three smaller, more tractable issues. The first, to be able to discover likely sites where relevant data and data services might be found, the second, to be able to deeply crawl the sites discovered to find any data and services which might be present. Lastly, to leverage the use of semantic technologies to characterize the services and data found, and to filter out everything but those relevant to the geosciences. To address the first two challenges BCube uses an adapted version of Apache Nutch (which originated Hadoop), a web scale crawler, and Amazon's ElasticMapReduce service for flexibility and cost effectiveness. For characterization of the services found, BCube is examining existing web service ontologies for their applicability to our needs and will re-use and/or extend these in order to query for services with specific well-defined characteristics in scientific datasets such as the use of geospatial namespaces. The original proposal for the crawler won a grant from Amazon's academic program, which allowed us to become operational; we successfully tested the Bcube Crawler at web scale obtaining a significant corpus, sizeable enough to enable work on characterization of the services and data found. There is still plenty of work to be done, doing "smart crawls" by managing the frontier, developing and enhancing our scoring algorithms and fully implementing the semantic characterization technologies. We describe the current status of the project, our successes and issues encountered. The final goal of the BCube crawler project is to provide relevant data services to other projects on the EarthCube stack and third party partners so they can be brokered and used by a wider scientific community.

  12. Health-Terrain: Visualizing Large Scale Health Data

    DTIC Science & Technology

    2014-12-01

    systems can only be realized if the quality of emerging large medical databases can be characterized and the meaning of the data understood. For this...Designed and tested an evaluation procedure for health data visualization system. This visualization framework offers a real time and web-based solution...rule is shown in the table, with the quality measures of each rule including the support, confidence, Laplace, Gain, p-s, lift and Conviction. We

  13. Identifying unproven cancer treatments on the health web: addressing accuracy, generalizability and scalability.

    PubMed

    Aphinyanaphongs, Yin; Fu, Lawrence D; Aliferis, Constantin F

    2013-01-01

    Building machine learning models that identify unproven cancer treatments on the Health Web is a promising approach for dealing with the dissemination of false and dangerous information to vulnerable health consumers. Aside from the obvious requirement of accuracy, two issues are of practical importance in deploying these models in real world applications. (a) Generalizability: The models must generalize to all treatments (not just the ones used in the training of the models). (b) Scalability: The models can be applied efficiently to billions of documents on the Health Web. First, we provide methods and related empirical data demonstrating strong accuracy and generalizability. Second, by combining the MapReduce distributed architecture and high dimensionality compression via Markov Boundary feature selection, we show how to scale the application of the models to WWW-scale corpora. The present work provides evidence that (a) a very small subset of unproven cancer treatments is sufficient to build a model to identify unproven treatments on the web; (b) unproven treatments use distinct language to market their claims and this language is learnable; (c) through distributed parallelization and state of the art feature selection, it is possible to prepare the corpora and build and apply models with large scalability.

  14. Web-based scoring of the dicentric assay, a collaborative biodosimetric scoring strategy for population triage in large scale radiation accidents.

    PubMed

    Romm, H; Ainsbury, E; Bajinskis, A; Barnard, S; Barquinero, J F; Barrios, L; Beinke, C; Puig-Casanovas, R; Deperas-Kaminska, M; Gregoire, E; Oestreicher, U; Lindholm, C; Moquet, J; Rothkamm, K; Sommer, S; Thierens, H; Vral, A; Vandersickel, V; Wojcik, A

    2014-05-01

    In the case of a large scale radiation accident high throughput methods of biological dosimetry for population triage are needed to identify individuals requiring clinical treatment. The dicentric assay performed in web-based scoring mode may be a very suitable technique. Within the MULTIBIODOSE EU FP7 project a network is being established of 8 laboratories with expertise in dose estimations based on the dicentric assay. Here, the manual dicentric assay was tested in a web-based scoring mode. More than 23,000 high resolution images of metaphase spreads (only first mitosis) were captured by four laboratories and established as image galleries on the internet (cloud). The galleries included images of a complete dose effect curve (0-5.0 Gy) and three types of irradiation scenarios simulating acute whole body, partial body and protracted exposure. The blood samples had been irradiated in vitro with gamma rays at the University of Ghent, Belgium. Two laboratories provided image galleries from Fluorescence plus Giemsa stained slides (3 h colcemid) and the image galleries from the other two laboratories contained images from Giemsa stained preparations (24 h colcemid). Each of the 8 participating laboratories analysed 3 dose points of the dose effect curve (scoring 100 cells for each point) and 3 unknown dose points (50 cells) for each of the 3 simulated irradiation scenarios. At first all analyses were performed in a QuickScan Mode without scoring individual chromosomes, followed by conventional scoring (only complete cells, 46 centromeres). The calibration curves obtained using these two scoring methods were very similar, with no significant difference in the linear-quadratic curve coefficients. Analysis of variance showed a significant effect of dose on the yield of dicentrics, but no significant effect of the laboratories, different methods of slide preparation or different incubation times used for colcemid. The results obtained to date within the MULTIBIODOSE project by a network of 8 collaborating laboratories throughout Europe are very promising. The dicentric assay in the web based scoring mode as a high throughput scoring strategy is a useful application for biodosimetry in the case of a large scale radiation accident.

  15. A Multi-scale Cognitive Approach to Intrusion Detection and Response

    DTIC Science & Technology

    2015-12-28

    the behavior of the traffic on the network, either by using mathematical formulas or by replaying packet streams. As a result, simulators depend...large scale. Summary of the most important results We obtained a powerful machine, which has 768 cores and 1.25 TB memory . RBG has been...time. Each client is configured with 1GB memory , 10 GB disk space, and one 100M Ethernet interface. The server nodes include web servers

  16. Decision Support Framework Implementation Of The Web-Based Environmental Decision Analysis DASEES: Decision Analysis For A Sustainable Environment, Economy, And Society

    EPA Science Inventory

    Solutions to pervasive environmental problems often are not amenable to a straightforward application of science-based actions. These problems encompass large-scale environmental policy questions where environmental concerns, economic constraints, and societal values conflict ca...

  17. Mutual Intelligibility between Closely Related Languages in Europe

    ERIC Educational Resources Information Center

    Gooskens, Charlotte; van Heuven, Vincent J.; Golubovic, Jelena; Schüppert, Anja; Swarte, Femke; Voigt, Stefanie

    2018-01-01

    By means of a large-scale web-based investigation, we established the degree of mutual intelligibility of 16 closely related spoken languages within the Germanic, Slavic and Romance language families in Europe. We first present the results of a selection of 1833 listeners representing the mutual intelligibility between young, educated Europeans…

  18. SCIMITAR: Scalable Stream-Processing for Sensor Information Brokering

    DTIC Science & Technology

    2013-11-01

    IaaS) cloud frameworks including Amazon Web Services and Eucalyptus . For load testing, we used The Grinder [9], a Java load testing framework that...internal Eucalyptus cluster which we could not scale as large as the Amazon environment due to a lack of computation resources. We recreated our

  19. Focused Crawling of the Deep Web Using Service Class Descriptions

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Rocco, D; Liu, L; Critchlow, T

    2004-06-21

    Dynamic Web data sources--sometimes known collectively as the Deep Web--increase the utility of the Web by providing intuitive access to data repositories anywhere that Web access is available. Deep Web services provide access to real-time information, like entertainment event listings, or present a Web interface to large databases or other data repositories. Recent studies suggest that the size and growth rate of the dynamic Web greatly exceed that of the static Web, yet dynamic content is often ignored by existing search engine indexers owing to the technical challenges that arise when attempting to search the Deep Web. To address thesemore » challenges, we present DynaBot, a service-centric crawler for discovering and clustering Deep Web sources offering dynamic content. DynaBot has three unique characteristics. First, DynaBot utilizes a service class model of the Web implemented through the construction of service class descriptions (SCDs). Second, DynaBot employs a modular, self-tuning system architecture for focused crawling of the DeepWeb using service class descriptions. Third, DynaBot incorporates methods and algorithms for efficient probing of the Deep Web and for discovering and clustering Deep Web sources and services through SCD-based service matching analysis. Our experimental results demonstrate the effectiveness of the service class discovery, probing, and matching algorithms and suggest techniques for efficiently managing service discovery in the face of the immense scale of the Deep Web.« less

  20. Design and implementation of a distributed large-scale spatial database system based on J2EE

    NASA Astrophysics Data System (ADS)

    Gong, Jianya; Chen, Nengcheng; Zhu, Xinyan; Zhang, Xia

    2003-03-01

    With the increasing maturity of distributed object technology, CORBA, .NET and EJB are universally used in traditional IT field. However, theories and practices of distributed spatial database need farther improvement in virtue of contradictions between large scale spatial data and limited network bandwidth or between transitory session and long transaction processing. Differences and trends among of CORBA, .NET and EJB are discussed in details, afterwards the concept, architecture and characteristic of distributed large-scale seamless spatial database system based on J2EE is provided, which contains GIS client application, web server, GIS application server and spatial data server. Moreover the design and implementation of components of GIS client application based on JavaBeans, the GIS engine based on servlet, the GIS Application server based on GIS enterprise JavaBeans(contains session bean and entity bean) are explained.Besides, the experiments of relation of spatial data and response time under different conditions are conducted, which proves that distributed spatial database system based on J2EE can be used to manage, distribute and share large scale spatial data on Internet. Lastly, a distributed large-scale seamless image database based on Internet is presented.

  1. The effect of top-level domains and advertisements on health web-site credibility.

    PubMed

    Walther, Joseph B; Wang, Zuoming; Loh, Tracy

    2004-09-03

    Concerns over health information on the Internet have generated efforts to enhance credibility markers; yet how users actually assess the credibility of online health information is largely unknown. This study set out to (1) establish a parsimonious and valid questionnaire instrument to measure credibility of Internet health information by drawing on various previous measures of source, news, and other credibility scales; and (2) to identify the effects of Web-site domains and advertising on credibility perceptions. Respondents (N = 156) examined one of 12 Web-site mock-ups and completed credibility scales in a 3 x 2 x 2 between-subjects experimental design. Factor analysis and validity checks were used for item reduction, and analysis of variance was employed for hypothesis testing of Web-site features' effects. In an attempt to construct a credibility instrument, three dimensions of credibility (safety, trustworthiness, and dynamism) were retained, reflecting traditional credibility sub-themes, but composed of items from disparate sources. When testing the effect of the presence or absence of advertising on a Web site on credibility, we found that this depends on the site's domain, with a trend for advertisements having deleterious effects on the credibility of sites with .org domain, but positive effects on sites with .com or .edu domains. Health-information Web-site providers should select domains purposefully when they can, especially if they must accept on-site advertising. Credibility perceptions may not be invariant or stable, but rather are sensitive to topic and context. Future research may employ these findings in order to compare other forms of health-information delivery to optimal Web-site features.

  2. Design and Implementation of Distributed Crawler System Based on Scrapy

    NASA Astrophysics Data System (ADS)

    Fan, Yuhao

    2018-01-01

    At present, some large-scale search engines at home and abroad only provide users with non-custom search services, and a single-machine web crawler cannot sovle the difficult task. In this paper, Through the study and research of the original Scrapy framework, the original Scrapy framework is improved by combining Scrapy and Redis, a distributed crawler system based on Web information Scrapy framework is designed and implemented, and Bloom Filter algorithm is applied to dupefilter modul to reduce memory consumption. The movie information captured from douban is stored in MongoDB, so that the data can be processed and analyzed. The results show that distributed crawler system based on Scrapy framework is more efficient and stable than the single-machine web crawler system.

  3. Galaxy evolution in the metric of the cosmic web

    NASA Astrophysics Data System (ADS)

    Kraljic, K.; Arnouts, S.; Pichon, C.; Laigle, C.; de la Torre, S.; Vibert, D.; Cadiou, C.; Dubois, Y.; Treyer, M.; Schimd, C.; Codis, S.; de Lapparent, V.; Devriendt, J.; Hwang, H. S.; Le Borgne, D.; Malavasi, N.; Milliard, B.; Musso, M.; Pogosyan, D.; Alpaslan, M.; Bland-Hawthorn, J.; Wright, A. H.

    2018-02-01

    The role of the cosmic web in shaping galaxy properties is investigated in the Galaxy And Mass Assembly (GAMA) spectroscopic survey in the redshift range 0.03 ≤ z ≤ 0.25. The stellar mass, u - r dust corrected colour and specific star formation rate (sSFR) of galaxies are analysed as a function of their distances to the 3D cosmic web features, such as nodes, filaments and walls, as reconstructed by DisPerSE. Significant mass and type/colour gradients are found for the whole population, with more massive and/or passive galaxies being located closer to the filament and wall than their less massive and/or star-forming counterparts. Mass segregation persists among the star-forming population alone. The red fraction of galaxies increases when closing in on nodes, and on filaments regardless of the distance to nodes. Similarly, the star-forming population reddens (or lowers its sSFR) at fixed mass when closing in on filament, implying that some quenching takes place. These trends are also found in the state-of-the-art hydrodynamical simulation HORIZON-AGN. These results suggest that on top of stellar mass and large-scale density, the traceless component of the tides from the anisotropic large-scale environment also shapes galactic properties. An extension of excursion theory accounting for filamentary tides provides a qualitative explanation in terms of anisotropic assembly bias: at a given mass, the accretion rate varies with the orientation and distance to filaments. It also explains the absence of type/colour gradients in the data on smaller, non-linear scales.

  4. Enabling Interactive Measurements from Large Coverage Microscopy

    PubMed Central

    Bajcsy, Peter; Vandecreme, Antoine; Amelot, Julien; Chalfoun, Joe; Majurski, Michael; Brady, Mary

    2017-01-01

    Microscopy could be an important tool for characterizing stem cell products if quantitative measurements could be collected over multiple spatial and temporal scales. With the cells changing states over time and being several orders of magnitude smaller than cell products, modern microscopes are already capable of imaging large spatial areas, repeat imaging over time, and acquiring images over several spectra. However, characterizing stem cell products from such large image collections is challenging because of data size, required computations, and lack of interactive quantitative measurements needed to determine release criteria. We present a measurement web system consisting of available algorithms, extensions to a client-server framework using Deep Zoom, and the configuration know-how to provide the information needed for inspecting the quality of a cell product. The cell and other data sets are accessible via the prototype web-based system at http://isg.nist.gov/deepzoomweb. PMID:28663600

  5. Web-Based and Mobile Stress Management Intervention for Employees: A Randomized Controlled Trial.

    PubMed

    Heber, Elena; Lehr, Dirk; Ebert, David Daniel; Berking, Matthias; Riper, Heleen

    2016-01-27

    Work-related stress is highly prevalent among employees and is associated with adverse mental health consequences. Web-based interventions offer the opportunity to deliver effective solutions on a large scale; however, the evidence is limited and the results conflicting. This randomized controlled trial evaluated the efficacy of guided Web- and mobile-based stress management training for employees. A total of 264 employees with elevated symptoms of stress (Perceived Stress Scale-10, PSS-10≥22) were recruited from the general working population and randomly assigned to an Internet-based stress management intervention (iSMI) or waitlist control group. The intervention (GET.ON Stress) was based on Lazarus's transactional model of stress, consisted of seven sessions, and applied both well-established problem solving and more recently developed emotion regulation strategies. Participants also had the opportunity to request automatic text messages on their mobile phone along with the iSMI. Participants received written feedback on every completed session from an e-coach. The primary outcome was perceived stress (PSS-10). Web-based self-report assessments for both groups were scheduled at baseline, 7 weeks, and 6 months. At 12 months, an extended follow-up was carried out for the iSMI group only. An intention-to-treat analysis of covariance revealed significantly large effect differences between iSMI and waitlist control groups for perceived stress at posttest (F1,261=58.08, P<.001; Cohen's d=0.83) and at the 6-month follow-up (F1,261=80.17, P<.001; Cohen's d=1.02). The effects in the iSMI group were maintained at 12-month follow-up. This Web- and mobile-based intervention has proven effective in reducing stress in employees in the long term. Internet-based stress management interventions should be further pursued as a valuable alternative to face-to-face interventions. German Clinical Trials Register (DRKS): 00004749; http://drks-neu.uniklinik-freiburg.de/ drks_web/setLocale_EN.do (Archived by WebCite at http://www.webcitation.org/6e8rl98nl).

  6. Plugin free remote visualization in the browser

    NASA Astrophysics Data System (ADS)

    Tamm, Georg; Slusallek, Philipp

    2015-01-01

    Today, users access information and rich media from anywhere using the web browser on their desktop computers, tablets or smartphones. But the web evolves beyond media delivery. Interactive graphics applications like visualization or gaming become feasible as browsers advance in the functionality they provide. However, to deliver large-scale visualization to thin clients like mobile devices, a dedicated server component is necessary. Ideally, the client runs directly within the browser the user is accustomed to, requiring no installation of a plugin or native application. In this paper, we present the state-of-the-art of technologies which enable plugin free remote rendering in the browser. Further, we describe a remote visualization system unifying these technologies. The system transfers rendering results to the client as images or as a video stream. We utilize the upcoming World Wide Web Consortium (W3C) conform Web Real-Time Communication (WebRTC) standard, and the Native Client (NaCl) technology built into Chrome, to deliver video with low latency.

  7. Lagrangian methods of cosmic web classification

    NASA Astrophysics Data System (ADS)

    Fisher, J. D.; Faltenbacher, A.; Johnson, M. S. T.

    2016-05-01

    The cosmic web defines the large-scale distribution of matter we see in the Universe today. Classifying the cosmic web into voids, sheets, filaments and nodes allows one to explore structure formation and the role environmental factors have on halo and galaxy properties. While existing studies of cosmic web classification concentrate on grid-based methods, this work explores a Lagrangian approach where the V-web algorithm proposed by Hoffman et al. is implemented with techniques borrowed from smoothed particle hydrodynamics. The Lagrangian approach allows one to classify individual objects (e.g. particles or haloes) based on properties of their nearest neighbours in an adaptive manner. It can be applied directly to a halo sample which dramatically reduces computational cost and potentially allows an application of this classification scheme to observed galaxy samples. Finally, the Lagrangian nature admits a straightforward inclusion of the Hubble flow negating the necessity of a visually defined threshold value which is commonly employed by grid-based classification methods.

  8. NaviCell Web Service for network-based data visualization.

    PubMed

    Bonnet, Eric; Viara, Eric; Kuperstein, Inna; Calzone, Laurence; Cohen, David P A; Barillot, Emmanuel; Zinovyev, Andrei

    2015-07-01

    Data visualization is an essential element of biological research, required for obtaining insights and formulating new hypotheses on mechanisms of health and disease. NaviCell Web Service is a tool for network-based visualization of 'omics' data which implements several data visual representation methods and utilities for combining them together. NaviCell Web Service uses Google Maps and semantic zooming to browse large biological network maps, represented in various formats, together with different types of the molecular data mapped on top of them. For achieving this, the tool provides standard heatmaps, barplots and glyphs as well as the novel map staining technique for grasping large-scale trends in numerical values (such as whole transcriptome) projected onto a pathway map. The web service provides a server mode, which allows automating visualization tasks and retrieving data from maps via RESTful (standard HTTP) calls. Bindings to different programming languages are provided (Python and R). We illustrate the purpose of the tool with several case studies using pathway maps created by different research groups, in which data visualization provides new insights into molecular mechanisms involved in systemic diseases such as cancer and neurodegenerative diseases. © The Author(s) 2015. Published by Oxford University Press on behalf of Nucleic Acids Research.

  9. Online self-administered training for post-traumatic stress disorder treatment providers: design and methods for a randomized, prospective intervention study

    PubMed Central

    2012-01-01

    This paper presents the rationale and methods for a randomized controlled evaluation of web-based training in motivational interviewing, goal setting, and behavioral task assignment. Web-based training may be a practical and cost-effective way to address the need for large-scale mental health training in evidence-based practice; however, there is a dearth of well-controlled outcome studies of these approaches. For the current trial, 168 mental health providers treating post-traumatic stress disorder (PTSD) were assigned to web-based training plus supervision, web-based training, or training-as-usual (control). A novel standardized patient (SP) assessment was developed and implemented for objective measurement of changes in clinical skills, while on-line self-report measures were used for assessing changes in knowledge, perceived self-efficacy, and practice related to cognitive behavioral therapy (CBT) techniques. Eligible participants were all actively involved in mental health treatment of veterans with PTSD. Study methodology illustrates ways of developing training content, recruiting participants, and assessing knowledge, perceived self-efficacy, and competency-based outcomes, and demonstrates the feasibility of conducting prospective studies of training efficacy or effectiveness in large healthcare systems. PMID:22583520

  10. NaviCell Web Service for network-based data visualization

    PubMed Central

    Bonnet, Eric; Viara, Eric; Kuperstein, Inna; Calzone, Laurence; Cohen, David P. A.; Barillot, Emmanuel; Zinovyev, Andrei

    2015-01-01

    Data visualization is an essential element of biological research, required for obtaining insights and formulating new hypotheses on mechanisms of health and disease. NaviCell Web Service is a tool for network-based visualization of ‘omics’ data which implements several data visual representation methods and utilities for combining them together. NaviCell Web Service uses Google Maps and semantic zooming to browse large biological network maps, represented in various formats, together with different types of the molecular data mapped on top of them. For achieving this, the tool provides standard heatmaps, barplots and glyphs as well as the novel map staining technique for grasping large-scale trends in numerical values (such as whole transcriptome) projected onto a pathway map. The web service provides a server mode, which allows automating visualization tasks and retrieving data from maps via RESTful (standard HTTP) calls. Bindings to different programming languages are provided (Python and R). We illustrate the purpose of the tool with several case studies using pathway maps created by different research groups, in which data visualization provides new insights into molecular mechanisms involved in systemic diseases such as cancer and neurodegenerative diseases. PMID:25958393

  11. High-Resolution Climate Data Visualization through GIS- and Web-based Data Portals

    NASA Astrophysics Data System (ADS)

    WANG, X.; Huang, G.

    2017-12-01

    Sound decisions on climate change adaptation rely on an in-depth assessment of potential climate change impacts at regional and local scales, which usually requires finer resolution climate projections at both spatial and temporal scales. However, effective downscaling of global climate projections is practically difficult due to the lack of computational resources and/or long-term reference data. Although a large volume of downscaled climate data has been make available to the public, how to understand and interpret the large-volume climate data and how to make use of the data to drive impact assessment and adaptation studies are still challenging for both impact researchers and decision makers. Such difficulties have become major barriers preventing informed climate change adaptation planning at regional scales. Therefore, this research will explore new GIS- and web-based technologies to help visualize the large-volume regional climate data with high spatiotemporal resolutions. A user-friendly public data portal, named Climate Change Data Portal (CCDP, http://ccdp.network), will be established to allow intuitive and open access to high-resolution regional climate projections at local scales. The CCDP offers functions of visual representation through geospatial maps and data downloading for a variety of climate variables (e.g., temperature, precipitation, relative humidity, solar radiation, and wind) at multiple spatial resolutions (i.e., 25 - 50 km) and temporal resolutions (i.e., annual, seasonal, monthly, daily, and hourly). The vast amount of information the CCDP encompasses can provide a crucial basis for assessing impacts of climate change on local communities and ecosystems and for supporting better decision making under a changing climate.

  12. First Operational Experience With a High-Energy Physics Run Control System Based on Web Technologies

    NASA Astrophysics Data System (ADS)

    Bauer, Gerry; Beccati, Barbara; Behrens, Ulf; Biery, Kurt; Branson, James; Bukowiec, Sebastian; Cano, Eric; Cheung, Harry; Ciganek, Marek; Cittolin, Sergio; Coarasa Perez, Jose Antonio; Deldicque, Christian; Erhan, Samim; Gigi, Dominique; Glege, Frank; Gomez-Reino, Robert; Gulmini, Michele; Hatton, Derek; Hwong, Yi Ling; Loizides, Constantin; Ma, Frank; Masetti, Lorenzo; Meijers, Frans; Meschi, Emilio; Meyer, Andreas; Mommsen, Remigius K.; Moser, Roland; O'Dell, Vivian; Oh, Alexander; Orsini, Luciano; Paus, Christoph; Petrucci, Andrea; Pieri, Marco; Racz, Attila; Raginel, Olivier; Sakulin, Hannes; Sani, Matteo; Schieferdecker, Philipp; Schwick, Christoph; Shpakov, Dennis; Simon, Michal; Sumorok, Konstanty; Yoon, Andre Sungho

    2012-08-01

    Run control systems of modern high-energy particle physics experiments have requirements similar to those of today's Internet applications. The Compact Muon Solenoid (CMS) collaboration at CERN's Large Hadron Collider (LHC) therefore decided to build the run control system for its detector based on web technologies. The system is composed of Java Web Applications distributed over a set of Apache Tomcat servlet containers that connect to a database back-end. Users interact with the system through a web browser. The present paper reports on the successful scaling of the system from a small test setup to the production data acquisition system that comprises around 10.000 applications running on a cluster of about 1600 hosts. We report on operational aspects during the first phase of operation with colliding beams including performance, stability, integration with the CMS Detector Control System and tools to guide the operator.

  13. Integrating Cloud-Computing-Specific Model into Aircraft Design

    NASA Astrophysics Data System (ADS)

    Zhimin, Tian; Qi, Lin; Guangwen, Yang

    Cloud Computing is becoming increasingly relevant, as it will enable companies involved in spreading this technology to open the door to Web 3.0. In the paper, the new categories of services introduced will slowly replace many types of computational resources currently used. In this perspective, grid computing, the basic element for the large scale supply of cloud services, will play a fundamental role in defining how those services will be provided. The paper tries to integrate cloud computing specific model into aircraft design. This work has acquired good results in sharing licenses of large scale and expensive software, such as CFD (Computational Fluid Dynamics), UG, CATIA, and so on.

  14. Dusty Starbursts within a z=3 Large Scale Structure revealed by ALMA

    NASA Astrophysics Data System (ADS)

    Umehata, Hideki

    The role of the large-scale structure is one of the most important theme in studying galaxy formation and evolution. However, it has been still mystery especially at z>2. On the basis of our ALMA 1.1 mm observations in a z ~ 3 protocluster field, it is suggested that submillimeter galaxies (SMGs) preferentially reside in the densest environment at z ~ 3. Furthermore we find a rich cluster of AGN-host SMGs at the core of the protocluster, combining with Chandra X-ray data. Our results indicate the vigorous star-formation and accelerated super massive black hole (SMBH) growth in the node of the cosmic web.

  15. Cosmic void clumps

    NASA Astrophysics Data System (ADS)

    Lares, M.; Luparello, H. E.; Garcia Lambas, D.; Ruiz, A. N.; Ceccarelli, L.; Paz, D.

    2017-10-01

    Cosmic voids are of great interest given their relation to the large scale distribution of mass and the way they trace cosmic flows shaping the cosmic web. Here we show that the distribution of voids has, in consonance with the distribution of mass, a characteristic scale at which void pairs are preferentially located. We identify clumps of voids with similar environments and use them to define second order underdensities. Also, we characterize its properties and analyze its impact on the cosmic microwave background.

  16. Googling DNA sequences on the World Wide Web.

    PubMed

    Hajibabaei, Mehrdad; Singer, Gregory A C

    2009-11-10

    New web-based technologies provide an excellent opportunity for sharing and accessing information and using web as a platform for interaction and collaboration. Although several specialized tools are available for analyzing DNA sequence information, conventional web-based tools have not been utilized for bioinformatics applications. We have developed a novel algorithm and implemented it for searching species-specific genomic sequences, DNA barcodes, by using popular web-based methods such as Google. We developed an alignment independent character based algorithm based on dividing a sequence library (DNA barcodes) and query sequence to words. The actual search is conducted by conventional search tools such as freely available Google Desktop Search. We implemented our algorithm in two exemplar packages. We developed pre and post-processing software to provide customized input and output services, respectively. Our analysis of all publicly available DNA barcode sequences shows a high accuracy as well as rapid results. Our method makes use of conventional web-based technologies for specialized genetic data. It provides a robust and efficient solution for sequence search on the web. The integration of our search method for large-scale sequence libraries such as DNA barcodes provides an excellent web-based tool for accessing this information and linking it to other available categories of information on the web.

  17. Development of grid-like applications for public health using Web 2.0 mashup techniques.

    PubMed

    Scotch, Matthew; Yip, Kevin Y; Cheung, Kei-Hoi

    2008-01-01

    Development of public health informatics applications often requires the integration of multiple data sources. This process can be challenging due to issues such as different file formats, schemas, naming systems, and having to scrape the content of web pages. A potential solution to these system development challenges is the use of Web 2.0 technologies. In general, Web 2.0 technologies are new internet services that encourage and value information sharing and collaboration among individuals. In this case report, we describe the development and use of Web 2.0 technologies including Yahoo! Pipes within a public health application that integrates animal, human, and temperature data to assess the risk of West Nile Virus (WNV) outbreaks. The results of development and testing suggest that while Web 2.0 applications are reasonable environments for rapid prototyping, they are not mature enough for large-scale public health data applications. The application, in fact a "systems of systems," often failed due to varied timeouts for application response across web sites and services, internal caching errors, and software added to web sites by administrators to manage the load on their servers. In spite of these concerns, the results of this study demonstrate the potential value of grid computing and Web 2.0 approaches in public health informatics.

  18. Observing Interstellar and Intergalactic Magnetic Fields

    NASA Astrophysics Data System (ADS)

    Han, J. L.

    2017-08-01

    Observational results of interstellar and intergalactic magnetic fields are reviewed, including the fields in supernova remnants and loops, interstellar filaments and clouds, Hii regions and bubbles, the Milky Way and nearby galaxies, galaxy clusters, and the cosmic web. A variety of approaches are used to investigate these fields. The orientations of magnetic fields in interstellar filaments and molecular clouds are traced by polarized thermal dust emission and starlight polarization. The field strengths and directions along the line of sight in dense clouds and cores are measured by Zeeman splitting of emission or absorption lines. The large-scale magnetic fields in the Milky Way have been best probed by Faraday rotation measures of a large number of pulsars and extragalactic radio sources. The coherent Galactic magnetic fields are found to follow the spiral arms and have their direction reversals in arms and interarm regions in the disk. The azimuthal fields in the halo reverse their directions below and above the Galactic plane. The orientations of organized magnetic fields in nearby galaxies have been observed through polarized synchrotron emission. Magnetic fields in the intracluster medium have been indicated by diffuse radio halos, polarized radio relics, and Faraday rotations of embedded radio galaxies and background sources. Sparse evidence for very weak magnetic fields in the cosmic web is the detection of the faint radio bridge between the Coma cluster and A1367. Future observations should aim at the 3D tomography of the large-scale coherent magnetic fields in our Galaxy and nearby galaxies, a better description of intracluster field properties, and firm detections of intergalactic magnetic fields in the cosmic web.

  19. Collaboration in the Humanities, Arts and Social Sciences in Australia

    ERIC Educational Resources Information Center

    Haddow, Gaby; Xia, Jianhong; Willson, Michele

    2017-01-01

    This paper reports on the first large-scale quantitative investigation into collaboration, demonstrated in co-authorship, by Australian humanities, arts and social sciences (HASS) researchers. Web of Science data were extracted for Australian HASS publications, with a focus on the softer social sciences, over the period 2004-2013. The findings…

  20. Linking native and invader traits explains native spider population responses to plant invasion

    Treesearch

    Jennifer N. Smith; Douglas J. Emlen; Dean E. Pearson

    2016-01-01

    Theoretically, the functional traits of native species should determine how natives respond to invader-driven changes. To explore this idea, we simulated a large-scale plant invasion using dead spotted knapweed (Centaurea stoebe) stems to determine if native spiders' web-building behaviors could explain differences in spider population responses to...

  1. Connectivist Learning Objects and Learning Styles

    ERIC Educational Resources Information Center

    del Moral, M. Esther; Cernea, Ana; Villalustre, Lourdes

    2013-01-01

    The Web 2.0 brought in the use of social tools at a large scale in every area: a transformation which led to redefining the teaching-learning process. In this new context knowledge is distributed over network connections in an uncontrolled way - thus learning consists of recognizing relevant information patterns and constructing new connections.…

  2. Electronic Scientific Data & Literature Aggregation: A Review for Librarians

    ERIC Educational Resources Information Center

    Losoff, Barbara

    2009-01-01

    The advent of large-scale digital repositories, along with the need for sharing useful data world-wide, demands change to the current information structure. The merging of digital scientific data with scholarly literature has the potential to fulfill the Semantic Web design principles. This paper will identify factors leading to integration of…

  3. Network-based approaches to climate knowledge discovery

    NASA Astrophysics Data System (ADS)

    Budich, Reinhard; Nyberg, Per; Weigel, Tobias

    2011-11-01

    Climate Knowledge Discovery Workshop; Hamburg, Germany, 30 March to 1 April 2011 Do complex networks combined with semantic Web technologies offer the next generation of solutions in climate science? To address this question, a first Climate Knowledge Discovery (CKD) Workshop, hosted by the German Climate Computing Center (Deutsches Klimarechenzentrum (DKRZ)), brought together climate and computer scientists from major American and European laboratories, data centers, and universities, as well as representatives from industry, the broader academic community, and the semantic Web communities. The participants, representing six countries, were concerned with large-scale Earth system modeling and computational data analysis. The motivation for the meeting was the growing problem that climate scientists generate data faster than it can be interpreted and the need to prepare for further exponential data increases. Current analysis approaches are focused primarily on traditional methods, which are best suited for large-scale phenomena and coarse-resolution data sets. The workshop focused on the open discussion of ideas and technologies to provide the next generation of solutions to cope with the increasing data volumes in climate science.

  4. GenomeVx: simple web-based creation of editable circular chromosome maps.

    PubMed

    Conant, Gavin C; Wolfe, Kenneth H

    2008-03-15

    We describe GenomeVx, a web-based tool for making editable, publication-quality, maps of mitochondrial and chloroplast genomes and of large plasmids. These maps show the location of genes and chromosomal features as well as a position scale. The program takes as input either raw feature positions or GenBank records. In the latter case, features are automatically extracted and colored, an example of which is given. Output is in the Adobe Portable Document Format (PDF) and can be edited by programs such as Adobe Illustrator. GenomeVx is available at http://wolfe.gen.tcd.ie/GenomeVx

  5. Directed module detection in a large-scale expression compendium.

    PubMed

    Fu, Qiang; Lemmens, Karen; Sanchez-Rodriguez, Aminael; Thijs, Inge M; Meysman, Pieter; Sun, Hong; Fierro, Ana Carolina; Engelen, Kristof; Marchal, Kathleen

    2012-01-01

    Public online microarray databases contain tremendous amounts of expression data. Mining these data sources can provide a wealth of information on the underlying transcriptional networks. In this chapter, we illustrate how the web services COLOMBOS and DISTILLER can be used to identify condition-dependent coexpression modules by exploring compendia of public expression data. COLOMBOS is designed for user-specified query-driven analysis, whereas DISTILLER generates a global regulatory network overview. The user is guided through both web services by means of a case study in which condition-dependent coexpression modules comprising a gene of interest (i.e., "directed") are identified.

  6. Opal web services for biomedical applications.

    PubMed

    Ren, Jingyuan; Williams, Nadya; Clementi, Luca; Krishnan, Sriram; Li, Wilfred W

    2010-07-01

    Biomedical applications have become increasingly complex, and they often require large-scale high-performance computing resources with a large number of processors and memory. The complexity of application deployment and the advances in cluster, grid and cloud computing require new modes of support for biomedical research. Scientific Software as a Service (sSaaS) enables scalable and transparent access to biomedical applications through simple standards-based Web interfaces. Towards this end, we built a production web server (http://ws.nbcr.net) in August 2007 to support the bioinformatics application called MEME. The server has grown since to include docking analysis with AutoDock and AutoDock Vina, electrostatic calculations using PDB2PQR and APBS, and off-target analysis using SMAP. All the applications on the servers are powered by Opal, a toolkit that allows users to wrap scientific applications easily as web services without any modification to the scientific codes, by writing simple XML configuration files. Opal allows both web forms-based access and programmatic access of all our applications. The Opal toolkit currently supports SOAP-based Web service access to a number of popular applications from the National Biomedical Computation Resource (NBCR) and affiliated collaborative and service projects. In addition, Opal's programmatic access capability allows our applications to be accessed through many workflow tools, including Vision, Kepler, Nimrod/K and VisTrails. From mid-August 2007 to the end of 2009, we have successfully executed 239,814 jobs. The number of successfully executed jobs more than doubled from 205 to 411 per day between 2008 and 2009. The Opal-enabled service model is useful for a wide range of applications. It provides for interoperation with other applications with Web Service interfaces, and allows application developers to focus on the scientific tool and workflow development. Web server availability: http://ws.nbcr.net.

  7. SETTER: web server for RNA structure comparison

    PubMed Central

    Čech, Petr; Svozil, Daniel; Hoksza, David

    2012-01-01

    The recent discoveries of regulatory non-coding RNAs changed our view of RNA as a simple information transfer molecule. Understanding the architecture and function of active RNA molecules requires methods for comparing and analyzing their 3D structures. While structural alignment of short RNAs is achievable in a reasonable amount of time, large structures represent much bigger challenge. Here, we present the SETTER web server for the RNA structure pairwise comparison utilizing the SETTER (SEcondary sTructure-based TERtiary Structure Similarity Algorithm) algorithm. The SETTER method divides an RNA structure into the set of non-overlapping structural elements called generalized secondary structure units (GSSUs). The SETTER algorithm scales as O(n2) with the size of a GSSUs and as O(n) with the number of GSSUs in the structure. This scaling gives SETTER its high speed as the average size of the GSSU remains constant irrespective of the size of the structure. However, the favorable speed of the algorithm does not compromise its accuracy. The SETTER web server together with the stand-alone implementation of the SETTER algorithm are freely accessible at http://siret.cz/setter. PMID:22693209

  8. Analysis and visualization of Arabidopsis thaliana GWAS using web 2.0 technologies.

    PubMed

    Huang, Yu S; Horton, Matthew; Vilhjálmsson, Bjarni J; Seren, Umit; Meng, Dazhe; Meyer, Christopher; Ali Amer, Muhammad; Borevitz, Justin O; Bergelson, Joy; Nordborg, Magnus

    2011-01-01

    With large-scale genomic data becoming the norm in biological studies, the storing, integrating, viewing and searching of such data have become a major challenge. In this article, we describe the development of an Arabidopsis thaliana database that hosts the geographic information and genetic polymorphism data for over 6000 accessions and genome-wide association study (GWAS) results for 107 phenotypes representing the largest collection of Arabidopsis polymorphism data and GWAS results to date. Taking advantage of a series of the latest web 2.0 technologies, such as Ajax (Asynchronous JavaScript and XML), GWT (Google-Web-Toolkit), MVC (Model-View-Controller) web framework and Object Relationship Mapper, we have created a web-based application (web app) for the database, that offers an integrated and dynamic view of geographic information, genetic polymorphism and GWAS results. Essential search functionalities are incorporated into the web app to aid reverse genetics research. The database and its web app have proven to be a valuable resource to the Arabidopsis community. The whole framework serves as an example of how biological data, especially GWAS, can be presented and accessed through the web. In the end, we illustrate the potential to gain new insights through the web app by two examples, showcasing how it can be used to facilitate forward and reverse genetics research. Database URL: http://arabidopsis.usc.edu/

  9. The Effect of Top-Level Domains and Advertisements on Health Web Site Credibility

    PubMed Central

    Wang, Zuoming; Loh, Tracy

    2004-01-01

    Background Concerns over health information on the Internet have generated efforts to enhance credibility markers; yet how users actually assess the credibility of online health information is largely unknown. Objective This study set out to (1) establish a parsimonious and valid questionnaire instrument to measure credibility of Internet health information by drawing on various previous measures of source, news, and other credibility scales; and (2) to identify the effects of Web-site domains and advertising on credibility perceptions. Methods Respondents (N = 156) examined one of 12 Web-site mock-ups and completed credibility scales in a 3 x 2 x 2 between-subjects experimental design. Factor analysis and validity checks were used for item reduction, and analysis of variance was employed for hypothesis testing of Web-site features' effects. Results In an attempt to construct a credibility instrument, three dimensions of credibility (safety, trustworthiness, and dynamism) were retained, reflecting traditional credibility sub-themes, but composed of items from disparate sources. When testing the effect of the presence or absence of advertising on a Web site on credibility, we found that this depends on the site's domain, with a trend for advertisements having deleterious effects on the credibility of sites with .org domain, but positive effects on sites with .com or .edu domains. Conclusions Health-information Web-site providers should select domains purposefully when they can, especially if they must accept on-site advertising. Credibility perceptions may not be invariant or stable, but rather are sensitive to topic and context. Future research may employ these findings in order to compare other forms of health-information delivery to optimal Web-site features. PMID:15471750

  10. Cosmic web and star formation activity in galaxies at z ∼ 1

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Darvish, B.; Mobasher, B.; Sales, L. V.

    We investigate the role of the delineated cosmic web/filaments on star formation activity by exploring a sample of 425 narrow-band selected Hα emitters, as well as 2846 color-color selected underlying star-forming galaxies for a large-scale structure at z = 0.84 in the COSMOS field from the HiZELS survey. Using the scale-independent Multi-scale Morphology Filter algorithm, we are able to quantitatively describe the density field and disentangle it into its major components: fields, filaments, and clusters. We show that the observed median star formation rate (SFR), stellar mass, specific SFR, the mean SFR-mass relation, and its scatter for both Hα emittersmore » and underlying star-forming galaxies do not strongly depend on different classes of environment, in agreement with previous studies. However, the fraction of Hα emitters varies with environment and is enhanced in filamentary structures at z ∼ 1. We propose mild galaxy-galaxy interactions as the possible physical agent for the elevation of the fraction of Hα star-forming galaxies in filaments. Our results show that filaments are the likely physical environments that are often classed as the 'intermediate' densities and that the cosmic web likely plays a major role in galaxy formation and evolution which has so far been poorly investigated.« less

  11. StructRNAfinder: an automated pipeline and web server for RNA families prediction.

    PubMed

    Arias-Carrasco, Raúl; Vásquez-Morán, Yessenia; Nakaya, Helder I; Maracaja-Coutinho, Vinicius

    2018-02-17

    The function of many noncoding RNAs (ncRNAs) depend upon their secondary structures. Over the last decades, several methodologies have been developed to predict such structures or to use them to functionally annotate RNAs into RNA families. However, to fully perform this analysis, researchers should utilize multiple tools, which require the constant parsing and processing of several intermediate files. This makes the large-scale prediction and annotation of RNAs a daunting task even to researchers with good computational or bioinformatics skills. We present an automated pipeline named StructRNAfinder that predicts and annotates RNA families in transcript or genome sequences. This single tool not only displays the sequence/structural consensus alignments for each RNA family, according to Rfam database but also provides a taxonomic overview for each assigned functional RNA. Moreover, we implemented a user-friendly web service that allows researchers to upload their own nucleotide sequences in order to perform the whole analysis. Finally, we provided a stand-alone version of StructRNAfinder to be used in large-scale projects. The tool was developed under GNU General Public License (GPLv3) and is freely available at http://structrnafinder.integrativebioinformatics.me . The main advantage of StructRNAfinder relies on the large-scale processing and integrating the data obtained by each tool and database employed along the workflow, of which several files are generated and displayed in user-friendly reports, useful for downstream analyses and data exploration.

  12. Class size as related to the use of technology, educational practices, and outcomes in Web-based nursing courses.

    PubMed

    Burruss, Nancy M; Billings, Diane M; Brownrigg, Vicki; Skiba, Diane J; Connors, Helen R

    2009-01-01

    With the expanding numbers of nursing students enrolled in Web-based courses and the shortage of faculty, class sizes are increasing. This exploratory descriptive study examined class size in relation to the use of technology and to particular educational practices and outcomes. The sample consisted of undergraduate (n = 265) and graduate (n = 863) students enrolled in fully Web-based nursing courses. The Evaluating Educational Uses of Web-based Courses in Nursing survey (Billings, D., Connors, H., Skiba, D. (2001). Benchmarking best practices in Web-based nursing courses. Advances in Nursing Science, 23, 41--52) and the Social Presence Scale (Gunawardena, C. N., Zittle, F. J. (1997). Social presence as a predictor of satisfaction within a computer-mediated conferencing environment. The American Journal of Distance Education, 11, 9-26.) were used to gather data about the study variables. Class sizes were defined as very small (1 to 10 students), small (11 to 20 students), medium (21 to 30 students), large (31 to 40 students), and very large (41 students and above). Descriptive and inferential statistics were used to analyze the data. There were significant differences by class size in students' perceptions of active participation in learning, student-faculty interaction, peer interaction, and connectedness. Some differences by class size between undergraduate and graduate students were also found, and these require further study.

  13. Stormbow: A Cloud-Based Tool for Reads Mapping and Expression Quantification in Large-Scale RNA-Seq Studies

    PubMed Central

    Zhao, Shanrong; Prenger, Kurt; Smith, Lance

    2013-01-01

    RNA-Seq is becoming a promising replacement to microarrays in transcriptome profiling and differential gene expression study. Technical improvements have decreased sequencing costs and, as a result, the size and number of RNA-Seq datasets have increased rapidly. However, the increasing volume of data from large-scale RNA-Seq studies poses a practical challenge for data analysis in a local environment. To meet this challenge, we developed Stormbow, a cloud-based software package, to process large volumes of RNA-Seq data in parallel. The performance of Stormbow has been tested by practically applying it to analyse 178 RNA-Seq samples in the cloud. In our test, it took 6 to 8 hours to process an RNA-Seq sample with 100 million reads, and the average cost was $3.50 per sample. Utilizing Amazon Web Services as the infrastructure for Stormbow allows us to easily scale up to handle large datasets with on-demand computational resources. Stormbow is a scalable, cost effective, and open-source based tool for large-scale RNA-Seq data analysis. Stormbow can be freely downloaded and can be used out of box to process Illumina RNA-Seq datasets. PMID:25937948

  14. Stormbow: A Cloud-Based Tool for Reads Mapping and Expression Quantification in Large-Scale RNA-Seq Studies.

    PubMed

    Zhao, Shanrong; Prenger, Kurt; Smith, Lance

    2013-01-01

    RNA-Seq is becoming a promising replacement to microarrays in transcriptome profiling and differential gene expression study. Technical improvements have decreased sequencing costs and, as a result, the size and number of RNA-Seq datasets have increased rapidly. However, the increasing volume of data from large-scale RNA-Seq studies poses a practical challenge for data analysis in a local environment. To meet this challenge, we developed Stormbow, a cloud-based software package, to process large volumes of RNA-Seq data in parallel. The performance of Stormbow has been tested by practically applying it to analyse 178 RNA-Seq samples in the cloud. In our test, it took 6 to 8 hours to process an RNA-Seq sample with 100 million reads, and the average cost was $3.50 per sample. Utilizing Amazon Web Services as the infrastructure for Stormbow allows us to easily scale up to handle large datasets with on-demand computational resources. Stormbow is a scalable, cost effective, and open-source based tool for large-scale RNA-Seq data analysis. Stormbow can be freely downloaded and can be used out of box to process Illumina RNA-Seq datasets.

  15. Advancing translational research with the Semantic Web.

    PubMed

    Ruttenberg, Alan; Clark, Tim; Bug, William; Samwald, Matthias; Bodenreider, Olivier; Chen, Helen; Doherty, Donald; Forsberg, Kerstin; Gao, Yong; Kashyap, Vipul; Kinoshita, June; Luciano, Joanne; Marshall, M Scott; Ogbuji, Chimezie; Rees, Jonathan; Stephens, Susie; Wong, Gwendolyn T; Wu, Elizabeth; Zaccagnini, Davide; Hongsermeier, Tonya; Neumann, Eric; Herman, Ivan; Cheung, Kei-Hoi

    2007-05-09

    A fundamental goal of the U.S. National Institute of Health (NIH) "Roadmap" is to strengthen Translational Research, defined as the movement of discoveries in basic research to application at the clinical level. A significant barrier to translational research is the lack of uniformly structured data across related biomedical domains. The Semantic Web is an extension of the current Web that enables navigation and meaningful use of digital resources by automatic processes. It is based on common formats that support aggregation and integration of data drawn from diverse sources. A variety of technologies have been built on this foundation that, together, support identifying, representing, and reasoning across a wide range of biomedical data. The Semantic Web Health Care and Life Sciences Interest Group (HCLSIG), set up within the framework of the World Wide Web Consortium, was launched to explore the application of these technologies in a variety of areas. Subgroups focus on making biomedical data available in RDF, working with biomedical ontologies, prototyping clinical decision support systems, working on drug safety and efficacy communication, and supporting disease researchers navigating and annotating the large amount of potentially relevant literature. We present a scenario that shows the value of the information environment the Semantic Web can support for aiding neuroscience researchers. We then report on several projects by members of the HCLSIG, in the process illustrating the range of Semantic Web technologies that have applications in areas of biomedicine. Semantic Web technologies present both promise and challenges. Current tools and standards are already adequate to implement components of the bench-to-bedside vision. On the other hand, these technologies are young. Gaps in standards and implementations still exist and adoption is limited by typical problems with early technology, such as the need for a critical mass of practitioners and installed base, and growing pains as the technology is scaled up. Still, the potential of interoperable knowledge sources for biomedicine, at the scale of the World Wide Web, merits continued work.

  16. Advancing translational research with the Semantic Web

    PubMed Central

    Ruttenberg, Alan; Clark, Tim; Bug, William; Samwald, Matthias; Bodenreider, Olivier; Chen, Helen; Doherty, Donald; Forsberg, Kerstin; Gao, Yong; Kashyap, Vipul; Kinoshita, June; Luciano, Joanne; Marshall, M Scott; Ogbuji, Chimezie; Rees, Jonathan; Stephens, Susie; Wong, Gwendolyn T; Wu, Elizabeth; Zaccagnini, Davide; Hongsermeier, Tonya; Neumann, Eric; Herman, Ivan; Cheung, Kei-Hoi

    2007-01-01

    Background A fundamental goal of the U.S. National Institute of Health (NIH) "Roadmap" is to strengthen Translational Research, defined as the movement of discoveries in basic research to application at the clinical level. A significant barrier to translational research is the lack of uniformly structured data across related biomedical domains. The Semantic Web is an extension of the current Web that enables navigation and meaningful use of digital resources by automatic processes. It is based on common formats that support aggregation and integration of data drawn from diverse sources. A variety of technologies have been built on this foundation that, together, support identifying, representing, and reasoning across a wide range of biomedical data. The Semantic Web Health Care and Life Sciences Interest Group (HCLSIG), set up within the framework of the World Wide Web Consortium, was launched to explore the application of these technologies in a variety of areas. Subgroups focus on making biomedical data available in RDF, working with biomedical ontologies, prototyping clinical decision support systems, working on drug safety and efficacy communication, and supporting disease researchers navigating and annotating the large amount of potentially relevant literature. Results We present a scenario that shows the value of the information environment the Semantic Web can support for aiding neuroscience researchers. We then report on several projects by members of the HCLSIG, in the process illustrating the range of Semantic Web technologies that have applications in areas of biomedicine. Conclusion Semantic Web technologies present both promise and challenges. Current tools and standards are already adequate to implement components of the bench-to-bedside vision. On the other hand, these technologies are young. Gaps in standards and implementations still exist and adoption is limited by typical problems with early technology, such as the need for a critical mass of practitioners and installed base, and growing pains as the technology is scaled up. Still, the potential of interoperable knowledge sources for biomedicine, at the scale of the World Wide Web, merits continued work. PMID:17493285

  17. Semantic Web technologies for the big data in life sciences.

    PubMed

    Wu, Hongyan; Yamaguchi, Atsuko

    2014-08-01

    The life sciences field is entering an era of big data with the breakthroughs of science and technology. More and more big data-related projects and activities are being performed in the world. Life sciences data generated by new technologies are continuing to grow in not only size but also variety and complexity, with great speed. To ensure that big data has a major influence in the life sciences, comprehensive data analysis across multiple data sources and even across disciplines is indispensable. The increasing volume of data and the heterogeneous, complex varieties of data are two principal issues mainly discussed in life science informatics. The ever-evolving next-generation Web, characterized as the Semantic Web, is an extension of the current Web, aiming to provide information for not only humans but also computers to semantically process large-scale data. The paper presents a survey of big data in life sciences, big data related projects and Semantic Web technologies. The paper introduces the main Semantic Web technologies and their current situation, and provides a detailed analysis of how Semantic Web technologies address the heterogeneous variety of life sciences big data. The paper helps to understand the role of Semantic Web technologies in the big data era and how they provide a promising solution for the big data in life sciences.

  18. Large Scale Landslide Database System Established for the Reservoirs in Southern Taiwan

    NASA Astrophysics Data System (ADS)

    Tsai, Tsai-Tsung; Tsai, Kuang-Jung; Shieh, Chjeng-Lun

    2017-04-01

    Typhoon Morakot seriously attack southern Taiwan awaken the public awareness of large scale landslide disasters. Large scale landslide disasters produce large quantity of sediment due to negative effects on the operating functions of reservoirs. In order to reduce the risk of these disasters within the study area, the establishment of a database for hazard mitigation / disaster prevention is necessary. Real time data and numerous archives of engineering data, environment information, photo, and video, will not only help people make appropriate decisions, but also bring the biggest concern for people to process and value added. The study tried to define some basic data formats / standards from collected various types of data about these reservoirs and then provide a management platform based on these formats / standards. Meanwhile, in order to satisfy the practicality and convenience, the large scale landslide disasters database system is built both provide and receive information abilities, which user can use this large scale landslide disasters database system on different type of devices. IT technology progressed extreme quick, the most modern system might be out of date anytime. In order to provide long term service, the system reserved the possibility of user define data format /standard and user define system structure. The system established by this study was based on HTML5 standard language, and use the responsive web design technology. This will make user can easily handle and develop this large scale landslide disasters database system.

  19. Web tools for large-scale 3D biological images and atlases

    PubMed Central

    2012-01-01

    Background Large-scale volumetric biomedical image data of three or more dimensions are a significant challenge for distributed browsing and visualisation. Many images now exceed 10GB which for most users is too large to handle in terms of computer RAM and network bandwidth. This is aggravated when users need to access tens or hundreds of such images from an archive. Here we solve the problem for 2D section views through archive data delivering compressed tiled images enabling users to browse through very-large volume data in the context of a standard web-browser. The system provides an interactive visualisation for grey-level and colour 3D images including multiple image layers and spatial-data overlay. Results The standard Internet Imaging Protocol (IIP) has been extended to enable arbitrary 2D sectioning of 3D data as well a multi-layered images and indexed overlays. The extended protocol is termed IIP3D and we have implemented a matching server to deliver the protocol and a series of Ajax/Javascript client codes that will run in an Internet browser. We have tested the server software on a low-cost linux-based server for image volumes up to 135GB and 64 simultaneous users. The section views are delivered with response times independent of scale and orientation. The exemplar client provided multi-layer image views with user-controlled colour-filtering and overlays. Conclusions Interactive browsing of arbitrary sections through large biomedical-image volumes is made possible by use of an extended internet protocol and efficient server-based image tiling. The tools open the possibility of enabling fast access to large image archives without the requirement of whole image download and client computers with very large memory configurations. The system was demonstrated using a range of medical and biomedical image data extending up to 135GB for a single image volume. PMID:22676296

  20. Open source database of images DEIMOS: extension for large-scale subjective image quality assessment

    NASA Astrophysics Data System (ADS)

    Vítek, Stanislav

    2014-09-01

    DEIMOS (Database of Images: Open Source) is an open-source database of images and video sequences for testing, verification and comparison of various image and/or video processing techniques such as compression, reconstruction and enhancement. This paper deals with extension of the database allowing performing large-scale web-based subjective image quality assessment. Extension implements both administrative and client interface. The proposed system is aimed mainly at mobile communication devices, taking into account advantages of HTML5 technology; it means that participants don't need to install any application and assessment could be performed using web browser. The assessment campaign administrator can select images from the large database and then apply rules defined by various test procedure recommendations. The standard test procedures may be fully customized and saved as a template. Alternatively the administrator can define a custom test, using images from the pool and other components, such as evaluating forms and ongoing questionnaires. Image sequence is delivered to the online client, e.g. smartphone or tablet, as a fully automated assessment sequence or viewer can decide on timing of the assessment if required. Environmental data and viewing conditions (e.g. illumination, vibrations, GPS coordinates, etc.), may be collected and subsequently analyzed.

  1. Benthic primary producers are key to sustain the Wadden Sea food web: stable carbon isotope analysis at landscape scale.

    PubMed

    Christianen, M J A; Middelburg, J J; Holthuijsen, S J; Jouta, J; Compton, T J; van der Heide, T; Piersma, T; Sinninghe Damsté, J S; van der Veer, H W; Schouten, S; Olff, H

    2017-06-01

    Coastal food webs can be supported by local benthic or pelagic primary producers and by the import of organic matter. Distinguishing between these energy sources is essential for our understanding of ecosystem functioning. However, the relative contribution of these components to the food web at the landscape scale is often unclear, as many studies lack good taxonomic and spatial resolution across large areas. Here, using stable carbon isotopes, we report on the primary carbon sources for consumers and their spatial variability across one of the world's largest intertidal ecosystems (Dutch Wadden Sea; 1460 km 2 intertidal surface area), at an exceptionally high taxonomic (178 species) and spatial resolution (9,165 samples from 839 locations). The absence of overlap in δ 13 C values between consumers and terrestrial organic matter suggests that benthic and pelagic producers dominate carbon input into this food web. In combination with the consistent enrichment of benthic primary producers (δ 13 C -16.3‰) relative to pelagic primary producers (δ 13 C -18.8) across the landscape, this allowed the use of a two-food-source isotope-mixing model. This spatially resolved modelling revealed that benthic primary producers (microphytobenthos) are the most important energy source for the majority of consumers at higher trophic levels (worms, molluscs, crustaceans, fish, and birds), and thus to the whole food web. In addition, we found large spatial heterogeneity in the δ 13 C values of benthic primary producers (δ 13 C -19.2 to -11.5‰) and primary consumers (δ 13 C -25.5 to -9.9‰), emphasizing the need for spatially explicit sampling of benthic and pelagic primary producers in coastal ecosystems. Our findings have important implications for our understanding of the functioning of ecological networks and for the management of coastal ecosystems. © 2017 by the Ecological Society of America.

  2. Social Facilitation Effects by Pedagogical Conversational Agent: Lexical Network Analysis in an Online Explanation Task

    ERIC Educational Resources Information Center

    Hayashi, Yugo

    2015-01-01

    The present study investigates web-based learning activities of undergraduate students who generate explanations about a key concept taught in a large-scale classroom. The present study used an online system with Pedagogical Conversational Agent (PCA), asked to explain about the key concept from different points and provided suggestions and…

  3. The Effect of Gestational Age on Symptom Severity in Children with Autism Spectrum Disorder

    ERIC Educational Resources Information Center

    Movsas, Tammy Z.; Paneth, Nigel

    2012-01-01

    Between 2006 and 2010, two research-validated instruments, Social Communication Questionnaire (SCQ) and Social Responsiveness Scale (SRS) were filled out online by 4,188 mothers of Autism Spectrum Disorder (ASD) children, aged 4-21, as part of voluntary parental participation in a large web-based registry. Univariate and multivariate linear…

  4. Investigating the Effect of Academic Procrastination on the Frequency and Variety of Academic Misconduct: A Panel Study

    ERIC Educational Resources Information Center

    Patrzek, Justine; Sattler, Sebastian; van Veen, Floris; Grunschel, Carola; Fries, Stefan

    2015-01-01

    In prior studies, academic procrastination has been discussed as an influencing factor of academic misconduct. However, empirical studies were conducted solely cross-sectionally and investigated only a few forms of academic misconduct. This large scale web-based study examined the responses of between 1359 and 2207 participants from different…

  5. Multisite Randomized Controlled Trial Examining Intelligent Tutoring of Structure Strategy for Fifth-Grade Readers

    ERIC Educational Resources Information Center

    Wijekumar, Kausalai; Meyer, Bonnie J. F.; Lei, Pui-Wa; Lin, Yu-Chu; Johnson, Lori A.; Spielvogel, James A.; Shurmatz, Kathryn M.; Ray, Melissa; Cook, Michael

    2014-01-01

    This article reports on a large scale randomized controlled trial to study the efficacy of a web-based intelligent tutoring system for the structure strategy designed to improve content area reading comprehension. The research was conducted with 128 fifth-grade classrooms within 12 school districts in rural and suburban settings. Classrooms within…

  6. Adult Siblings of Individuals with Down Syndrome versus with Autism: Findings from a Large-Scale US Survey

    ERIC Educational Resources Information Center

    Hodapp, R. M.; Urbano, R. C.

    2007-01-01

    Background: As adults with Down syndrome live increasingly longer lives, their adult siblings will most likely assume caregiving responsibilities. Yet little is known about either the sibling relationship or the general functioning of these adult siblings. Using a national, web-based survey, this study compared adult siblings of individuals with…

  7. A Unique Design for High-Impact Safety and Awareness Training

    ERIC Educational Resources Information Center

    Calandra, Brendan; Harmon, Stephen W.

    2012-01-01

    The authors were asked to design and develop a large-scale, web-based learning environment that would effectively assist international aid workers in conducting their daily tasks in the field, at home and in the office in a safer and more secure fashion. The design, development and dissemination of the training needed to be done quickly,…

  8. E-Mentoring for Social Equity: Review of Research to Inform Program Development

    ERIC Educational Resources Information Center

    Single, Peg Boyle; Single, Richard M.

    2005-01-01

    The advent of user-friendly email programs and web browsers created possibilities for widespread use of e-mentoring programs. In this review of the research, we presented the history of e-mentoring programs and defined e-mentoring and structured e-mentoring programs, focusing on large-scale e-mentoring programs that addressed issues of social…

  9. Understanding the Relationships between Interest in Online Math Games and Academic Performance

    ERIC Educational Resources Information Center

    Zhang, M.

    2015-01-01

    Although the Internet is widely used by students in both formal and informal environments, little is known about how and where youth spend their time online. Using Internet search and Web analytics data, this study discovered a large-scale phenomenon associated with the poor performance of elementary school students in the USA that has been…

  10. Developing Server-Side Infrastructure for Large-Scale E-Learning of Web Technology

    ERIC Educational Resources Information Center

    Simpkins, Neil

    2010-01-01

    The growth of E-business has made experience in server-side technology an increasingly important area for educators. Server-side skills are in increasing demand and recognised to be of relatively greater value than comparable client-side aspects (Ehie, 2002). In response to this, many educational organisations have developed E-business courses,…

  11. Taking Open Innovation to the Molecular Level - Strengths and Limitations.

    PubMed

    Zdrazil, Barbara; Blomberg, Niklas; Ecker, Gerhard F

    2012-08-01

    The ever-growing availability of large-scale open data and its maturation is having a significant impact on industrial drug-discovery, as well as on academic and non-profit research. As industry is changing to an 'open innovation' business concept, precompetitive initiatives and strong public-private partnerships including academic research cooperation partners are gaining more and more importance. Now, the bioinformatics and cheminformatics communities are seeking for web tools which allow the integration of this large volume of life science datasets available in the public domain. Such a data exploitation tool would ideally be able to answer complex biological questions by formulating only one search query. In this short review/perspective, we outline the use of semantic web approaches for data and knowledge integration. Further, we discuss strengths and current limitations of public available data retrieval tools and integrated platforms.

  12. TCGA4U: A Web-Based Genomic Analysis Platform To Explore And Mine TCGA Genomic Data For Translational Research.

    PubMed

    Huang, Zhenzhen; Duan, Huilong; Li, Haomin

    2015-01-01

    Large-scale human cancer genomics projects, such as TCGA, generated large genomics data for further study. Exploring and mining these data to obtain meaningful analysis results can help researchers find potential genomics alterations that intervene the development and metastasis of tumors. We developed a web-based gene analysis platform, named TCGA4U, which used statistics methods and models to help translational investigators explore, mine and visualize human cancer genomic characteristic information from the TCGA datasets. Furthermore, through Gene Ontology (GO) annotation and clinical data integration, the genomic data were transformed into biological process, molecular function, cellular component and survival curves to help researchers identify potential driver genes. Clinical researchers without expertise in data analysis will benefit from such a user-friendly genomic analysis platform.

  13. GEOMetaCuration: a web-based application for accurate manual curation of Gene Expression Omnibus metadata

    PubMed Central

    Li, Zhao; Li, Jin; Yu, Peng

    2018-01-01

    Abstract Metadata curation has become increasingly important for biological discovery and biomedical research because a large amount of heterogeneous biological data is currently freely available. To facilitate efficient metadata curation, we developed an easy-to-use web-based curation application, GEOMetaCuration, for curating the metadata of Gene Expression Omnibus datasets. It can eliminate mechanical operations that consume precious curation time and can help coordinate curation efforts among multiple curators. It improves the curation process by introducing various features that are critical to metadata curation, such as a back-end curation management system and a curator-friendly front-end. The application is based on a commonly used web development framework of Python/Django and is open-sourced under the GNU General Public License V3. GEOMetaCuration is expected to benefit the biocuration community and to contribute to computational generation of biological insights using large-scale biological data. An example use case can be found at the demo website: http://geometacuration.yubiolab.org. Database URL: https://bitbucket.com/yubiolab/GEOMetaCuration PMID:29688376

  14. Filaments from the galaxy distribution and from the velocity field in the local universe

    NASA Astrophysics Data System (ADS)

    Libeskind, Noam I.; Tempel, Elmo; Hoffman, Yehuda; Tully, R. Brent; Courtois, Hélène

    2015-10-01

    The cosmic web that characterizes the large-scale structure of the Universe can be quantified by a variety of methods. For example, large redshift surveys can be used in combination with point process algorithms to extract long curvilinear filaments in the galaxy distribution. Alternatively, given a full 3D reconstruction of the velocity field, kinematic techniques can be used to decompose the web into voids, sheets, filaments and knots. In this Letter, we look at how two such algorithms - the Bisous model and the velocity shear web - compare with each other in the local Universe (within 100 Mpc), finding good agreement. This is both remarkable and comforting, given that the two methods are radically different in ideology and applied to completely independent and different data sets. Unsurprisingly, the methods are in better agreement when applied to unbiased and complete data sets, like cosmological simulations, than when applied to observational samples. We conclude that more observational data is needed to improve on these methods, but that both methods are most likely properly tracing the underlying distribution of matter in the Universe.

  15. SurveyWiz and factorWiz: JavaScript Web pages that make HTML forms for research on the Internet.

    PubMed

    Birnbaum, M H

    2000-05-01

    SurveyWiz and factorWiz are Web pages that act as wizards to create HTML forms that enable one to collect data via the Web. SurveyWiz allows the user to enter survey questions or personality test items with a mixture of text boxes and scales of radio buttons. One can add demographic questions of age, sex, education, and nationality with the push of a button. FactorWiz creates the HTML for within-subjects, two-factor designs as large as 9 x 9, or higher order factorial designs up to 81 cells. The user enters levels of the row and column factors, which can be text, images, or other multimedia. FactorWiz generates the stimulus combinations, randomizes their order, and creates the page. In both programs HTML is displayed in a window, and the user copies it to a text editor to save it. When uploaded to a Web server and supported by a CGI script, the created Web pages allow data to be collected, coded, and saved on the server. These programs are intended to assist researchers and students in quickly creating studies that can be administered via the Web.

  16. TreeVector: scalable, interactive, phylogenetic trees for the web.

    PubMed

    Pethica, Ralph; Barker, Gary; Kovacs, Tim; Gough, Julian

    2010-01-28

    Phylogenetic trees are complex data forms that need to be graphically displayed to be human-readable. Traditional techniques of plotting phylogenetic trees focus on rendering a single static image, but increases in the production of biological data and large-scale analyses demand scalable, browsable, and interactive trees. We introduce TreeVector, a Scalable Vector Graphics-and Java-based method that allows trees to be integrated and viewed seamlessly in standard web browsers with no extra software required, and can be modified and linked using standard web technologies. There are now many bioinformatics servers and databases with a range of dynamic processes and updates to cope with the increasing volume of data. TreeVector is designed as a framework to integrate with these processes and produce user-customized phylogenies automatically. We also address the strengths of phylogenetic trees as part of a linked-in browsing process rather than an end graphic for print. TreeVector is fast and easy to use and is available to download precompiled, but is also open source. It can also be run from the web server listed below or the user's own web server. It has already been deployed on two recognized and widely used database Web sites.

  17. The Australian Longitudinal Study on Women's Health: Using Focus Groups to Inform Recruitment

    PubMed Central

    2016-01-01

    Background Recruitment and retention of participants to large-scale, longitudinal studies can be a challenge, particularly when trying to target young women. Qualitative inquiries with members of the target population can prove valuable in assisting with the development of effective recruiting techniques. Researchers in the current study made use of focus group methodology to identify how to encourage young women aged 18-23 to participate in a national cohort online survey. Objective Our objectives were to gain insight into how to encourage young women to participate in a large-scale, longitudinal health survey, as well as to evaluate the survey instrument and mode of administration. Methods The Australian Longitudinal Study on Women’s Health used focus group methodology to learn how to encourage young women to participate in a large-scale, longitudinal Web-based health survey and to evaluate the survey instrument and mode of administration. Nineteen groups, involving 75 women aged 18-23 years, were held in remote, regional, and urban areas of New South Wales and Queensland. Results Focus groups were held in 2 stages, with discussions lasting from 19 minutes to over 1 hour. The focus groups allowed concord to be reached regarding survey promotion using social media, why personal information was needed, strategies to ensure confidentiality, how best to ask sensitive questions, and survey design for ease of completion. Recruitment into the focus groups proved difficult: the groups varied in size between 1 and 8 participants, with the majority conducted with 2 participants. Conclusions Intense recruitment efforts and variation in final focus group numbers highlights the “hard to reach” character of young women. However, the benefits of conducting focus group discussions as a preparatory stage to the recruitment of a large cohort for a longitudinal Web-based health survey were upheld. PMID:26902160

  18. An interactive web application for the dissemination of human systems immunology data.

    PubMed

    Speake, Cate; Presnell, Scott; Domico, Kelly; Zeitner, Brad; Bjork, Anna; Anderson, David; Mason, Michael J; Whalen, Elizabeth; Vargas, Olivia; Popov, Dimitry; Rinchai, Darawan; Jourde-Chiche, Noemie; Chiche, Laurent; Quinn, Charlie; Chaussabel, Damien

    2015-06-19

    Systems immunology approaches have proven invaluable in translational research settings. The current rate at which large-scale datasets are generated presents unique challenges and opportunities. Mining aggregates of these datasets could accelerate the pace of discovery, but new solutions are needed to integrate the heterogeneous data types with the contextual information that is necessary for interpretation. In addition, enabling tools and technologies facilitating investigators' interaction with large-scale datasets must be developed in order to promote insight and foster knowledge discovery. State of the art application programming was employed to develop an interactive web application for browsing and visualizing large and complex datasets. A collection of human immune transcriptome datasets were loaded alongside contextual information about the samples. We provide a resource enabling interactive query and navigation of transcriptome datasets relevant to human immunology research. Detailed information about studies and samples are displayed dynamically; if desired the associated data can be downloaded. Custom interactive visualizations of the data can be shared via email or social media. This application can be used to browse context-rich systems-scale data within and across systems immunology studies. This resource is publicly available online at [Gene Expression Browser Landing Page ( https://gxb.benaroyaresearch.org/dm3/landing.gsp )]. The source code is also available openly [Gene Expression Browser Source Code ( https://github.com/BenaroyaResearch/gxbrowser )]. We have developed a data browsing and visualization application capable of navigating increasingly large and complex datasets generated in the context of immunological studies. This intuitive tool ensures that, whether taken individually or as a whole, such datasets generated at great effort and expense remain interpretable and a ready source of insight for years to come.

  19. Catalytic, conductive, and transparent platinum nanofiber webs for FTO-free dye-sensitized solar cells.

    PubMed

    Kim, Jongwook; Kang, Jonghyun; Jeong, Uiyoung; Kim, Heesuk; Lee, Hyunjung

    2013-04-24

    We report a multifunctional platinium nanofiber (PtNF) web that can act as a catalyst layer in dye-sensitized solar cell (DSSC) to simultaneously function as a transparent counter electrode (CE), i.e., without the presence of an indium-doped tin oxide (ITO) or fluorine-doped tin oxide (FTO) glass. This PtNF web can be easily produced by electrospinning, which is highly cost-effective and suitable for large-area industrial-scale production. Electrospun PtNFs are straight and have a length of a few micrometers, with a common diameter of 40-70 nm. Each nanofiber is composed of compact, crystalline Pt grains and they are well-fused and highly interconnected, which should be helpful to provide an efficient conductive network for free electron transport and a large surface area for electrocatalytic behavior. A PtNF web is served as a counter electrode in DSSC and the photovoltaic performance increases up to a power efficiency of 6.0%. It reaches up to 83% of that in a conventional DSSC using a Pt-coated FTO glass as a counter electrode. Newly designed DSSCs containing PtNF webs display highly stable photoelectric conversion efficiencies, and excellent catalytic, conductive, and transparent properties, as well as long-term stability. Also, while the DSSC function is retained, the fabrication cost is reduced by eliminating the transparent conducting layer on the counter electrode. The presented method of fabricating DSSCs based on a PtNF web can be extended to other electrocatalytic optoelectronic devices that combine superior catalytic activity with high conductivity and transparency.

  20. Why do Scale-Free Networks Emerge in Nature? From Gradient Networks to Transport Efficiency

    NASA Astrophysics Data System (ADS)

    Toroczkai, Zoltan

    2004-03-01

    It has recently been recognized [1,2,3] that a large number of complex networks are scale-free (having a power-law degree distribution). Examples include citation networks [4], the internet [5], the world-wide-web [6], cellular metabolic networks [7], protein interaction networks [8], the sex-web [9] and alliance networks in the U.S. biotechnology industry [10]. The existence of scale-free networks in such diverse systems suggests that there is a simple underlying common reason for their development. Here, we propose that scale-free networks emerge because they ensure efficient transport of some entity. We show that for flows generated by gradients of a scalar "potential'' distributed on a network, non scale-free networks, e.g., random graphs [11], will become maximally congested, while scale-free networks will ensure efficient transport in the large network size limit. [1] R. Albert and A.-L. Barabási, Rev.Mod.Phys. 74, 47 (2002). [2] M.E.J. Newman, SIAM Rev. 45, 167 (2003). [3] S.N. Dorogovtsev and J.F.F. Mendes, Evolution of Networks: From Biological Nets to the Internet and WWW, Oxford Univ. Press, Oxford, 2003. [4] S. Redner, Eur.Phys.J. B, 4, 131 (1998). [5] M. Faloutsos, P. Faloutsos and C. Faloutsos Comp.Comm.Rev. 29, 251 (1999). [6] R. Albert, H. Jeong, and A.L. Barabási, Nature 401, 130 (1999). [7] H. Jeong et.al. Nature 407, 651 (2000). [8] H. Jeong, S. Mason, A.-L. Barabási and Z. N. Oltvai, Nature 411, 41 (2001). [9] F. Liljeros et. al. Nature 411 907 (2000). [10] W. W. Powell, D. R. White, K. W. Koput and J. Owen-Smith Am.J.Soc. in press. [11] B. Bollobás, Random Graphs, Second Edition, Cambridge University Press (2001).

  1. Interactive Visualization of Large-Scale Hydrological Data using Emerging Technologies in Web Systems and Parallel Programming

    NASA Astrophysics Data System (ADS)

    Demir, I.; Krajewski, W. F.

    2013-12-01

    As geoscientists are confronted with increasingly massive datasets from environmental observations to simulations, one of the biggest challenges is having the right tools to gain scientific insight from the data and communicate the understanding to stakeholders. Recent developments in web technologies make it easy to manage, visualize and share large data sets with general public. Novel visualization techniques and dynamic user interfaces allow users to interact with data, and modify the parameters to create custom views of the data to gain insight from simulations and environmental observations. This requires developing new data models and intelligent knowledge discovery techniques to explore and extract information from complex computational simulations or large data repositories. Scientific visualization will be an increasingly important component to build comprehensive environmental information platforms. This presentation provides an overview of the trends and challenges in the field of scientific visualization, and demonstrates information visualization and communication tools developed within the light of these challenges.

  2. Scaling Up: From Web-Enhanced Courses to a Web-Enhanced Curriculum

    ERIC Educational Resources Information Center

    Wood, Robert E.

    2004-01-01

    In the past decade, the most important technological innovation in higher education has been the enhancement of academic courses with Web-based information and tools. Given the success and popularity of numerous Web-based innovations, the author and two colleagues wondered whether the benefits of technology use could be scaled up from the course…

  3. Spider web-inspired acoustic metamaterials

    NASA Astrophysics Data System (ADS)

    Miniaci, Marco; Krushynska, Anastasiia; Movchan, Alexander B.; Bosia, Federico; Pugno, Nicola M.

    2016-08-01

    Spider silk is a remarkable example of bio-material with superior mechanical characteristics. Its multilevel structural organization of dragline and viscid silk leads to unusual and tunable properties, extensively studied from a quasi-static point of view. In this study, inspired by the Nephila spider orb web architecture, we propose a design for mechanical metamaterials based on its periodic repetition. We demonstrate that spider-web metamaterial structure plays an important role in the dynamic response and wave attenuation mechanisms. The capability of the resulting structure to inhibit elastic wave propagation in sub-wavelength frequency ranges is assessed, and parametric studies are performed to derive optimal configurations and constituent mechanical properties. The results show promise for the design of innovative lightweight structures for tunable vibration damping and impact protection, or the protection of large scale infrastructure such as suspended bridges.

  4. 3Drefine: an interactive web server for efficient protein structure refinement

    PubMed Central

    Bhattacharya, Debswapna; Nowotny, Jackson; Cao, Renzhi; Cheng, Jianlin

    2016-01-01

    3Drefine is an interactive web server for consistent and computationally efficient protein structure refinement with the capability to perform web-based statistical and visual analysis. The 3Drefine refinement protocol utilizes iterative optimization of hydrogen bonding network combined with atomic-level energy minimization on the optimized model using a composite physics and knowledge-based force fields for efficient protein structure refinement. The method has been extensively evaluated on blind CASP experiments as well as on large-scale and diverse benchmark datasets and exhibits consistent improvement over the initial structure in both global and local structural quality measures. The 3Drefine web server allows for convenient protein structure refinement through a text or file input submission, email notification, provided example submission and is freely available without any registration requirement. The server also provides comprehensive analysis of submissions through various energy and statistical feedback and interactive visualization of multiple refined models through the JSmol applet that is equipped with numerous protein model analysis tools. The web server has been extensively tested and used by many users. As a result, the 3Drefine web server conveniently provides a useful tool easily accessible to the community. The 3Drefine web server has been made publicly available at the URL: http://sysbio.rnet.missouri.edu/3Drefine/. PMID:27131371

  5. Technological Networks

    NASA Astrophysics Data System (ADS)

    Mitra, Bivas

    The study of networks in the form of mathematical graph theory is one of the fundamental pillars of discrete mathematics. However, recent years have witnessed a substantial new movement in network research. The focus of the research is shifting away from the analysis of small graphs and the properties of individual vertices or edges to consideration of statistical properties of large scale networks. This new approach has been driven largely by the availability of technological networks like the Internet [12], World Wide Web network [2], etc. that allow us to gather and analyze data on a scale far larger than previously possible. At the same time, technological networks have evolved as a socio-technological system, as the concepts of social systems that are based on self-organization theory have become unified in technological networks [13]. In today’s society, we have a simple and universal access to great amounts of information and services. These information services are based upon the infrastructure of the Internet and the World Wide Web. The Internet is the system composed of ‘computers’ connected by cables or some other form of physical connections. Over this physical network, it is possible to exchange e-mails, transfer files, etc. On the other hand, the World Wide Web (commonly shortened to the Web) is a system of interlinked hypertext documents accessed via the Internet where nodes represent web pages and links represent hyperlinks between the pages. Peer-to-peer (P2P) networks [26] also have recently become a popular medium through which huge amounts of data can be shared. P2P file sharing systems, where files are searched and downloaded among peers without the help of central servers, have emerged as a major component of Internet traffic. An important advantage in P2P networks is that all clients provide resources, including bandwidth, storage space, and computing power. In this chapter, we discuss these technological networks in detail. The review is organized as follows. Section 2 presents an introduction to the Internet and different protocols related to it. This section also specifies the socio-technological properties of the Internet, like scale invariance, the small-world property, network resilience, etc. Section 3 describes the P2P networks, their categorization, and other related issues like search, stability, etc. Section 4 concludes the chapter.

  6. SIFTER search: a web server for accurate phylogeny-based protein function prediction

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Sahraeian, Sayed M.; Luo, Kevin R.; Brenner, Steven E.

    We are awash in proteins discovered through high-throughput sequencing projects. As only a minuscule fraction of these have been experimentally characterized, computational methods are widely used for automated annotation. Here, we introduce a user-friendly web interface for accurate protein function prediction using the SIFTER algorithm. SIFTER is a state-of-the-art sequence-based gene molecular function prediction algorithm that uses a statistical model of function evolution to incorporate annotations throughout the phylogenetic tree. Due to the resources needed by the SIFTER algorithm, running SIFTER locally is not trivial for most users, especially for large-scale problems. The SIFTER web server thus provides access tomore » precomputed predictions on 16 863 537 proteins from 232 403 species. Users can explore SIFTER predictions with queries for proteins, species, functions, and homologs of sequences not in the precomputed prediction set. Lastly, the SIFTER web server is accessible at http://sifter.berkeley.edu/ and the source code can be downloaded.« less

  7. SIFTER search: a web server for accurate phylogeny-based protein function prediction

    DOE PAGES

    Sahraeian, Sayed M.; Luo, Kevin R.; Brenner, Steven E.

    2015-05-15

    We are awash in proteins discovered through high-throughput sequencing projects. As only a minuscule fraction of these have been experimentally characterized, computational methods are widely used for automated annotation. Here, we introduce a user-friendly web interface for accurate protein function prediction using the SIFTER algorithm. SIFTER is a state-of-the-art sequence-based gene molecular function prediction algorithm that uses a statistical model of function evolution to incorporate annotations throughout the phylogenetic tree. Due to the resources needed by the SIFTER algorithm, running SIFTER locally is not trivial for most users, especially for large-scale problems. The SIFTER web server thus provides access tomore » precomputed predictions on 16 863 537 proteins from 232 403 species. Users can explore SIFTER predictions with queries for proteins, species, functions, and homologs of sequences not in the precomputed prediction set. Lastly, the SIFTER web server is accessible at http://sifter.berkeley.edu/ and the source code can be downloaded.« less

  8. G2S: a web-service for annotating genomic variants on 3D protein structures.

    PubMed

    Wang, Juexin; Sheridan, Robert; Sumer, S Onur; Schultz, Nikolaus; Xu, Dong; Gao, Jianjiong

    2018-06-01

    Accurately mapping and annotating genomic locations on 3D protein structures is a key step in structure-based analysis of genomic variants detected by recent large-scale sequencing efforts. There are several mapping resources currently available, but none of them provides a web API (Application Programming Interface) that supports programmatic access. We present G2S, a real-time web API that provides automated mapping of genomic variants on 3D protein structures. G2S can align genomic locations of variants, protein locations, or protein sequences to protein structures and retrieve the mapped residues from structures. G2S API uses REST-inspired design and it can be used by various clients such as web browsers, command terminals, programming languages and other bioinformatics tools for bringing 3D structures into genomic variant analysis. The webserver and source codes are freely available at https://g2s.genomenexus.org. g2s@genomenexus.org. Supplementary data are available at Bioinformatics online.

  9. Climate change impacts in multispecies systems: drought alters food web size structure in a field experiment

    PubMed Central

    Woodward, Guy; Brown, Lee E.; Edwards, Francois K.; Hudson, Lawrence N.; Milner, Alexander M.; Reuman, Daniel C.; Ledger, Mark E.

    2012-01-01

    Experimental data from intergenerational field manipulations of entire food webs are scarce, yet such approaches are essential for gauging impacts of environmental change in natural systems. We imposed 2 years of intermittent drought on stream channels in a replicated field trial, to measure food web responses to simulated climate change. Drought triggered widespread losses of species and links, with larger taxa and those that were rare for their size, many of which were predatory, being especially vulnerable. Many network properties, including size–scaling relationships within food chains, changed in response to drought. Other properties, such as connectance, were unaffected. These findings highlight the need for detailed experimental data from different organizational levels, from pairwise links to the entire food web. The loss of not only large species, but also those that were rare for their size, provides a newly refined way to gauge likely impacts that may be applied more generally to other systems and/or impacts. PMID:23007087

  10. Mangrove clearing impacts on macrofaunal assemblages and benthic food webs in a tropical estuary.

    PubMed

    Bernardino, Angelo Fraga; Gomes, Luiz Eduardo de Oliveira; Hadlich, Heliatrice Louise; Andrades, Ryan; Correa, Lucas Barreto

    2018-01-01

    Despite over 21,000ha of mangrove forests being removed per year in Brazil, ecological changes following mangrove deforestation have been overlooked. Here we evaluated changes in benthic macrofaunal assemblages and food-webs at a mangrove removal and natural sites in a tropical estuary in Eastern Brazil. The impacted site had coarser sediment particle sizes suggesting significant changes in sedimentation processes after forest clearing. Spatial differences in macrofaunal abundance, biomass and diversity were not directly associated with the removal of mangrove forests, supporting recolonization of impacted areas by estuarine fauna. However, benthic assemblage composition, infaunal δ 13 C signatures and food-web diversity markedly differed at the impacted site being strongly related to sedimentary changes. The loss of infaunal trophic diversity that followed mangrove removal suggests that large-scale forest clearing may impact estuarine food webs, with potential consequences to nearby coastal ecosystems given the high clearing rate of mangrove forests in Brazil. Copyright © 2017 Elsevier Ltd. All rights reserved.

  11. A Data Management System Integrating Web-based Training and Randomized Trials: Requirements, Experiences and Recommendations.

    PubMed

    Muroff, Jordana; Amodeo, Maryann; Larson, Mary Jo; Carey, Margaret; Loftin, Ralph D

    2011-01-01

    This article describes a data management system (DMS) developed to support a large-scale randomized study of an innovative web-course that was designed to improve substance abuse counselors' knowledge and skills in applying a substance abuse treatment method (i.e., cognitive behavioral therapy; CBT). The randomized trial compared the performance of web-course-trained participants (intervention group) and printed-manual-trained participants (comparison group) to determine the effectiveness of the web-course in teaching CBT skills. A single DMS was needed to support all aspects of the study: web-course delivery and management, as well as randomized trial management. The authors briefly reviewed several other systems that were described as built either to handle randomized trials or to deliver and evaluate web-based training. However it was clear that these systems fell short of meeting our needs for simultaneous, coordinated management of the web-course and the randomized trial. New England Research Institute's (NERI) proprietary Advanced Data Entry and Protocol Tracking (ADEPT) system was coupled with the web-programmed course and customized for our purposes. This article highlights the requirements for a DMS that operates at the intersection of web-based course management systems and randomized clinical trial systems, and the extent to which the coupled, customized ADEPT satisfied those requirements. Recommendations are included for institutions and individuals considering conducting randomized trials and web-based training programs, and seeking a DMS that can meet similar requirements.

  12. Exploring network operations for data and information networks

    NASA Astrophysics Data System (ADS)

    Yao, Bing; Su, Jing; Ma, Fei; Wang, Xiaomin; Zhao, Xiyang; Yao, Ming

    2017-01-01

    Barabási and Albert, in 1999, formulated scale-free models based on some real networks: World-Wide Web, Internet, metabolic and protein networks, language or sexual networks. Scale-free networks not only appear around us, but also have high qualities in the world. As known, high quality information networks can transfer feasibly and efficiently data, clearly, their topological structures are very important for data safety. We build up network operations for constructing large scale of dynamic networks from smaller scale of network models having good property and high quality. We focus on the simplest operators to formulate complex operations, and are interesting on the closeness of operations to desired network properties.

  13. FQC Dashboard: integrates FastQC results into a web-based, interactive, and extensible FASTQ quality control tool

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Brown, Joseph; Pirrung, Meg; McCue, Lee Ann

    FQC is software that facilitates large-scale quality control of FASTQ files by carrying out a QC protocol, parsing results, and aggregating quality metrics within and across experiments into an interactive dashboard. The dashboard utilizes human-readable configuration files to manipulate the pages and tabs, and is extensible with CSV data.

  14. E-Learning in Engineering Education: A Theoretical and Empirical Study of the Algerian Higher Education Institution

    ERIC Educational Resources Information Center

    Benchicou, Soraya; Aichouni, Mohamed; Nehari, Driss

    2010-01-01

    Technology-mediated education or e-learning is growing globally both in scale and delivery capacity due to the large diffusion of the ubiquitous information and communication technologies (ICT) in general and the web technologies in particular. This statement has not yet been fully supported by research, especially in developing countries such as…

  15. Global change in the trophic functioning of marine food webs

    PubMed Central

    Gascuel, Didier; Colléter, Mathieu; Palomares, Maria L. D.; Du Pontavice, Hubert; Pauly, Daniel; Cheung, William W. L.

    2017-01-01

    The development of fisheries in the oceans, and other human drivers such as climate warming, have led to changes in species abundance, assemblages, trophic interactions, and ultimately in the functioning of marine food webs. Here, using a trophodynamic approach and global databases of catches and life history traits of marine species, we tested the hypothesis that anthropogenic ecological impacts may have led to changes in the global parameters defining the transfers of biomass within the food web. First, we developed two indicators to assess such changes: the Time Cumulated Indicator (TCI) measuring the residence time of biomass within the food web, and the Efficiency Cumulated Indicator (ECI) quantifying the fraction of secondary production reaching the top of the trophic chain. Then, we assessed, at the large marine ecosystem scale, the worldwide change of these two indicators over the 1950–2010 time-periods. Global trends were identified and cluster analyses were used to characterize the variability of trends between ecosystems. Results showed that the most common pattern over the study period is a global decrease in TCI, while the ECI indicator tends to increase. Thus, changes in species assemblages would induce faster and apparently more efficient biomass transfers in marine food webs. Results also suggested that the main driver of change over that period had been the large increase in fishing pressure. The largest changes occurred in ecosystems where ‘fishing down the marine food web’ are most intensive. PMID:28800358

  16. Managing Large Scale Project Analysis Teams through a Web Accessible Database

    NASA Technical Reports Server (NTRS)

    O'Neil, Daniel A.

    2008-01-01

    Large scale space programs analyze thousands of requirements while mitigating safety, performance, schedule, and cost risks. These efforts involve a variety of roles with interdependent use cases and goals. For example, study managers and facilitators identify ground-rules and assumptions for a collection of studies required for a program or project milestone. Task leaders derive product requirements from the ground rules and assumptions and describe activities to produce needed analytical products. Disciplined specialists produce the specified products and load results into a file management system. Organizational and project managers provide the personnel and funds to conduct the tasks. Each role has responsibilities to establish information linkages and provide status reports to management. Projects conduct design and analysis cycles to refine designs to meet the requirements and implement risk mitigation plans. At the program level, integrated design and analysis cycles studies are conducted to eliminate every 'to-be-determined' and develop plans to mitigate every risk. At the agency level, strategic studies analyze different approaches to exploration architectures and campaigns. This paper describes a web-accessible database developed by NASA to coordinate and manage tasks at three organizational levels. Other topics in this paper cover integration technologies and techniques for process modeling and enterprise architectures.

  17. The Cell Collective: Toward an open and collaborative approach to systems biology

    PubMed Central

    2012-01-01

    Background Despite decades of new discoveries in biomedical research, the overwhelming complexity of cells has been a significant barrier to a fundamental understanding of how cells work as a whole. As such, the holistic study of biochemical pathways requires computer modeling. Due to the complexity of cells, it is not feasible for one person or group to model the cell in its entirety. Results The Cell Collective is a platform that allows the world-wide scientific community to create these models collectively. Its interface enables users to build and use models without specifying any mathematical equations or computer code - addressing one of the major hurdles with computational research. In addition, this platform allows scientists to simulate and analyze the models in real-time on the web, including the ability to simulate loss/gain of function and test what-if scenarios in real time. Conclusions The Cell Collective is a web-based platform that enables laboratory scientists from across the globe to collaboratively build large-scale models of various biological processes, and simulate/analyze them in real time. In this manuscript, we show examples of its application to a large-scale model of signal transduction. PMID:22871178

  18. The Turkish Version of Web-Based Learning Platform Evaluation Scale: Reliability and Validity Study

    ERIC Educational Resources Information Center

    Dag, Funda

    2016-01-01

    The purpose of this study is to determine the language equivalence and the validity and reliability of the Turkish version of the "Web-Based Learning Platform Evaluation Scale" ("Web Tabanli Ögrenme Ortami Degerlendirme Ölçegi" [WTÖODÖ]) used in the selection and evaluation of web-based learning environments. Within this scope,…

  19. Structure Discovery in Large Semantic Graphs Using Extant Ontological Scaling and Descriptive Statistics

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    al-Saffar, Sinan; Joslyn, Cliff A.; Chappell, Alan R.

    As semantic datasets grow to be very large and divergent, there is a need to identify and exploit their inherent semantic structure for discovery and optimization. Towards that end, we present here a novel methodology to identify the semantic structures inherent in an arbitrary semantic graph dataset. We first present the concept of an extant ontology as a statistical description of the semantic relations present amongst the typed entities modeled in the graph. This serves as a model of the underlying semantic structure to aid in discovery and visualization. We then describe a method of ontological scaling in which themore » ontology is employed as a hierarchical scaling filter to infer different resolution levels at which the graph structures are to be viewed or analyzed. We illustrate these methods on three large and publicly available semantic datasets containing more than one billion edges each. Keywords-Semantic Web; Visualization; Ontology; Multi-resolution Data Mining;« less

  20. Statistical Analysis of Large-Scale Structure of Universe

    NASA Astrophysics Data System (ADS)

    Tugay, A. V.

    While galaxy cluster catalogs were compiled many decades ago, other structural elements of cosmic web are detected at definite level only in the newest works. For example, extragalactic filaments were described by velocity field and SDSS galaxy distribution during the last years. Large-scale structure of the Universe could be also mapped in the future using ATHENA observations in X-rays and SKA in radio band. Until detailed observations are not available for the most volume of Universe, some integral statistical parameters can be used for its description. Such methods as galaxy correlation function, power spectrum, statistical moments and peak statistics are commonly used with this aim. The parameters of power spectrum and other statistics are important for constraining the models of dark matter, dark energy, inflation and brane cosmology. In the present work we describe the growth of large-scale density fluctuations in one- and three-dimensional case with Fourier harmonics of hydrodynamical parameters. In result we get power-law relation for the matter power spectrum.

  1. Large Scale Deformation of the Western U.S. Cordillera

    NASA Technical Reports Server (NTRS)

    Bennett, Richard A.

    2002-01-01

    Over the past couple of years, with support from NASA, we used a large collection of data from GPS, VLBI, SLR, and DORIS networks which span the Western U.S. Cordillera (WUSC) to precisely quantify present-day large-scale crustal deformations in a single uniform reference frame. Our work was roughly divided into an analysis of these space geodetic observations to infer the deformation field across and within the entire plate boundary zone, and an investigation of the implications of this deformation field regarding plate boundary dynamics. Following the determination of the first generation WUSC velocity solution, we placed high priority on the dissemination of the velocity estimates. With in-kind support from the Smithsonian Astrophysical Observatory, we constructed a web-site which allows anyone to access the data, and to determine their own velocity reference frame.

  2. Large Scale Deformation of the Western U.S. Cordillera

    NASA Technical Reports Server (NTRS)

    Bennett, Richard A.

    2002-01-01

    Over the past couple of years, with support from NASA, we used a large collection of data from GPS, VLBI, SLR, and DORIS networks which span the Westem U.S. Cordillera (WUSC) to precisely quantify present-day large-scale crustal deformations in a single uniform reference frame. Our work was roughly divided into an analysis of these space geodetic observations to infer the deformation field across and within the entire plate boundary zone, and an investigation of the implications of this deformation field regarding plate boundary dynamics. Following the determination of the first generation WUSC velocity solution, we placed high priority on the dissemination of the velocity estimates. With in-kind support from the Smithsonian Astrophysical Observatory, we constructed a web-site which allows anyone to access the data, and to determine their own velocity reference frame.

  3. MEMOSys: Bioinformatics platform for genome-scale metabolic models

    PubMed Central

    2011-01-01

    Background Recent advances in genomic sequencing have enabled the use of genome sequencing in standard biological and biotechnological research projects. The challenge is how to integrate the large amount of data in order to gain novel biological insights. One way to leverage sequence data is to use genome-scale metabolic models. We have therefore designed and implemented a bioinformatics platform which supports the development of such metabolic models. Results MEMOSys (MEtabolic MOdel research and development System) is a versatile platform for the management, storage, and development of genome-scale metabolic models. It supports the development of new models by providing a built-in version control system which offers access to the complete developmental history. Moreover, the integrated web board, the authorization system, and the definition of user roles allow collaborations across departments and institutions. Research on existing models is facilitated by a search system, references to external databases, and a feature-rich comparison mechanism. MEMOSys provides customizable data exchange mechanisms using the SBML format to enable analysis in external tools. The web application is based on the Java EE framework and offers an intuitive user interface. It currently contains six annotated microbial metabolic models. Conclusions We have developed a web-based system designed to provide researchers a novel application facilitating the management and development of metabolic models. The system is freely available at http://www.icbi.at/MEMOSys. PMID:21276275

  4. Allometric scaling enhances stability in complex food webs.

    PubMed

    Brose, Ulrich; Williams, Richard J; Martinez, Neo D

    2006-11-01

    Classic local stability theory predicts that complex ecological networks are unstable and are unlikely to persist despite empiricists' abundant documentation of such complexity in nature. This contradiction has puzzled biologists for decades. While some have explored how stability may be achieved in small modules of a few interacting species, rigorous demonstrations of how large complex and ecologically realistic networks dynamically persist remain scarce and inadequately understood. Here, we help fill this void by combining structural models of complex food webs with nonlinear bioenergetic models of population dynamics parameterized by biological rates that are allometrically scaled to populations' average body masses. Increasing predator-prey body mass ratios increase population persistence up to a saturation level that is reached by invertebrate and ectotherm vertebrate predators when being 10 or 100 times larger than their prey respectively. These values are corroborated by empirical predator-prey body mass ratios from a global data base. Moreover, negative effects of diversity (i.e. species richness) on stability (i.e. population persistence) become neutral or positive relationships at these empirical ratios. These results demonstrate that the predator-prey body mass ratios found in nature may be key to enabling persistence of populations in complex food webs and stabilizing the diversity of natural ecosystems.

  5. Linking Native and Invader Traits Explains Native Spider Population Responses to Plant Invasion.

    PubMed

    Smith, Jennifer N; Emlen, Douglas J; Pearson, Dean E

    2016-01-01

    Theoretically, the functional traits of native species should determine how natives respond to invader-driven changes. To explore this idea, we simulated a large-scale plant invasion using dead spotted knapweed (Centaurea stoebe) stems to determine if native spiders' web-building behaviors could explain differences in spider population responses to structural changes arising from C. stoebe invasion. After two years, irregular web-spiders were >30 times more abundant and orb weavers were >23 times more abundant on simulated invasion plots compared to controls. Additionally, irregular web-spiders on simulated invasion plots built webs that were 4.4 times larger and 5.0 times more likely to capture prey, leading to >2-fold increases in recruitment. Orb-weavers showed no differences in web size or prey captures between treatments. Web-spider responses to simulated invasion mimicked patterns following natural invasions, confirming that C. stoebe's architecture is likely the primary attribute driving native spider responses to these invasions. Differences in spider responses were attributable to differences in web construction behaviors relative to historic web substrate constraints. Orb-weavers in this system constructed webs between multiple plants, so they were limited by the overall quantity of native substrates but not by the architecture of individual native plant species. Irregular web-spiders built their webs within individual plants and were greatly constrained by the diminutive architecture of native plant substrates, so they were limited both by quantity and quality of native substrates. Evaluating native species traits in the context of invader-driven change can explain invasion outcomes and help to identify factors limiting native populations.

  6. Linking Native and Invader Traits Explains Native Spider Population Responses to Plant Invasion

    PubMed Central

    Emlen, Douglas J.; Pearson, Dean E.

    2016-01-01

    Theoretically, the functional traits of native species should determine how natives respond to invader-driven changes. To explore this idea, we simulated a large-scale plant invasion using dead spotted knapweed (Centaurea stoebe) stems to determine if native spiders’ web-building behaviors could explain differences in spider population responses to structural changes arising from C. stoebe invasion. After two years, irregular web-spiders were >30 times more abundant and orb weavers were >23 times more abundant on simulated invasion plots compared to controls. Additionally, irregular web-spiders on simulated invasion plots built webs that were 4.4 times larger and 5.0 times more likely to capture prey, leading to >2-fold increases in recruitment. Orb-weavers showed no differences in web size or prey captures between treatments. Web-spider responses to simulated invasion mimicked patterns following natural invasions, confirming that C. stoebe’s architecture is likely the primary attribute driving native spider responses to these invasions. Differences in spider responses were attributable to differences in web construction behaviors relative to historic web substrate constraints. Orb-weavers in this system constructed webs between multiple plants, so they were limited by the overall quantity of native substrates but not by the architecture of individual native plant species. Irregular web-spiders built their webs within individual plants and were greatly constrained by the diminutive architecture of native plant substrates, so they were limited both by quantity and quality of native substrates. Evaluating native species traits in the context of invader-driven change can explain invasion outcomes and help to identify factors limiting native populations. PMID:27082240

  7. Testing of a prototype Web based intervention for adolescent mothers on postpartum depression.

    PubMed

    Logsdon, M Cynthia; Barone, Michael; Lynch, Tania; Robertson, Ashley; Myers, John; Morrison, David; York, Sara; Gregg, Jennifer

    2013-08-01

    This article describes testing of a prototype Web site for adolescent mothers with postpartum depression; providing proof of concept. Participants (N=138) were recruited from a public school-based program for adolescent parents and completed the Mental Health Acceptability Scale, Stigma Scale for Receiving Psychological Help, and Attitudes Towards Seeking Professional Psychological Help Scale before, and after, the Web site intervention. They also provided feedback on the usability of the Web site. Attitudes related to depression and treatment (ATSPPH) improved after viewing the Web site (p=.023). Feedback on the Web site indicated that it was easy to use (77%), reflecting highly acceptable score for product usability. The data provide the foundation for the launch of the Web site from prototype to product and more comprehensive testing. The creation and testing of informational text messages will be added to the Web site to increase the interactivity and dose of the intervention. Copyright © 2013 Elsevier Inc. All rights reserved.

  8. Visualization and Analysis of Multi-scale Land Surface Products via Giovanni Portals

    NASA Technical Reports Server (NTRS)

    Shen, Suhung; Kempler, Steven J.; Gerasimov, Irina V.

    2013-01-01

    Large volumes of MODIS land data products at multiple spatial resolutions have been integrated into the Giovanni online analysis system to support studies on land cover and land use changes,focused on the Northern Eurasia and Monsoon Asia regions through the LCLUC program. Giovanni (Goddard Interactive Online Visualization ANd aNalysis Infrastructure) is a Web-based application developed by the NASA Goddard Earth Sciences Data and Information Services Center (GES DISC), providing a simple and intuitive way to visualize, analyze, and access Earth science remotely-sensed and modeled data.Customized Giovanni Web portals (Giovanni-NEESPI andGiovanni-MAIRS) have been created to integrate land, atmospheric,cryospheric, and societal products, enabling researchers to do quick exploration and basic analyses of land surface changes, and their relationships to climate, at global and regional scales. This presentation shows a sample Giovanni portal page, lists selected data products in the system, and illustrates potential analyses with imagesand time-series at global and regional scales, focusing on climatology and anomaly analysis. More information is available at the GES DISCMAIRS data support project portal: http:disc.sci.gsfc.nasa.govmairs.

  9. A new probe of the magnetic field power spectrum in cosmic web filaments

    NASA Astrophysics Data System (ADS)

    Hales, Christopher A.; Greiner, Maksim; Ensslin, Torsten A.

    2015-08-01

    Establishing the properties of magnetic fields on scales larger than galaxy clusters is critical for resolving the unknown origin and evolution of galactic and cluster magnetism. More generally, observations of magnetic fields on cosmic scales are needed for assessing the impacts of magnetism on cosmology, particle physics, and structure formation over the full history of the Universe. However, firm observational evidence for magnetic fields in large scale structure remains elusive. In an effort to address this problem, we have developed a novel statistical method to infer the magnetic field power spectrum in cosmic web filaments using observation of the two-point correlation of Faraday rotation measures from a dense grid of extragalactic radio sources. Here we describe our approach, which embeds and extends the pioneering work of Kolatt (1998) within the context of Information Field Theory (a statistical theory for Bayesian inference on spatially distributed signals; Enfllin et al., 2009). We describe prospects for observation, for example with forthcoming data from the ultra-deep JVLA CHILES Con Pol survey and future surveys with the SKA.

  10. Web-video-mining-supported workflow modeling for laparoscopic surgeries.

    PubMed

    Liu, Rui; Zhang, Xiaoli; Zhang, Hao

    2016-11-01

    As quality assurance is of strong concern in advanced surgeries, intelligent surgical systems are expected to have knowledge such as the knowledge of the surgical workflow model (SWM) to support their intuitive cooperation with surgeons. For generating a robust and reliable SWM, a large amount of training data is required. However, training data collected by physically recording surgery operations is often limited and data collection is time-consuming and labor-intensive, severely influencing knowledge scalability of the surgical systems. The objective of this research is to solve the knowledge scalability problem in surgical workflow modeling with a low cost and labor efficient way. A novel web-video-mining-supported surgical workflow modeling (webSWM) method is developed. A novel video quality analysis method based on topic analysis and sentiment analysis techniques is developed to select high-quality videos from abundant and noisy web videos. A statistical learning method is then used to build the workflow model based on the selected videos. To test the effectiveness of the webSWM method, 250 web videos were mined to generate a surgical workflow for the robotic cholecystectomy surgery. The generated workflow was evaluated by 4 web-retrieved videos and 4 operation-room-recorded videos, respectively. The evaluation results (video selection consistency n-index ≥0.60; surgical workflow matching degree ≥0.84) proved the effectiveness of the webSWM method in generating robust and reliable SWM knowledge by mining web videos. With the webSWM method, abundant web videos were selected and a reliable SWM was modeled in a short time with low labor cost. Satisfied performances in mining web videos and learning surgery-related knowledge show that the webSWM method is promising in scaling knowledge for intelligent surgical systems. Copyright © 2016 Elsevier B.V. All rights reserved.

  11. TV Audience Measurement with Big Data.

    PubMed

    Hill, Shawndra

    2014-06-01

    TV audience measurement involves estimating the number of viewers tuned into a TV show at any given time as well as their demographics. First introduced shortly after commercial television broadcasting began in the late 1940s, audience measurement allowed the business of television to flourish by offering networks a way to quantify the monetary value of TV audiences for advertisers, who pay for the estimated number of eyeballs watching during commercials. The first measurement techniques suffered from multiple limitations because reliable, large-scale data were costly to acquire. Yet despite these limitations, measurement standards remained largely unchanged for decades until devices such as cable boxes, video-on-demand boxes, and cell phones, as well as web apps, Internet browser clicks, web queries, and social media activity, resulted in an explosion of digitally available data. TV viewers now leave digital traces that can be used to track almost every aspect of their daily lives, allowing the potential for large-scale aggregation across data sources for individual users and groups and enabling the tracking of more people on more dimensions for more shows. Data are now more comprehensive, available in real time, and cheaper to acquire, enabling accurate and fine-grained TV audience measurement. In this article, I discuss the evolution of audience measurement and what the recent data explosion means for the TV industry and academic research.

  12. Cooperative capture of large prey solves scaling challenge faced by spider societies

    PubMed Central

    Yip, Eric C.; Powers, Kimberly S.; Avilés, Leticia

    2008-01-01

    A decrease in the surface area per unit volume is a well known constraint setting limits to the size of organisms at both the cellular and whole-organismal levels. Similar constraints may apply to social groups as they grow in size. The communal three-dimensional webs that social spiders build function ecologically as single units that intercept prey through their surface and should thus be subject to this constraint. Accordingly, we show that web prey capture area per spider, and thus number of insects captured per capita, decreases with colony size in a neotropical social spider. Prey biomass intake per capita, however, peaks at intermediate colony sizes because the spiders forage cooperatively and larger colonies capture increasingly large insects. A peaked prey biomass intake function would explain not only why these spiders live in groups and cooperate but also why they disperse only at large colony sizes, thus addressing both sociality and colony size range in this social spider. These findings may also explain the conspicuous absence of social spiders from higher latitudes and higher elevations, areas that we have previously shown to harbor considerably fewer insects of the largest size classes than the lowland tropical rainforests where social spiders thrive. Our findings thus illustrate the relevance of scaling laws to the size and functioning of levels of organization above the individual. PMID:18689677

  13. Large-scale Scanning Transmission Electron Microscopy (Nanotomy) of Healthy and Injured Zebrafish Brain.

    PubMed

    Kuipers, Jeroen; Kalicharan, Ruby D; Wolters, Anouk H G; van Ham, Tjakko J; Giepmans, Ben N G

    2016-05-25

    Large-scale 2D electron microscopy (EM), or nanotomy, is the tissue-wide application of nanoscale resolution electron microscopy. Others and we previously applied large scale EM to human skin pancreatic islets, tissue culture and whole zebrafish larvae(1-7). Here we describe a universally applicable method for tissue-scale scanning EM for unbiased detection of sub-cellular and molecular features. Nanotomy was applied to investigate the healthy and a neurodegenerative zebrafish brain. Our method is based on standardized EM sample preparation protocols: Fixation with glutaraldehyde and osmium, followed by epoxy-resin embedding, ultrathin sectioning and mounting of ultrathin-sections on one-hole grids, followed by post staining with uranyl and lead. Large-scale 2D EM mosaic images are acquired using a scanning EM connected to an external large area scan generator using scanning transmission EM (STEM). Large scale EM images are typically ~ 5 - 50 G pixels in size, and best viewed using zoomable HTML files, which can be opened in any web browser, similar to online geographical HTML maps. This method can be applied to (human) tissue, cross sections of whole animals as well as tissue culture(1-5). Here, zebrafish brains were analyzed in a non-invasive neuronal ablation model. We visualize within a single dataset tissue, cellular and subcellular changes which can be quantified in various cell types including neurons and microglia, the brain's macrophages. In addition, nanotomy facilitates the correlation of EM with light microscopy (CLEM)(8) on the same tissue, as large surface areas previously imaged using fluorescent microscopy, can subsequently be subjected to large area EM, resulting in the nano-anatomy (nanotomy) of tissues. In all, nanotomy allows unbiased detection of features at EM level in a tissue-wide quantifiable manner.

  14. Large-scale Scanning Transmission Electron Microscopy (Nanotomy) of Healthy and Injured Zebrafish Brain

    PubMed Central

    Kuipers, Jeroen; Kalicharan, Ruby D.; Wolters, Anouk H. G.

    2016-01-01

    Large-scale 2D electron microscopy (EM), or nanotomy, is the tissue-wide application of nanoscale resolution electron microscopy. Others and we previously applied large scale EM to human skin pancreatic islets, tissue culture and whole zebrafish larvae1-7. Here we describe a universally applicable method for tissue-scale scanning EM for unbiased detection of sub-cellular and molecular features. Nanotomy was applied to investigate the healthy and a neurodegenerative zebrafish brain. Our method is based on standardized EM sample preparation protocols: Fixation with glutaraldehyde and osmium, followed by epoxy-resin embedding, ultrathin sectioning and mounting of ultrathin-sections on one-hole grids, followed by post staining with uranyl and lead. Large-scale 2D EM mosaic images are acquired using a scanning EM connected to an external large area scan generator using scanning transmission EM (STEM). Large scale EM images are typically ~ 5 - 50 G pixels in size, and best viewed using zoomable HTML files, which can be opened in any web browser, similar to online geographical HTML maps. This method can be applied to (human) tissue, cross sections of whole animals as well as tissue culture1-5. Here, zebrafish brains were analyzed in a non-invasive neuronal ablation model. We visualize within a single dataset tissue, cellular and subcellular changes which can be quantified in various cell types including neurons and microglia, the brain's macrophages. In addition, nanotomy facilitates the correlation of EM with light microscopy (CLEM)8 on the same tissue, as large surface areas previously imaged using fluorescent microscopy, can subsequently be subjected to large area EM, resulting in the nano-anatomy (nanotomy) of tissues. In all, nanotomy allows unbiased detection of features at EM level in a tissue-wide quantifiable manner. PMID:27285162

  15. Engineering web maps with gradual content zoom based on streaming vector data

    NASA Astrophysics Data System (ADS)

    Huang, Lina; Meijers, Martijn; Šuba, Radan; van Oosterom, Peter

    2016-04-01

    Vario-scale data structures have been designed to support gradual content zoom and the progressive transfer of vector data, for use with arbitrary map scales. The focus to date has been on the server side, especially on how to convert geographic data into the proposed vario-scale structures by means of automated generalisation. This paper contributes to the ongoing vario-scale research by focusing on the client side and communication, particularly on how this works in a web-services setting. It is claimed that these functionalities are urgently needed, as many web-based applications, both desktop and mobile, require gradual content zoom, progressive transfer and a high performance level. The web-client prototypes developed in this paper make it possible to assess the behaviour of vario-scale data and to determine how users will actually see the interactions. Several different options of web-services communication architectures are possible in a vario-scale setting. These options are analysed and tested with various web-client prototypes, with respect to functionality, ease of implementation and performance (amount of transmitted data and response times). We show that the vario-scale data structure can fit in with current web-based architectures and efforts to standardise map distribution on the internet. However, to maximise the benefits of vario-scale data, a client needs to be aware of this structure. When a client needs a map to be refined (by means of a gradual content zoom operation), only the 'missing' data will be requested. This data will be sent incrementally to the client from a server. In this way, the amount of data transferred at one time is reduced, shortening the transmission time. In addition to these conceptual architecture aspects, there are many implementation and tooling design decisions at play. These will also be elaborated on in this paper. Based on the experiments conducted, we conclude that the vario-scale approach indeed supports gradual content zoom and the progressive web transfer of vector data. This is a big step forward in making vector data at arbitrary map scales available to larger user groups.

  16. Repurposing of open data through large scale hydrological modelling - hypeweb.smhi.se

    NASA Astrophysics Data System (ADS)

    Strömbäck, Lena; Andersson, Jafet; Donnelly, Chantal; Gustafsson, David; Isberg, Kristina; Pechlivanidis, Ilias; Strömqvist, Johan; Arheimer, Berit

    2015-04-01

    Hydrological modelling demands large amounts of spatial data, such as soil properties, land use, topography, lakes and reservoirs, ice and snow coverage, water management (e.g. irrigation patterns and regulations), meteorological data and observed water discharge in rivers. By using such data, the hydrological model will in turn provide new data that can be used for new purposes (i.e. re-purposing). This presentation will give an example of how readily available open data from public portals have been re-purposed by using the Hydrological Predictions for the Environment (HYPE) model in a number of large-scale model applications covering numerous subbasins and rivers. HYPE is a dynamic, semi-distributed, process-based, and integrated catchment model. The model output is launched as new Open Data at the web site www.hypeweb.smhi.se to be used for (i) Climate change impact assessments on water resources and dynamics; (ii) The European Water Framework Directive (WFD) for characterization and development of measure programs to improve the ecological status of water bodies; (iii) Design variables for infrastructure constructions; (iv) Spatial water-resource mapping; (v) Operational forecasts (1-10 days and seasonal) on floods and droughts; (vi) Input to oceanographic models for operational forecasts and marine status assessments; (vii) Research. The following regional domains have been modelled so far with different resolutions (number of subbasins within brackets): Sweden (37 000), Europe (35 000), Arctic basin (30 000), La Plata River (6 000), Niger River (800), Middle-East North-Africa (31 000), and the Indian subcontinent (6 000). The Hype web site provides several interactive web applications for exploring results from the models. The user can explore an overview of various water variables for historical and future conditions. Moreover the user can explore and download historical time series of discharge for each basin and explore the performance of the model towards observed river flow. The presentation will describe the Open Data sources used, show the functionality of the web site and discuss model performance and experience from this world-wide hydrological modelling of multi-basins using open data.

  17. xQTL workbench: a scalable web environment for multi-level QTL analysis.

    PubMed

    Arends, Danny; van der Velde, K Joeri; Prins, Pjotr; Broman, Karl W; Möller, Steffen; Jansen, Ritsert C; Swertz, Morris A

    2012-04-01

    xQTL workbench is a scalable web platform for the mapping of quantitative trait loci (QTLs) at multiple levels: for example gene expression (eQTL), protein abundance (pQTL), metabolite abundance (mQTL) and phenotype (phQTL) data. Popular QTL mapping methods for model organism and human populations are accessible via the web user interface. Large calculations scale easily on to multi-core computers, clusters and Cloud. All data involved can be uploaded and queried online: markers, genotypes, microarrays, NGS, LC-MS, GC-MS, NMR, etc. When new data types come available, xQTL workbench is quickly customized using the Molgenis software generator. xQTL workbench runs on all common platforms, including Linux, Mac OS X and Windows. An online demo system, installation guide, tutorials, software and source code are available under the LGPL3 license from http://www.xqtl.org. m.a.swertz@rug.nl.

  18. xQTL workbench: a scalable web environment for multi-level QTL analysis

    PubMed Central

    Arends, Danny; van der Velde, K. Joeri; Prins, Pjotr; Broman, Karl W.; Möller, Steffen; Jansen, Ritsert C.; Swertz, Morris A.

    2012-01-01

    Summary: xQTL workbench is a scalable web platform for the mapping of quantitative trait loci (QTLs) at multiple levels: for example gene expression (eQTL), protein abundance (pQTL), metabolite abundance (mQTL) and phenotype (phQTL) data. Popular QTL mapping methods for model organism and human populations are accessible via the web user interface. Large calculations scale easily on to multi-core computers, clusters and Cloud. All data involved can be uploaded and queried online: markers, genotypes, microarrays, NGS, LC-MS, GC-MS, NMR, etc. When new data types come available, xQTL workbench is quickly customized using the Molgenis software generator. Availability: xQTL workbench runs on all common platforms, including Linux, Mac OS X and Windows. An online demo system, installation guide, tutorials, software and source code are available under the LGPL3 license from http://www.xqtl.org. Contact: m.a.swertz@rug.nl PMID:22308096

  19. Large scale healthcare data integration and analysis using the semantic web.

    PubMed

    Timm, John; Renly, Sondra; Farkash, Ariel

    2011-01-01

    Healthcare data interoperability can only be achieved when the semantics of the content is well defined and consistently implemented across heterogeneous data sources. Achieving these objectives of interoperability requires the collaboration of experts from several domains. This paper describes tooling that integrates Semantic Web technologies with common tools to facilitate cross-domain collaborative development for the purposes of data interoperability. Our approach is divided into stages of data harmonization and representation, model transformation, and instance generation. We applied our approach on Hypergenes, an EU funded project, where we use our method to the Essential Hypertension disease model using a CDA template. Our domain expert partners include clinical providers, clinical domain researchers, healthcare information technology experts, and a variety of clinical data consumers. We show that bringing Semantic Web technologies into the healthcare interoperability toolkit increases opportunities for beneficial collaboration thus improving patient care and clinical research outcomes.

  20. Bringing modeling to the masses: A web based system to predict potential species distributions

    USGS Publications Warehouse

    Graham, Jim; Newman, Greg; Kumar, Sunil; Jarnevich, Catherine S.; Young, Nick; Crall, Alycia W.; Stohlgren, Thomas J.; Evangelista, Paul

    2010-01-01

    Predicting current and potential species distributions and abundance is critical for managing invasive species, preserving threatened and endangered species, and conserving native species and habitats. Accurate predictive models are needed at local, regional, and national scales to guide field surveys, improve monitoring, and set priorities for conservation and restoration. Modeling capabilities, however, are often limited by access to software and environmental data required for predictions. To address these needs, we built a comprehensive web-based system that: (1) maintains a large database of field data; (2) provides access to field data and a wealth of environmental data; (3) accesses values in rasters representing environmental characteristics; (4) runs statistical spatial models; and (5) creates maps that predict the potential species distribution. The system is available online at www.niiss.org, and provides web-based tools for stakeholders to create potential species distribution models and maps under current and future climate scenarios.

  1. Automation of Hessian-Based Tubularity Measure Response Function in 3D Biomedical Images.

    PubMed

    Dzyubak, Oleksandr P; Ritman, Erik L

    2011-01-01

    The blood vessels and nerve trees consist of tubular objects interconnected into a complex tree- or web-like structure that has a range of structural scale 5 μm diameter capillaries to 3 cm aorta. This large-scale range presents two major problems; one is just making the measurements, and the other is the exponential increase of component numbers with decreasing scale. With the remarkable increase in the volume imaged by, and resolution of, modern day 3D imagers, it is almost impossible to make manual tracking of the complex multiscale parameters from those large image data sets. In addition, the manual tracking is quite subjective and unreliable. We propose a solution for automation of an adaptive nonsupervised system for tracking tubular objects based on multiscale framework and use of Hessian-based object shape detector incorporating National Library of Medicine Insight Segmentation and Registration Toolkit (ITK) image processing libraries.

  2. On the connectivity of the cosmic web: theory and implications for cosmology and galaxy formation

    NASA Astrophysics Data System (ADS)

    Codis, Sandrine; Pogosyan, Dmitri; Pichon, Christophe

    2018-06-01

    Cosmic connectivity and multiplicity, i.e. the number of filaments globally or locally connected to a given cluster is a natural probe of the growth of structure and in particular of the nature of dark energy. It is also a critical ingredient driving the assembly history of galaxies as it controls mass and angular momentum accretion. The connectivity of the cosmic web is investigated here via the persistent skeleton. This tool identifies topologically the ridges of the cosmic landscape which allows us to investigate how the nodes of the cosmic web are connected together. When applied to Gaussian random fields corresponding to the high redshift universe, it is found that on average the nodes are connected to exactly κ = 4 neighbours in two dimensions and ˜6.1 in three dimensions. Investigating spatial dimensions up to d = 6, typical departures from a cubic lattice κ = 2d are shown to scale like the power 7/4 of the dimension. These numbers strongly depend on the height of the peaks: the higher the peak the larger the connectivity. Predictions from first principles based on peak theory are shown to reproduce well the connectivity and multiplicity of Gaussian random fields and cosmological simulations. As an illustration, connectivity is quantified in galaxy lensing convergence maps and large dark haloes catalogues. As a function of redshift and scale the mean connectivity decreases in a cosmology-dependent way. As a function of halo mass it scales like 10/3 times the log of the mass. Implications on galactic scales are discussed.

  3. Web-based segmentation and display of three-dimensional radiologic image data.

    PubMed

    Silverstein, J; Rubenstein, J; Millman, A; Panko, W

    1998-01-01

    In many clinical circumstances, viewing sequential radiological image data as three-dimensional models is proving beneficial. However, designing customized computer-generated radiological models is beyond the scope of most physicians, due to specialized hardware and software requirements. We have created a simple method for Internet users to remotely construct and locally display three-dimensional radiological models using only a standard web browser. Rapid model construction is achieved by distributing the hardware intensive steps to a remote server. Once created, the model is automatically displayed on the requesting browser and is accessible to multiple geographically distributed users. Implementation of our server software on large scale systems could be of great service to the worldwide medical community.

  4. Raising Student Awareness of the Use of English for Specific Business Purposes in the European Context: A Staff-Student Project

    ERIC Educational Resources Information Center

    Nickerson, C.; Gerritsen, M.; Meurs, F.v.

    2005-01-01

    This Research Note reports on a large-scale staff-student project focussing on the use of English for Specific Business Purposes in a number of promotional genres (TV commercials, annual reports, corporate web-sites, print advertising) within several of the EU member states: Belgium, the Netherlands, France, Germany and Spain. The project as a…

  5. Global-scale carbon and energy flows through the marine planktonic food web: An analysis with a coupled physical-biological model

    NASA Astrophysics Data System (ADS)

    Stock, Charles A.; Dunne, John P.; John, Jasmin G.

    2014-01-01

    Global-scale planktonic ecosystem models exhibit large differences in simulated net primary production (NPP) and assessment of planktonic food web fluxes beyond primary producers has been limited, diminishing confidence in carbon flux estimates from these models. In this study, a global ocean-ice-ecosystem model was assessed against a suite of observation-based planktonic food web flux estimates, many of which were not considered in previous modeling studies. The simulation successfully captured cross-biome differences and similarities in these fluxes after calibration of a limited number of highly uncertain yet influential parameters. The resulting comprehensive carbon budgets suggested that shortened food webs, elevated growth efficiencies, and tight consumer-resource coupling enable oceanic upwelling systems to support 45% of pelagic mesozooplankton production despite accounting for only 22% of ocean area and 34% of NPP. In seasonally stratified regions (42% of ocean area and 40% of NPP), weakened consumer-resource coupling tempers mesozooplankton production to 41% and enhances export below 100 m to 48% of the global total. In oligotrophic systems (36% of ocean area and 26% of NPP), the dominance of small phytoplankton and low consumer growth efficiencies supported only 14% of mesozooplankton production and 17% of export globally. Bacterial production, in contrast, was maintained in nearly constant proportion to primary production across biomes through the compensating effects of increased partitioning of NPP to the microbial food web in oligotrophic ecosystems and increased bacterial growth efficiencies in more productive areas. Cross-biome differences in mesozooplankton trophic level were muted relative to those invoked by previous work such that significant differences in consumer growth efficiencies and the strength of consumer-resource coupling were needed to explain sharp cross-biome differences in mesozooplankton production. Lastly, simultaneous consideration of multiple flux constraints supports a highly distributed view of respiration across the planktonic food web rather than one dominated by heterotrophic bacteria. The solution herein is unlikely unique in its ability to explain observed cross-biome energy flow patterns and notable misfits remain. Resolution of existing uncertainties in observed biome-scale productivity and increasingly mechanistic physical and biological model components should yield significant refinements to estimates herein.

  6. Textual and visual content-based anti-phishing: a Bayesian approach.

    PubMed

    Zhang, Haijun; Liu, Gang; Chow, Tommy W S; Liu, Wenyin

    2011-10-01

    A novel framework using a Bayesian approach for content-based phishing web page detection is presented. Our model takes into account textual and visual contents to measure the similarity between the protected web page and suspicious web pages. A text classifier, an image classifier, and an algorithm fusing the results from classifiers are introduced. An outstanding feature of this paper is the exploration of a Bayesian model to estimate the matching threshold. This is required in the classifier for determining the class of the web page and identifying whether the web page is phishing or not. In the text classifier, the naive Bayes rule is used to calculate the probability that a web page is phishing. In the image classifier, the earth mover's distance is employed to measure the visual similarity, and our Bayesian model is designed to determine the threshold. In the data fusion algorithm, the Bayes theory is used to synthesize the classification results from textual and visual content. The effectiveness of our proposed approach was examined in a large-scale dataset collected from real phishing cases. Experimental results demonstrated that the text classifier and the image classifier we designed deliver promising results, the fusion algorithm outperforms either of the individual classifiers, and our model can be adapted to different phishing cases. © 2011 IEEE

  7. WebMedSA: a web-based framework for segmenting and annotating medical images using biomedical ontologies

    NASA Astrophysics Data System (ADS)

    Vega, Francisco; Pérez, Wilson; Tello, Andrés.; Saquicela, Victor; Espinoza, Mauricio; Solano-Quinde, Lizandro; Vidal, Maria-Esther; La Cruz, Alexandra

    2015-12-01

    Advances in medical imaging have fostered medical diagnosis based on digital images. Consequently, the number of studies by medical images diagnosis increases, thus, collaborative work and tele-radiology systems are required to effectively scale up to this diagnosis trend. We tackle the problem of the collaborative access of medical images, and present WebMedSA, a framework to manage large datasets of medical images. WebMedSA relies on a PACS and supports the ontological annotation, as well as segmentation and visualization of the images based on their semantic description. Ontological annotations can be performed directly on the volumetric image or at different image planes (e.g., axial, coronal, or sagittal); furthermore, annotations can be complemented after applying a segmentation technique. WebMedSA is based on three main steps: (1) RDF-ization process for extracting, anonymizing, and serializing metadata comprised in DICOM medical images into RDF/XML; (2) Integration of different biomedical ontologies (using L-MOM library), making this approach ontology independent; and (3) segmentation and visualization of annotated data which is further used to generate new annotations according to expert knowledge, and validation. Initial user evaluations suggest that WebMedSA facilitates the exchange of knowledge between radiologists, and provides the basis for collaborative work among them.

  8. 'Sciencenet'--towards a global search and share engine for all scientific knowledge.

    PubMed

    Lütjohann, Dominic S; Shah, Asmi H; Christen, Michael P; Richter, Florian; Knese, Karsten; Liebel, Urban

    2011-06-15

    Modern biological experiments create vast amounts of data which are geographically distributed. These datasets consist of petabytes of raw data and billions of documents. Yet to the best of our knowledge, a search engine technology that searches and cross-links all different data types in life sciences does not exist. We have developed a prototype distributed scientific search engine technology, 'Sciencenet', which facilitates rapid searching over this large data space. By 'bringing the search engine to the data', we do not require server farms. This platform also allows users to contribute to the search index and publish their large-scale data to support e-Science. Furthermore, a community-driven method guarantees that only scientific content is crawled and presented. Our peer-to-peer approach is sufficiently scalable for the science web without performance or capacity tradeoff. The free to use search portal web page and the downloadable client are accessible at: http://sciencenet.kit.edu. The web portal for index administration is implemented in ASP.NET, the 'AskMe' experiment publisher is written in Python 2.7, and the backend 'YaCy' search engine is based on Java 1.6.

  9. A performance study of WebDav access to storages within the Belle II collaboration

    NASA Astrophysics Data System (ADS)

    Pardi, S.; Russo, G.

    2017-10-01

    WebDav and HTTP are becoming popular protocols for data access in the High Energy Physics community. The most used Grid and Cloud storage solutions provide such kind of interfaces, in this scenario tuning and performance evaluation became crucial aspects to promote the adoption of these protocols within the Belle II community. In this work, we present the results of a large-scale test activity, made with the goal to evaluate performances and reliability of the WebDav protocol, and study a possible adoption for the user analysis. More specifically, we considered a pilot infrastructure composed by a set of storage elements configured with the WebDav interface, hosted at the Belle II sites. The performance tests include a comparison with xrootd and gridftp. As reference tests we used a set of analysis jobs running under the Belle II software framework, accessing the input data with the ROOT I/O library, in order to simulate as much as possible a realistic user activity. The final analysis shows the possibility to achieve promising performances with WebDav on different storage systems, and gives an interesting feedback, for Belle II community and for other high energy physics experiments.

  10. Winners don't take all: Characterizing the competition for links on the web

    NASA Astrophysics Data System (ADS)

    Pennock, David M.; Flake, Gary W.; Lawrence, Steve; Glover, Eric J.; Giles, C. Lee

    2002-04-01

    As a whole, the World Wide Web displays a striking "rich get richer" behavior, with a relatively small number of sites receiving a disproportionately large share of hyperlink references and traffic. However, hidden in this skewed global distribution, we discover a qualitatively different and considerably less biased link distribution among subcategories of pagesfor example, among all university homepages or all newspaper homepages. Although the connectivity distribution over the entire web is close to a pure power law, we find that the distribution within specific categories is typically unimodal on a log scale, with the location of the mode, and thus the extent of the rich get richer phenomenon, varying across different categories. Similar distributions occur in many other naturally occurring networks, including research paper citations, movie actor collaborations, and United States power grid connections. A simple generative model, incorporating a mixture of preferential and uniform attachment, quantifies the degree to which the rich nodes grow richer, and how new (and poorly connected) nodes can compete. The model accurately accounts for the true connectivity distributions of category-specific web pages, the web as a whole, and other social networks.

  11. 3Drefine: an interactive web server for efficient protein structure refinement.

    PubMed

    Bhattacharya, Debswapna; Nowotny, Jackson; Cao, Renzhi; Cheng, Jianlin

    2016-07-08

    3Drefine is an interactive web server for consistent and computationally efficient protein structure refinement with the capability to perform web-based statistical and visual analysis. The 3Drefine refinement protocol utilizes iterative optimization of hydrogen bonding network combined with atomic-level energy minimization on the optimized model using a composite physics and knowledge-based force fields for efficient protein structure refinement. The method has been extensively evaluated on blind CASP experiments as well as on large-scale and diverse benchmark datasets and exhibits consistent improvement over the initial structure in both global and local structural quality measures. The 3Drefine web server allows for convenient protein structure refinement through a text or file input submission, email notification, provided example submission and is freely available without any registration requirement. The server also provides comprehensive analysis of submissions through various energy and statistical feedback and interactive visualization of multiple refined models through the JSmol applet that is equipped with numerous protein model analysis tools. The web server has been extensively tested and used by many users. As a result, the 3Drefine web server conveniently provides a useful tool easily accessible to the community. The 3Drefine web server has been made publicly available at the URL: http://sysbio.rnet.missouri.edu/3Drefine/. © The Author(s) 2016. Published by Oxford University Press on behalf of Nucleic Acids Research.

  12. Warm-hot baryons comprise 5-10 per cent of filaments in the cosmic web.

    PubMed

    Eckert, Dominique; Jauzac, Mathilde; Shan, HuanYuan; Kneib, Jean-Paul; Erben, Thomas; Israel, Holger; Jullo, Eric; Klein, Matthias; Massey, Richard; Richard, Johan; Tchernin, Céline

    2015-12-03

    Observations of the cosmic microwave background indicate that baryons account for 5 per cent of the Universe's total energy content. In the local Universe, the census of all observed baryons falls short of this estimate by a factor of two. Cosmological simulations indicate that the missing baryons have not condensed into virialized haloes, but reside throughout the filaments of the cosmic web (where matter density is larger than average) as a low-density plasma at temperatures of 10(5)-10(7) kelvin, known as the warm-hot intergalactic medium. There have been previous claims of the detection of warm-hot baryons along the line of sight to distant blazars and of hot gas between interacting clusters. These observations were, however, unable to trace the large-scale filamentary structure, or to estimate the total amount of warm-hot baryons in a representative volume of the Universe. Here we report X-ray observations of filamentary structures of gas at 10(7) kelvin associated with the galaxy cluster Abell 2744. Previous observations of this cluster were unable to resolve and remove coincidental X-ray point sources. After subtracting these, we find hot gas structures that are coherent over scales of 8 megaparsecs. The filaments coincide with over-densities of galaxies and dark matter, with 5-10 per cent of their mass in baryonic gas. This gas has been heated up by the cluster's gravitational pull and is now feeding its core. Our findings strengthen evidence for a picture of the Universe in which a large fraction of the missing baryons reside in the filaments of the cosmic web.

  13. Large-scale structure non-Gaussianities with modal methods

    NASA Astrophysics Data System (ADS)

    Schmittfull, Marcel

    2016-10-01

    Relying on a separable modal expansion of the bispectrum, the implementation of a fast estimator for the full bispectrum of a 3d particle distribution is presented. The computational cost of accurate bispectrum estimation is negligible relative to simulation evolution, so the bispectrum can be used as a standard diagnostic whenever the power spectrum is evaluated. As an application, the time evolution of gravitational and primordial dark matter bispectra was measured in a large suite of N-body simulations. The bispectrum shape changes characteristically when the cosmic web becomes dominated by filaments and halos, therefore providing a quantitative probe of 3d structure formation. Our measured bispectra are determined by ~ 50 coefficients, which can be used as fitting formulae in the nonlinear regime and for non-Gaussian initial conditions. We also compare the measured bispectra with predictions from the Effective Field Theory of Large Scale Structures (EFTofLSS).

  14. Shift in a Large River Fish Assemblage: Body-Size and Trophic Structure Dynamics

    PubMed Central

    Broadway, Kyle J.; Pyron, Mark; Gammon, James R.; Murry, Brent A.

    2015-01-01

    As the intensity and speed of environmental change increase at both local and global scales it is imperative that we gain a better understanding of the ecological implications of community shifts. While there has been substantial progress toward understanding the drivers and subsequent responses of community change (e.g. lake trophic state), the ecological impacts of food web changes are far less understood. We analyzed Wabash River fish assemblage data collected from 1974-2008, to evaluate temporal variation in body-size structure and functional group composition. Two parameters derived from annual community size-spectra were our major response variables: (1) the regression slope is an index of ecological efficiency and predator-prey biomass ratios, and (2) spectral elevation (regression midpoint height) is a proxy for food web capacity. We detected a large assemblage shift, over at least a seven year period, defined by dramatic changes in abundance (measured as catch-per-unit-effort) of the dominant functional feeding groups among two time periods; from an assemblage dominated by planktivore-omnivores to benthic invertivores. There was a concurrent increase in ecological efficiency (slopes increased over time) following the shift associated with an increase in large-bodied low trophic level fish. Food web capacity remained relatively stable with no clear temporal trends. Thus, increased ecological efficiency occurred simultaneous to a compensatory response that shifted biomass among functional feeding groups. PMID:25902144

  15. Promoting Teachers' Positive Attitude towards Web Use: A Study in Web Site Development

    ERIC Educational Resources Information Center

    Akpinar, Yavuz; Bayramoglu, Yusuf

    2008-01-01

    The purpose of the study was to examine effects of a compact training for developing web sites on teachers' web attitude, as composed of: web self efficacy, perceived web enjoyment, perceived web usefulness and behavioral intention to use the web. To measure the related constructs, the Web Attitude Scale was adapted into Turkish and tested with a…

  16. Creating a gold medal Olympic and Paralympics health care team: a satisfaction survey of the mobile medical unit/polyclinic team training for the Vancouver 2010 winter games

    PubMed Central

    2013-01-01

    Background The mobile medical unit/polyclinic (MMU/PC) was an essential part of the medical services to support ill or injured Olympic or Paralympics family during the 2010 Olympic and Paralympics winter games. The objective of this study was to survey the satisfaction of the clinical staff that completed the training programs prior to deployment to the MMU. Methods Medical personnel who participated in at least one of the four training programs, including (1) week-end sessions; (2) web-based modules; (3) just-in-time training; and (4) daily simulation exercises were invited to participate in a web-based survey and comment on their level of satisfaction with training program. Results A total of 64 (out of 94 who were invited) physicians, nurses and respiratory therapists completed the survey. All participants reported favorably that the MMU/PC training positively impacted their knowledge, skills and team functions while deployed at the MMU/PC during the 2010 Olympic Games. However, components of the training program were valued differently depending on clinical job title, years of experience, and prior experience in large scale events. Respondents with little or no experience working in large scale events (45%) rated daily simulations as the most valuable component of the training program for strengthening competencies and knowledge in clinical skills for working in large scale events. Conclusion The multi-phase MMU/PC training was found to be beneficial for preparing the medical team for the 2010 Winter Games. In particular this survey demonstrates the effectiveness of simulation training programs on teamwork competencies in ad hoc groups. PMID:24225074

  17. SensorDB: a virtual laboratory for the integration, visualization and analysis of varied biological sensor data.

    PubMed

    Salehi, Ali; Jimenez-Berni, Jose; Deery, David M; Palmer, Doug; Holland, Edward; Rozas-Larraondo, Pablo; Chapman, Scott C; Georgakopoulos, Dimitrios; Furbank, Robert T

    2015-01-01

    To our knowledge, there is no software or database solution that supports large volumes of biological time series sensor data efficiently and enables data visualization and analysis in real time. Existing solutions for managing data typically use unstructured file systems or relational databases. These systems are not designed to provide instantaneous response to user queries. Furthermore, they do not support rapid data analysis and visualization to enable interactive experiments. In large scale experiments, this behaviour slows research discovery, discourages the widespread sharing and reuse of data that could otherwise inform critical decisions in a timely manner and encourage effective collaboration between groups. In this paper we present SensorDB, a web based virtual laboratory that can manage large volumes of biological time series sensor data while supporting rapid data queries and real-time user interaction. SensorDB is sensor agnostic and uses web-based, state-of-the-art cloud and storage technologies to efficiently gather, analyse and visualize data. Collaboration and data sharing between different agencies and groups is thereby facilitated. SensorDB is available online at http://sensordb.csiro.au.

  18. Occlusion assessment of intracranial aneurysms treated with the WEB device.

    PubMed

    Caroff, Jildaz; Mihalea, Cristian; Tuilier, Titien; Barreau, Xavier; Cognard, Christophe; Desal, Hubert; Pierot, Laurent; Arnoux, Armelle; Moret, Jacques; Spelle, Laurent

    2016-09-01

    The Woven EndoBridge (WEB) system is an innovative device under evaluation for its capacity to treat wide-neck bifurcation intracranial aneurysms. The purpose of this study is to evaluate the use of the different occlusion scales available in clinical practice. Seven WEB-experienced neurointerventionalists were provided with 30 angiographic follow-up data sets and asked to grade each evaluation point according to the Bicêtre Occlusion Scale Score (BOSS), firstly based on DSA images only then using additional C-Arm VasoCT analysis. This BOSS evaluation was then converted into the WEB Occlusion Scale (WOS) and into a dichotomized scale (complete occlusion or not). To estimate the inter-rater agreement among the seven raters, an overall kappa coefficient [1] and its standard error (SE) were computed. Using the five-grade BOSS, raters showed "moderate" agreement (kappa = 0.56). Using the three-grade WOS, agreement appeared slightly better (kappa = 0.59). Strongest inter-rater agreement was observed with a dichotomized version of the scale (complete occlusion or not), which enabled an "almost perfect" agreement (kappa = 0.88). VasoCT consistently enhanced the agreement particularly with regards depicting intra-WEB residual filling. The WOS is a consistent means to angiographically evaluate the WEB device efficiency. But the five-grade BOSS scale allows to identify aneurysm subgroups with differing risks of recurrence and/or rehemorrhage, which needs to be separated especially at the initial phase of evaluation of this innovative device. The additional use of VasoCT allows better inter-rater agreement in evaluating occlusion and specially in depicting intra-WEB persistent filling.

  19. Mining the Human Phenome using Semantic Web Technologies: A Case Study for Type 2 Diabetes

    PubMed Central

    Pathak, Jyotishman; Kiefer, Richard C.; Bielinski, Suzette J.; Chute, Christopher G.

    2012-01-01

    The ability to conduct genome-wide association studies (GWAS) has enabled new exploration of how genetic variations contribute to health and disease etiology. However, historically GWAS have been limited by inadequate sample size due to associated costs for genotyping and phenotyping of study subjects. This has prompted several academic medical centers to form “biobanks” where biospecimens linked to personal health information, typically in electronic health records (EHRs), are collected and stored on large number of subjects. This provides tremendous opportunities to discover novel genotype-phenotype associations and foster hypothesis generation. In this work, we study how emerging Semantic Web technologies can be applied in conjunction with clinical and genotype data stored at the Mayo Clinic Biobank to mine the phenotype data for genetic associations. In particular, we demonstrate the role of using Resource Description Framework (RDF) for representing EHR diagnoses and procedure data, and enable federated querying via standardized Web protocols to identify subjects genotyped with Type 2 Diabetes for discovering gene-disease associations. Our study highlights the potential of Web-scale data federation techniques to execute complex queries. PMID:23304343

  20. Mining the human phenome using semantic web technologies: a case study for Type 2 Diabetes.

    PubMed

    Pathak, Jyotishman; Kiefer, Richard C; Bielinski, Suzette J; Chute, Christopher G

    2012-01-01

    The ability to conduct genome-wide association studies (GWAS) has enabled new exploration of how genetic variations contribute to health and disease etiology. However, historically GWAS have been limited by inadequate sample size due to associated costs for genotyping and phenotyping of study subjects. This has prompted several academic medical centers to form "biobanks" where biospecimens linked to personal health information, typically in electronic health records (EHRs), are collected and stored on large number of subjects. This provides tremendous opportunities to discover novel genotype-phenotype associations and foster hypothesis generation. In this work, we study how emerging Semantic Web technologies can be applied in conjunction with clinical and genotype data stored at the Mayo Clinic Biobank to mine the phenotype data for genetic associations. In particular, we demonstrate the role of using Resource Description Framework (RDF) for representing EHR diagnoses and procedure data, and enable federated querying via standardized Web protocols to identify subjects genotyped with Type 2 Diabetes for discovering gene-disease associations. Our study highlights the potential of Web-scale data federation techniques to execute complex queries.

  1. The Fabric of the Universe: Exploring the Cosmic Web in 3D Prints and Woven Textiles

    NASA Astrophysics Data System (ADS)

    Diemer, Benedikt; Facio, Isaac

    2017-05-01

    We introduce The Fabric of the Universe, an art and science collaboration focused on exploring the cosmic web of dark matter with unconventional techniques and materials. We discuss two of our projects in detail. First, we describe a pipeline for translating three-dimensional (3D) density structures from N-body simulations into solid surfaces suitable for 3D printing, and present prints of a cosmological volume and of the infall region around a massive cluster halo. In these models, we discover wall-like features that are invisible in two-dimensional projections. Going beyond the sheer visualization of simulation data, we undertake an exploration of the cosmic web as a three-dimensional woven textile. To this end, we develop experimental 3D weaving techniques to create sphere-like and filamentary shapes and radically simplify a region of the cosmic web into a set of filaments and halos. We translate the resulting tree structure into a series of commands that can be executed by a digital weaving machine, and present a large-scale textile installation.

  2. TheCellMap.org: A Web-Accessible Database for Visualizing and Mining the Global Yeast Genetic Interaction Network

    PubMed Central

    Usaj, Matej; Tan, Yizhao; Wang, Wen; VanderSluis, Benjamin; Zou, Albert; Myers, Chad L.; Costanzo, Michael; Andrews, Brenda; Boone, Charles

    2017-01-01

    Providing access to quantitative genomic data is key to ensure large-scale data validation and promote new discoveries. TheCellMap.org serves as a central repository for storing and analyzing quantitative genetic interaction data produced by genome-scale Synthetic Genetic Array (SGA) experiments with the budding yeast Saccharomyces cerevisiae. In particular, TheCellMap.org allows users to easily access, visualize, explore, and functionally annotate genetic interactions, or to extract and reorganize subnetworks, using data-driven network layouts in an intuitive and interactive manner. PMID:28325812

  3. TheCellMap.org: A Web-Accessible Database for Visualizing and Mining the Global Yeast Genetic Interaction Network.

    PubMed

    Usaj, Matej; Tan, Yizhao; Wang, Wen; VanderSluis, Benjamin; Zou, Albert; Myers, Chad L; Costanzo, Michael; Andrews, Brenda; Boone, Charles

    2017-05-05

    Providing access to quantitative genomic data is key to ensure large-scale data validation and promote new discoveries. TheCellMap.org serves as a central repository for storing and analyzing quantitative genetic interaction data produced by genome-scale Synthetic Genetic Array (SGA) experiments with the budding yeast Saccharomyces cerevisiae In particular, TheCellMap.org allows users to easily access, visualize, explore, and functionally annotate genetic interactions, or to extract and reorganize subnetworks, using data-driven network layouts in an intuitive and interactive manner. Copyright © 2017 Usaj et al.

  4. SCALING AN URBAN EMERGENCY EVACUATION FRAMEWORK: CHALLENGES AND PRACTICES

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Karthik, Rajasekar; Lu, Wei

    2014-01-01

    Critical infrastructure disruption, caused by severe weather events, natural disasters, terrorist attacks, etc., has significant impacts on urban transportation systems. We built a computational framework to simulate urban transportation systems under critical infrastructure disruption in order to aid real-time emergency evacuation. This framework will use large scale datasets to provide a scalable tool for emergency planning and management. Our framework, World-Wide Emergency Evacuation (WWEE), integrates population distribution and urban infrastructure networks to model travel demand in emergency situations at global level. Also, a computational model of agent-based traffic simulation is used to provide an optimal evacuation plan for traffic operationmore » purpose [1]. In addition, our framework provides a web-based high resolution visualization tool for emergency evacuation modelers and practitioners. We have successfully tested our framework with scenarios in both United States (Alexandria, VA) and Europe (Berlin, Germany) [2]. However, there are still some major drawbacks for scaling this framework to handle big data workloads in real time. On our back-end, lack of proper infrastructure limits us in ability to process large amounts of data, run the simulation efficiently and quickly, and provide fast retrieval and serving of data. On the front-end, the visualization performance of microscopic evacuation results is still not efficient enough due to high volume data communication between server and client. We are addressing these drawbacks by using cloud computing and next-generation web technologies, namely Node.js, NoSQL, WebGL, Open Layers 3 and HTML5 technologies. We will describe briefly about each one and how we are using and leveraging these technologies to provide an efficient tool for emergency management organizations. Our early experimentation demonstrates that using above technologies is a promising approach to build a scalable and high performance urban emergency evacuation framework that can improve traffic mobility and safety under critical infrastructure disruption in today s socially connected world.« less

  5. Interval-level measurement with visual analogue scales in Internet-based research: VAS Generator.

    PubMed

    Reips, Ulf-Dietrich; Funke, Frederik

    2008-08-01

    The present article describes VAS Generator (www.vasgenerator.net), a free Web service for creating a wide range of visual analogue scales that can be used as measurement devices in Web surveys and Web experimentation, as well as for local computerized assessment. A step-by-step example for creating and implementing a visual analogue scale with visual feedback is given. VAS Generator and the scales it generates work independently of platforms and use the underlying languages HTML and JavaScript. Results from a validation study with 355 participants are reported and show that the scales generated with VAS Generator approximate an interval-scale level. In light of previous research on visual analogue versus categorical (e.g., radio button) scales in Internet-based research, we conclude that categorical scales only reach ordinal-scale level, and thus visual analogue scales are to be preferred whenever possible.

  6. A Scalable proxy cache for Grid Data Access

    NASA Astrophysics Data System (ADS)

    Cristian Cirstea, Traian; Just Keijser, Jan; Koeroo, Oscar Arthur; Starink, Ronald; Templon, Jeffrey Alan

    2012-12-01

    We describe a prototype grid proxy cache system developed at Nikhef, motivated by a desire to construct the first building block of a future https-based Content Delivery Network for grid infrastructures. Two goals drove the project: firstly to provide a “native view” of the grid for desktop-type users, and secondly to improve performance for physics-analysis type use cases, where multiple passes are made over the same set of data (residing on the grid). We further constrained the design by requiring that the system should be made of standard components wherever possible. The prototype that emerged from this exercise is a horizontally-scalable, cooperating system of web server / cache nodes, fronted by a customized webDAV server. The webDAV server is custom only in the sense that it supports http redirects (providing horizontal scaling) and that the authentication module has, as back end, a proxy delegation chain that can be used by the cache nodes to retrieve files from the grid. The prototype was deployed at Nikhef and tested at a scale of several terabytes of data and approximately one hundred fast cores of computing. Both small and large files were tested, in a number of scenarios, and with various numbers of cache nodes, in order to understand the scaling properties of the system. For properly-dimensioned cache-node hardware, the system showed speedup of several integer factors for the analysis-type use cases. These results and others are presented and discussed.

  7. Accelerating Software Development through Agile Practices--A Case Study of a Small-Scale, Time-Intensive Web Development Project at a College-Level IT Competition

    ERIC Educational Resources Information Center

    Zhang, Xuesong; Dorn, Bradley

    2012-01-01

    Agile development has received increasing interest both in industry and academia due to its benefits in developing software quickly, meeting customer needs, and keeping pace with the rapidly changing requirements. However, agile practices and scrum in particular have been mainly tested in mid- to large-size projects. In this paper, we present…

  8. Introduction to CERN

    ScienceCinema

    Heuer, R.-D.

    2018-02-19

    Summer Student Lecture Programme Introduction. The mission of CERN; push back the frontiers of knowledge, e.g. the secrets of the Big Bang...what was the matter like within the first moments of the Universe's existence? You have to develop new technologies for accelerators and detectors (also information technology--the Web and the GRID and medicine--diagnosis and therapy). There are three key technology areas at CERN; accelerating, particle detection, large-scale computing.

  9. Behind the Scenes with OpenLearn: The Challenges of Researching the Provision of Open Educational Resources

    ERIC Educational Resources Information Center

    Godwin, Stephen; McAndrew, Patrick; Santos, Andreia

    2008-01-01

    Web-enabled technology is now being applied on a large scale. In this paper we look at open access provision of teaching and learning leading to many users with varying patterns and motivations for use. This has provided us with a research challenge to find methods that help us understand and explain such initiatives. We describe ways to model the…

  10. CrosstalkNet: A Visualization Tool for Differential Co-expression Networks and Communities.

    PubMed

    Manem, Venkata; Adam, George Alexandru; Gruosso, Tina; Gigoux, Mathieu; Bertos, Nicholas; Park, Morag; Haibe-Kains, Benjamin

    2018-04-15

    Variations in physiological conditions can rewire molecular interactions between biological compartments, which can yield novel insights into gain or loss of interactions specific to perturbations of interest. Networks are a promising tool to elucidate intercellular interactions, yet exploration of these large-scale networks remains a challenge due to their high dimensionality. To retrieve and mine interactions, we developed CrosstalkNet, a user friendly, web-based network visualization tool that provides a statistical framework to infer condition-specific interactions coupled with a community detection algorithm for bipartite graphs to identify significantly dense subnetworks. As a case study, we used CrosstalkNet to mine a set of 54 and 22 gene-expression profiles from breast tumor and normal samples, respectively, with epithelial and stromal compartments extracted via laser microdissection. We show how CrosstalkNet can be used to explore large-scale co-expression networks and to obtain insights into the biological processes that govern cross-talk between different tumor compartments. Significance: This web application enables researchers to mine complex networks and to decipher novel biological processes in tumor epithelial-stroma cross-talk as well as in other studies of intercompartmental interactions. Cancer Res; 78(8); 2140-3. ©2018 AACR . ©2018 American Association for Cancer Research.

  11. -A curated transcriptomic dataset collection relevant to embryonic development associated with in vitro fertilization in healthy individuals and patients with polycystic ovary syndrome.

    PubMed

    Mackeh, Rafah; Boughorbel, Sabri; Chaussabel, Damien; Kino, Tomoshige

    2017-01-01

    The collection of large-scale datasets available in public repositories is rapidly growing and providing opportunities to identify and fill gaps in different fields of biomedical research. However, users of these datasets should be able to selectively browse datasets related to their field of interest. Here we made available a collection of transcriptome datasets related to human follicular cells from normal individuals or patients with polycystic ovary syndrome, in the process of their development, during in vitro fertilization. After RNA-seq dataset exclusion and careful selection based on study description and sample information, 12 datasets, encompassing a total of 85 unique transcriptome profiles, were identified in NCBI Gene Expression Omnibus and uploaded to the Gene Expression Browser (GXB), a web application specifically designed for interactive query and visualization of integrated large-scale data. Once annotated in GXB, multiple sample grouping has been made in order to create rank lists to allow easy data interpretation and comparison. The GXB tool also allows the users to browse a single gene across multiple projects to evaluate its expression profiles in multiple biological systems/conditions in a web-based customized graphical views. The curated dataset is accessible at the following link: http://ivf.gxbsidra.org/dm3/landing.gsp.

  12. ­A curated transcriptomic dataset collection relevant to embryonic development associated with in vitro fertilization in healthy individuals and patients with polycystic ovary syndrome

    PubMed Central

    Mackeh, Rafah; Boughorbel, Sabri; Chaussabel, Damien; Kino, Tomoshige

    2017-01-01

    The collection of large-scale datasets available in public repositories is rapidly growing and providing opportunities to identify and fill gaps in different fields of biomedical research. However, users of these datasets should be able to selectively browse datasets related to their field of interest. Here we made available a collection of transcriptome datasets related to human follicular cells from normal individuals or patients with polycystic ovary syndrome, in the process of their development, during in vitro fertilization. After RNA-seq dataset exclusion and careful selection based on study description and sample information, 12 datasets, encompassing a total of 85 unique transcriptome profiles, were identified in NCBI Gene Expression Omnibus and uploaded to the Gene Expression Browser (GXB), a web application specifically designed for interactive query and visualization of integrated large-scale data. Once annotated in GXB, multiple sample grouping has been made in order to create rank lists to allow easy data interpretation and comparison. The GXB tool also allows the users to browse a single gene across multiple projects to evaluate its expression profiles in multiple biological systems/conditions in a web-based customized graphical views. The curated dataset is accessible at the following link: http://ivf.gxbsidra.org/dm3/landing.gsp. PMID:28413616

  13. WikiPEATia - a web based platform for assembling peatland data through ‘crowd sourcing’

    NASA Astrophysics Data System (ADS)

    Wisser, D.; Glidden, S.; Fieseher, C.; Treat, C. C.; Routhier, M.; Frolking, S. E.

    2009-12-01

    The Earth System Science community is realizing that peatlands are an important and unique terrestrial ecosystem that has not yet been well-integrated into large-scale earth system analyses. A major hurdle is the lack of accessible, geospatial data of peatland distribution, coupled with data on peatland properties (e.g., vegetation composition, peat depth, basal dates, soil chemistry, peatland class) at the global scale. This data, however, is available at the local scale. Although a comprehensive global database on peatlands probably lags similar data on more economically important ecosystems such as forests, grasslands, croplands, a large amount of field data have been collected over the past several decades. A few efforts have been made to map peatlands at large scales but existing data have not been assembled into a single geospatial database that is publicly accessible or do not depict data with a level of detail that is needed in the Earth System Science Community. A global peatland database would contribute to advances in a number of research fields such as hydrology, vegetation and ecosystem modeling, permafrost modeling, and earth system modeling. We present a Web 2.0 approach that uses state-of-the-art webserver and innovative online mapping technologies and is designed to create such a global database through ‘crowd-sourcing’. Primary functions of the online system include form-driven textual user input of peatland research metadata, spatial data input of peatland areas via a mapping interface, database editing and querying editing capabilities, as well as advanced visualization and data analysis tools. WikiPEATia provides an integrated information technology platform for assembling, integrating, and posting peatland-related geospatial datasets facilitates and encourages research community involvement. A successful effort will make existing peatland data much more useful to the research community, and will help to identify significant data gaps.

  14. Compilation and network analyses of cambrian food webs.

    PubMed

    Dunne, Jennifer A; Williams, Richard J; Martinez, Neo D; Wood, Rachel A; Erwin, Douglas H

    2008-04-29

    A rich body of empirically grounded theory has developed about food webs--the networks of feeding relationships among species within habitats. However, detailed food-web data and analyses are lacking for ancient ecosystems, largely because of the low resolution of taxa coupled with uncertain and incomplete information about feeding interactions. These impediments appear insurmountable for most fossil assemblages; however, a few assemblages with excellent soft-body preservation across trophic levels are candidates for food-web data compilation and topological analysis. Here we present plausible, detailed food webs for the Chengjiang and Burgess Shale assemblages from the Cambrian Period. Analyses of degree distributions and other structural network properties, including sensitivity analyses of the effects of uncertainty associated with Cambrian diet designations, suggest that these early Paleozoic communities share remarkably similar topology with modern food webs. Observed regularities reflect a systematic dependence of structure on the numbers of taxa and links in a web. Most aspects of Cambrian food-web structure are well-characterized by a simple "niche model," which was developed for modern food webs and takes into account this scale dependence. However, a few aspects of topology differ between the ancient and recent webs: longer path lengths between species and more species in feeding loops in the earlier Chengjiang web, and higher variability in the number of links per species for both Cambrian webs. Our results are relatively insensitive to the exclusion of low-certainty or random links. The many similarities between Cambrian and recent food webs point toward surprisingly strong and enduring constraints on the organization of complex feeding interactions among metazoan species. The few differences could reflect a transition to more strongly integrated and constrained trophic organization within ecosystems following the rapid diversification of species, body plans, and trophic roles during the Cambrian radiation. More research is needed to explore the generality of food-web structure through deep time and across habitats, especially to investigate potential mechanisms that could give rise to similar structure, as well as any differences.

  15. Food-web structure and ecosystem services: insights from the Serengeti.

    PubMed

    Dobson, Andy

    2009-06-27

    The central organizing theme of this paper is to discuss the dynamics of the Serengeti grassland ecosystem from the perspective of recent developments in food-web theory. The seasonal rainfall patterns that characterize the East African climate create an annually oscillating, large-scale, spatial mosaic of feeding opportunities for the larger ungulates in the Serengeti; this in turn creates a significant annual variation in the food available for their predators. At a smaller spatial scale, periodic fires during the dry season create patches of highly nutritious grazing that are eaten in preference to the surrounding older patches of less palatable vegetation. The species interactions between herbivores and plants, and carnivores and herbivores, are hierarchically nested in the Serengeti food web, with the largest bodied consumers on each trophic level having the broadest diets that include species from a large variety of different habitats in the ecosystem. The different major habitats of the Serengeti are also used in a nested fashion; the highly nutritious forage of the short grass plains is available only to the larger migratory species for a few months each year. The longer grass areas, the woodlands and kopjes (large partially wooded rocky islands in the surrounding mosaic of grassland) contain species that are resident throughout the year; these species often have smaller body size and more specialized diets than the migratory species. Only the larger herbivores and carnivores obtain their nutrition from all the different major habitat types in the ecosystem. The net effect of this is to create a nested hierarchy of subchains of energy flow within the larger Serengeti food web; these flows are seasonally forced by rainfall and operate at different rates in different major branches of the web. The nested structure that couples sequential trophic levels together interacts with annual seasonal variation in the fast and slow chains of nutrient flow in a way that is likely to be central to the stability of the whole web. If the Serengeti is to be successfully conserved as a fully functioning ecosystem, then it is essential that the full diversity of natural habitats be maintained within the greater Serengeti ecosystem. The best way to do this is by controlling the external forces that threaten the boundaries of the ecosystem and by balancing the economic services the park provides between local, national and international needs. I conclude by discussing how the ecosystem services provided by the Serengeti are driven by species on different trophic levels. Tourism provides the largest financial revenue to the national economy, but it could be better organized to provide more sustained revenue to the park. Ultimately, ecotourism needs to be developed in ways that take lessons from the structure of the Serengeti food webs, and in ways that provide tangible benefits to people living around the park while also improving the experience of all visitors.

  16. Food-web structure and ecosystem services: insights from the Serengeti

    PubMed Central

    Dobson, Andy

    2009-01-01

    The central organizing theme of this paper is to discuss the dynamics of the Serengeti grassland ecosystem from the perspective of recent developments in food-web theory. The seasonal rainfall patterns that characterize the East African climate create an annually oscillating, large-scale, spatial mosaic of feeding opportunities for the larger ungulates in the Serengeti; this in turn creates a significant annual variation in the food available for their predators. At a smaller spatial scale, periodic fires during the dry season create patches of highly nutritious grazing that are eaten in preference to the surrounding older patches of less palatable vegetation. The species interactions between herbivores and plants, and carnivores and herbivores, are hierarchically nested in the Serengeti food web, with the largest bodied consumers on each trophic level having the broadest diets that include species from a large variety of different habitats in the ecosystem. The different major habitats of the Serengeti are also used in a nested fashion; the highly nutritious forage of the short grass plains is available only to the larger migratory species for a few months each year. The longer grass areas, the woodlands and kopjes (large partially wooded rocky islands in the surrounding mosaic of grassland) contain species that are resident throughout the year; these species often have smaller body size and more specialized diets than the migratory species. Only the larger herbivores and carnivores obtain their nutrition from all the different major habitat types in the ecosystem. The net effect of this is to create a nested hierarchy of subchains of energy flow within the larger Serengeti food web; these flows are seasonally forced by rainfall and operate at different rates in different major branches of the web. The nested structure that couples sequential trophic levels together interacts with annual seasonal variation in the fast and slow chains of nutrient flow in a way that is likely to be central to the stability of the whole web. If the Serengeti is to be successfully conserved as a fully functioning ecosystem, then it is essential that the full diversity of natural habitats be maintained within the greater Serengeti ecosystem. The best way to do this is by controlling the external forces that threaten the boundaries of the ecosystem and by balancing the economic services the park provides between local, national and international needs. I conclude by discussing how the ecosystem services provided by the Serengeti are driven by species on different trophic levels. Tourism provides the largest financial revenue to the national economy, but it could be better organized to provide more sustained revenue to the park. Ultimately, ecotourism needs to be developed in ways that take lessons from the structure of the Serengeti food webs, and in ways that provide tangible benefits to people living around the park while also improving the experience of all visitors. PMID:19451118

  17. The utility of web mining for epidemiological research: studying the association between parity and cancer risk [Web Mining for Epidemiological Research. Assessing its Utility in Exploring the Association Between Parity and Cancer Risk

    DOE PAGES

    Tourassi, Georgia; Yoon, Hong-Jun; Xu, Songhua; ...

    2015-11-27

    Background: The World Wide Web has emerged as a powerful data source for epidemiological studies related to infectious disease surveillance. However, its potential for cancer-related epidemiological discoveries is largely unexplored. Methods: Using advanced web crawling and tailored information extraction procedures we automatically collected and analyzed the text content of 79,394 online obituary articles published between 1998 and 2014. The collected data included 51,911 cancer (27,330 breast; 9,470 lung; 6,496 pancreatic; 6,342 ovarian; 2,273 colon) and 27,483 non-cancer cases. With the derived information, we replicated a case-control study design to investigate the association between parity and cancer risk. Age-adjusted odds ratiosmore » (ORs) with 95% confidence intervals (CIs) were calculated for each cancer type and compared to those reported in large-scale epidemiological studies. Results: Parity was found to be associated with a significantly reduced risk of breast cancer (OR=0.78, 95% CI = 0.75 to 0.82), pancreatic cancer (OR=0.78, 95% CI = 0.72 to 0.83), colon cancer (OR=0.67, 95% CI = 0.60 to 0.74), and ovarian cancer (OR=0.58, 95% CI = 0.54 to 0.62). Marginal association was found for lung cancer prevalence (OR=0.87, 95% CI = 0.81 to 0.92). The linear trend between multi-parity and reduced cancer risk was dramatically more pronounced for breast and ovarian cancer than the other cancers included in the analysis. Conclusion: This large web-mining study on parity and cancer risk produced findings very similar to those reported with traditional observational studies. It may be used as a promising strategy to generate study hypotheses for guiding and prioritizing future epidemiological studies.« less

  18. The utility of web mining for epidemiological research: studying the association between parity and cancer risk [Web Mining for Epidemiological Research. Assessing its Utility in Exploring the Association Between Parity and Cancer Risk

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Tourassi, Georgia; Yoon, Hong-Jun; Xu, Songhua

    Background: The World Wide Web has emerged as a powerful data source for epidemiological studies related to infectious disease surveillance. However, its potential for cancer-related epidemiological discoveries is largely unexplored. Methods: Using advanced web crawling and tailored information extraction procedures we automatically collected and analyzed the text content of 79,394 online obituary articles published between 1998 and 2014. The collected data included 51,911 cancer (27,330 breast; 9,470 lung; 6,496 pancreatic; 6,342 ovarian; 2,273 colon) and 27,483 non-cancer cases. With the derived information, we replicated a case-control study design to investigate the association between parity and cancer risk. Age-adjusted odds ratiosmore » (ORs) with 95% confidence intervals (CIs) were calculated for each cancer type and compared to those reported in large-scale epidemiological studies. Results: Parity was found to be associated with a significantly reduced risk of breast cancer (OR=0.78, 95% CI = 0.75 to 0.82), pancreatic cancer (OR=0.78, 95% CI = 0.72 to 0.83), colon cancer (OR=0.67, 95% CI = 0.60 to 0.74), and ovarian cancer (OR=0.58, 95% CI = 0.54 to 0.62). Marginal association was found for lung cancer prevalence (OR=0.87, 95% CI = 0.81 to 0.92). The linear trend between multi-parity and reduced cancer risk was dramatically more pronounced for breast and ovarian cancer than the other cancers included in the analysis. Conclusion: This large web-mining study on parity and cancer risk produced findings very similar to those reported with traditional observational studies. It may be used as a promising strategy to generate study hypotheses for guiding and prioritizing future epidemiological studies.« less

  19. AdS and dS Entropy from String Junctions or The Function of Junction Conjunctions

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Silverstein, Eva M

    Flux compactifications of string theory exhibiting the possibility of discretely tuning the cosmological constant to small values have been constructed. The highly tuned vacua in this discretuum have curvature radii which scale as large powers of the flux quantum numbers, exponential in the number of cycles in the compactification. By the arguments of Susskind/Witten (in the AdS case) and Gibbons/Hawking (in the dS case), we expect correspondingly large entropies associated with these vacua. If they are to provide a dual description of these vacua on their Coulomb branch, branes traded for the flux need to account for this entropy atmore » the appropriate energy scale. In this note, we argue that simple string junctions and webs ending on the branes can account for this large entropy, obtaining a rough estimate for junction entropy that agrees with the existing rough estimates for the spacing of the discretuum. In particular, the brane entropy can account for the (A)dS entropy far away from string scale correspondence limits.« less

  20. Sign: large-scale gene network estimation environment for high performance computing.

    PubMed

    Tamada, Yoshinori; Shimamura, Teppei; Yamaguchi, Rui; Imoto, Seiya; Nagasaki, Masao; Miyano, Satoru

    2011-01-01

    Our research group is currently developing software for estimating large-scale gene networks from gene expression data. The software, called SiGN, is specifically designed for the Japanese flagship supercomputer "K computer" which is planned to achieve 10 petaflops in 2012, and other high performance computing environments including Human Genome Center (HGC) supercomputer system. SiGN is a collection of gene network estimation software with three different sub-programs: SiGN-BN, SiGN-SSM and SiGN-L1. In these three programs, five different models are available: static and dynamic nonparametric Bayesian networks, state space models, graphical Gaussian models, and vector autoregressive models. All these models require a huge amount of computational resources for estimating large-scale gene networks and therefore are designed to be able to exploit the speed of 10 petaflops. The software will be available freely for "K computer" and HGC supercomputer system users. The estimated networks can be viewed and analyzed by Cell Illustrator Online and SBiP (Systems Biology integrative Pipeline). The software project web site is available at http://sign.hgc.jp/ .

  1. The Vainshtein mechanism in the cosmic web

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Falck, Bridget; Koyama, Kazuya; Zhao, Gong-bo

    We investigate the dependence of the Vainshtein screening mechanism on the cosmic web morphology of both dark matter particles and halos as determined by ORIGAMI. Unlike chameleon and symmetron screening, which come into effect in regions of high density, Vainshtein screening instead depends on the dimensionality of the system, and screened bodies can still feel external fields. ORIGAMI is well-suited to this problem because it defines morphologies according to the dimensionality of the collapsing structure and does not depend on a smoothing scale or density threshold parameter. We find that halo particles are screened while filament, wall, and void particlesmore » are unscreened, and this is independent of the particle density. However, after separating halos according to their large scale cosmic web environment, we find no difference in the screening properties of halos in filaments versus halos in clusters. We find that the fifth force enhancement of dark matter particles in halos is greatest well outside the virial radius. We confirm the theoretical expectation that even if the internal field is suppressed by the Vainshtein mechanism, the object still feels the fifth force generated by the external fields, by measuring peculiar velocities and velocity dispersions of halos. Finally, we investigate the morphology and gravity model dependence of halo spins, concentrations, and shapes.« less

  2. Biodiversity and ecosystem functioning in evolving food webs.

    PubMed

    Allhoff, K T; Drossel, B

    2016-05-19

    We use computer simulations in order to study the interplay between biodiversity and ecosystem functioning (BEF) during both the formation and the ongoing evolution of large food webs. A species in our model is characterized by its own body mass, its preferred prey body mass and the width of its potential prey body mass spectrum. On an ecological time scale, population dynamics determines which species are viable and which ones go extinct. On an evolutionary time scale, new species emerge as modifications of existing ones. The network structure thus emerges and evolves in a self-organized manner. We analyse the relation between functional diversity and five community level measures of ecosystem functioning. These are the metabolic loss of the predator community, the total biomasses of the basal and the predator community, and the consumption rates on the basal community and within the predator community. Clear BEF relations are observed during the initial build-up of the networks, or when parameters are varied, causing bottom-up or top-down effects. However, ecosystem functioning measures fluctuate only very little during long-term evolution under constant environmental conditions, despite changes in functional diversity. This result supports the hypothesis that trophic cascades are weaker in more complex food webs. © 2016 The Author(s).

  3. Comparative ecology of widely distributed pelagic fish species in the North Atlantic: Implications for modelling climate and fisheries impacts

    NASA Astrophysics Data System (ADS)

    Trenkel, V. M.; Huse, G.; MacKenzie, B. R.; Alvarez, P.; Arrizabalaga, H.; Castonguay, M.; Goñi, N.; Grégoire, F.; Hátún, H.; Jansen, T.; Jacobsen, J. A.; Lehodey, P.; Lutcavage, M.; Mariani, P.; Melvin, G. D.; Neilson, J. D.; Nøttestad, L.; Óskarsson, G. J.; Payne, M. R.; Richardson, D. E.; Senina, I.; Speirs, D. C.

    2014-12-01

    This paper reviews the current knowledge on the ecology of widely distributed pelagic fish stocks in the North Atlantic basin with emphasis on their role in the food web and the factors determining their relationship with the environment. We consider herring (Clupea harengus), mackerel (Scomber scombrus), capelin (Mallotus villosus), blue whiting (Micromesistius poutassou), and horse mackerel (Trachurus trachurus), which have distributions extending beyond the continental shelf and predominantly occur on both sides of the North Atlantic. We also include albacore (Thunnus alalunga), bluefin tuna (Thunnus thynnus), swordfish (Xiphias gladius), and blue marlin (Makaira nigricans), which, by contrast, show large-scale migrations at the basin scale. We focus on the links between life history processes and the environment, horizontal and vertical distribution, spatial structure and trophic role. Many of these species carry out extensive migrations from spawning grounds to nursery and feeding areas. Large oceanographic features such as the North Atlantic subpolar gyre play an important role in determining spatial distributions and driving variations in stock size. Given the large biomasses of especially the smaller species considered here, these stocks can exert significant top-down pressures on the food web and are important in supporting higher trophic levels. The review reveals commonalities and differences between the ecology of widely distributed pelagic fish in the NE and NW Atlantic basins, identifies knowledge gaps and modelling needs that the EURO-BASIN project attempts to address.

  4. The utility of web mining for epidemiological research: studying the association between parity and cancer risk

    PubMed Central

    Yoon, Hong-Jun; Xu, Songhua; Han, Xuesong

    2016-01-01

    Background The World Wide Web has emerged as a powerful data source for epidemiological studies related to infectious disease surveillance. However, its potential for cancer-related epidemiological discoveries is largely unexplored. Methods Using advanced web crawling and tailored information extraction procedures, the authors automatically collected and analyzed the text content of 79 394 online obituary articles published between 1998 and 2014. The collected data included 51 911 cancer (27 330 breast; 9470 lung; 6496 pancreatic; 6342 ovarian; 2273 colon) and 27 483 non-cancer cases. With the derived information, the authors replicated a case-control study design to investigate the association between parity (i.e., childbearing) and cancer risk. Age-adjusted odds ratios (ORs) with 95% confidence intervals (CIs) were calculated for each cancer type and compared to those reported in large-scale epidemiological studies. Results Parity was found to be associated with a significantly reduced risk of breast cancer (OR = 0.78, 95% CI, 0.75-0.82), pancreatic cancer (OR = 0.78, 95% CI, 0.72-0.83), colon cancer (OR = 0.67, 95% CI, 0.60-0.74), and ovarian cancer (OR = 0.58, 95% CI, 0.54-0.62). Marginal association was found for lung cancer risk (OR = 0.87, 95% CI, 0.81-0.92). The linear trend between increased parity and reduced cancer risk was dramatically more pronounced for breast and ovarian cancer than the other cancers included in the analysis. Conclusion This large web-mining study on parity and cancer risk produced findings very similar to those reported with traditional observational studies. It may be used as a promising strategy to generate study hypotheses for guiding and prioritizing future epidemiological studies. PMID:26615183

  5. Evaluation of a Web-Based E-Learning Platform for Brief Motivational Interviewing by Nurses in Cardiovascular Care: A Pilot Study

    PubMed Central

    Cossette, Sylvie; Heppell, Sonia; Boyer, Louise; Mailhot, Tanya; Simard, Marie-Josée; Tanguay, Jean-Francois

    2016-01-01

    Background Brief motivational interviewing (MI) can contribute to reductions in morbidity and mortality related to coronary artery disease, through health behavior change. Brief MI, unlike more intensive interventions, was proposed to meet the needs of clinicians with little spare time. While the provision of face-to-face brief MI training on a large scale is complicated, Web-based e-learning is promising because of the flexibility it offers. Objective The primary objective of this pilot study was to examine the feasibility and acceptability of a Web-based e-learning platform for brief MI (MOTIV@CŒUR), which was evaluated by nurses in cardiovascular care. The secondary objective was to assess the preliminary effect of the training on nurses’ perceived brief MI skills and self-reported clinical use of brief MI. Methods We conducted a single-group, pre-post pilot study involving nurses working in a coronary care unit to evaluate MOTIV@CŒUR, which is a Web-based e-learning platform for brief MI, consisting of two sessions lasting 30 and 20 minutes. MOTIV@CŒUR covers 4 real-life clinical situations through role-modeling videos showing nurse-client interactions. A brief introduction to MI is followed by role playing, during which a nurse practitioner evaluates clients’ motivation to change and intervenes according to the principles of brief MI. The clinical situations target smoking, medication adherence, physical activity, and diet. Nurses were asked to complete both Web-based training sessions asynchronously within 20 days, which allowed assessment of the feasibility of the intervention. Data regarding acceptability and preliminary effects (perceived skills in brief MI, and self-reported clinical use of conviction and confidence interventions) were self-assessed through Web-based questionnaires 30 days (±5 days) after the first session. Results We enrolled 27 women and 4 men (mean age 37, SD 9 years) in March 2016. Of the 31 participants, 24 (77%, 95% CI 63%–91%) completed both sessions in ≤20 days. At 30 days, 28 of the 31 participants (90%) had completed at least one session. The training was rated as highly acceptable, with the highest scores observed for information quality (mean 6.26, SD 0.60; scale 0–7), perceived ease of use (mean 6.16, SD 0.78; scale 0–7), and system quality (mean 6.15, SD 0.58; scale 0–7). Posttraining scores for self-reported clinical use of confidence interventions were higher than pretraining scores (mean 34.72, SD 6.29 vs mean 31.48, SD 6.75, respectively; P=.03; scale 10–50). Other results were nonsignificant. Conclusions Brief MI training using a Web-based e-learning platform including role-modeling videos is both feasible and acceptable according to cardiovascular care nurses. Further research is required to evaluate the e-learning platform in a randomized controlled trial. Trial Registration International Standard Randomized Controlled Trial Number (ISRCTN): 16510888; http://www.isrctn.com/ISRCTN16510888 (Archived by WebCite at http://www.webcitation.org/6jf7dr7bx) PMID:27539960

  6. Evaluation of a Web-Based E-Learning Platform for Brief Motivational Interviewing by Nurses in Cardiovascular Care: A Pilot Study.

    PubMed

    Fontaine, Guillaume; Cossette, Sylvie; Heppell, Sonia; Boyer, Louise; Mailhot, Tanya; Simard, Marie-Josée; Tanguay, Jean-Francois

    2016-08-18

    Brief motivational interviewing (MI) can contribute to reductions in morbidity and mortality related to coronary artery disease, through health behavior change. Brief MI, unlike more intensive interventions, was proposed to meet the needs of clinicians with little spare time. While the provision of face-to-face brief MI training on a large scale is complicated, Web-based e-learning is promising because of the flexibility it offers. The primary objective of this pilot study was to examine the feasibility and acceptability of a Web-based e-learning platform for brief MI (MOTIV@CŒUR), which was evaluated by nurses in cardiovascular care. The secondary objective was to assess the preliminary effect of the training on nurses' perceived brief MI skills and self-reported clinical use of brief MI. We conducted a single-group, pre-post pilot study involving nurses working in a coronary care unit to evaluate MOTIV@CŒUR, which is a Web-based e-learning platform for brief MI, consisting of two sessions lasting 30 and 20 minutes. MOTIV@CŒUR covers 4 real-life clinical situations through role-modeling videos showing nurse-client interactions. A brief introduction to MI is followed by role playing, during which a nurse practitioner evaluates clients' motivation to change and intervenes according to the principles of brief MI. The clinical situations target smoking, medication adherence, physical activity, and diet. Nurses were asked to complete both Web-based training sessions asynchronously within 20 days, which allowed assessment of the feasibility of the intervention. Data regarding acceptability and preliminary effects (perceived skills in brief MI, and self-reported clinical use of conviction and confidence interventions) were self-assessed through Web-based questionnaires 30 days (±5 days) after the first session. We enrolled 27 women and 4 men (mean age 37, SD 9 years) in March 2016. Of the 31 participants, 24 (77%, 95% CI 63%-91%) completed both sessions in ≤20 days. At 30 days, 28 of the 31 participants (90%) had completed at least one session. The training was rated as highly acceptable, with the highest scores observed for information quality (mean 6.26, SD 0.60; scale 0-7), perceived ease of use (mean 6.16, SD 0.78; scale 0-7), and system quality (mean 6.15, SD 0.58; scale 0-7). Posttraining scores for self-reported clinical use of confidence interventions were higher than pretraining scores (mean 34.72, SD 6.29 vs mean 31.48, SD 6.75, respectively; P=.03; scale 10-50). Other results were nonsignificant. Brief MI training using a Web-based e-learning platform including role-modeling videos is both feasible and acceptable according to cardiovascular care nurses. Further research is required to evaluate the e-learning platform in a randomized controlled trial. International Standard Randomized Controlled Trial Number (ISRCTN): 16510888; http://www.isrctn.com/ISRCTN16510888 (Archived by WebCite at http://www.webcitation.org/6jf7dr7bx).

  7. Surfing for scoliosis: the quality of information available on the Internet.

    PubMed

    Mathur, Sameer; Shanti, Nael; Brkaric, Mario; Sood, Vivek; Kubeck, Justin; Paulino, Carl; Merola, Andrew A

    2005-12-01

    A cross section of Web sites accessible to the general public was surveyed. To evaluate the quality and accuracy of information on scoliosis that a patient might access on the Internet. The Internet is a rapidly expanding communications network with an estimated 765 million users worldwide by the year 2005. Medical information is one of the most common sources of inquires on the Web. More than 100 million Americans accessed the Internet for medical information in the year 2000. Undoubtedly, the use of the Internet for patient information needs will continue to expand as Internet access becomes more readily available. This expansion combined with the Internet's poorly regulated format can lead to problems in the quality of information available. Since the Internet operates on a global scale, implementing and enforcing standards have been difficult. The largely uncontrolled information can potentially negatively influence consumer health outcomes. To identify potential sites, five search engines were selected and the word "scoliosis" was entered into each search engine. A total of 50 Web sites were chosen for review. Each Web site was evaluated according to the type of Web site, quality content, and informational accuracy by three board-certified academic orthopedic surgeons, fellowship trained in spinal surgery, who each has been in practice for a minimum of 8 years. Each Web site was categorized as academic, commercial, physician, nonphysician health professional, and unidentified. In addition, each Web site was evaluated according to scoliosis-specific content using a point value system of 32 disease-specific key words pertinent to the care of scoliosis on an ordinal scale. A list of these words is given. Point values were given for the use of key words related to disease summary, classifications, treatment options, and complications. The accuracy of the individual Web site was evaluated by each spine surgeon using a scale of 1 to 4. A score of 1 represents that the examiner agreed with less than 25% of the information while a score of 4 represents greater than 75% agreement. Of the total 50 Web sites evaluated, 44% were academic, 18% were physician based, 16% were commercial, 12% were unidentified, and 10% were nonphysician health professionals. The quality content score (maximum, 32 points) for academic sites was 12.6 +/- 3.8, physician sites 11.3 +/- 4.0, commercial sites 11 +/- 4.2, unidentified 7.6 +/- 3.9, and nonphysician health professional site 7.0 +/- 1.8. The accuracy score (maximum, 12 points) was 6.6 +/- 2.4 for academic sites, 6.3 +/- 3.0 for physician-professional sites, 6.0 +/- 2.7 for unidentified sites, 5.5 +/- 3.8 for nonphysician professional sites, and 5.0 +/- 1.5 for commercial Web sites. The academic Web sites had the highest mean scores in both quality and accuracy content scores. The information about scoliosis on the Internet is of limited quality and poor information value. Although the majority of the Web sites were academic, the content quality and accuracy scores were still poor. The lowest scoring Web sites were the nonphysician professionals and the unidentified sites, which were often message boards. Overall, the highest scoring Web site related to both quality and accuracy of information was www.srs.org. This Web site was designed by the Scoliosis Research Society. The public and the medical communities need to be aware of these existing limitations of the Internet. Based on our review, the physician must assume primary responsibility of educating and counseling their patients.

  8. SOURCE EXPLORER: Towards Web Browser Based Tools for Astronomical Source Visualization and Analysis

    NASA Astrophysics Data System (ADS)

    Young, M. D.; Hayashi, S.; Gopu, A.

    2014-05-01

    As a new generation of large format, high-resolution imagers come online (ODI, DECAM, LSST, etc.) we are faced with the daunting prospect of astronomical images containing upwards of hundreds of thousands of identifiable sources. Visualizing and interacting with such large datasets using traditional astronomical tools appears to be unfeasible, and a new approach is required. We present here a method for the display and analysis of arbitrarily large source datasets using dynamically scaling levels of detail, enabling scientists to rapidly move from large-scale spatial overviews down to the level of individual sources and everything in-between. Based on the recognized standards of HTML5+JavaScript, we enable observers and archival users to interact with their images and sources from any modern computer without having to install specialized software. We demonstrate the ability to produce large-scale source lists from the images themselves, as well as overlaying data from publicly available source ( 2MASS, GALEX, SDSS, etc.) or user provided source lists. A high-availability cluster of computational nodes allows us to produce these source maps on demand and customized based on user input. User-generated source lists and maps are persistent across sessions and are available for further plotting, analysis, refinement, and culling.

  9. Low cost, scalable proteomics data analysis using Amazon's cloud computing services and open source search algorithms.

    PubMed

    Halligan, Brian D; Geiger, Joey F; Vallejos, Andrew K; Greene, Andrew S; Twigger, Simon N

    2009-06-01

    One of the major difficulties for many laboratories setting up proteomics programs has been obtaining and maintaining the computational infrastructure required for the analysis of the large flow of proteomics data. We describe a system that combines distributed cloud computing and open source software to allow laboratories to set up scalable virtual proteomics analysis clusters without the investment in computational hardware or software licensing fees. Additionally, the pricing structure of distributed computing providers, such as Amazon Web Services, allows laboratories or even individuals to have large-scale computational resources at their disposal at a very low cost per run. We provide detailed step-by-step instructions on how to implement the virtual proteomics analysis clusters as well as a list of current available preconfigured Amazon machine images containing the OMSSA and X!Tandem search algorithms and sequence databases on the Medical College of Wisconsin Proteomics Center Web site ( http://proteomics.mcw.edu/vipdac ).

  10. TOGA COARE Satellite data summaries available on the World Wide Web

    NASA Technical Reports Server (NTRS)

    Chen, S. S.; Houze, R. A., Jr.; Mapes, B. E.; Brodzick, S. R.; Yutler, S. E.

    1995-01-01

    Satellite data summary images and analysis plots from the Tropical Ocean Global Atmosphere Coupled Ocean-Atmosphere Response Experiment (TOGA COARE), which were initially prepared in the field at the Honiara Operations Center, are now available on the Internet via World Wide Web browsers such as Mosaic. These satellite data summaries consist of products derived from the Japanese Geosynchronous Meteorological Satellite IR data: a time-size series of the distribution of contiguous cold cloudiness areas, weekly percent high cloudiness (PHC) maps, and a five-month time-longitudinal diagram illustrating the zonal motion of large areas of cold cloudiness. The weekly PHC maps are overlaid with weekly mean 850-hPa wind calculated from the European Centre for Medium-Range Weather Forecasts (ECMWF) global analysis field and can be viewed as an animation loop. These satellite summaries provide an overview of spatial and temporal variabilities of the cloud population and a large-scale context for studies concerning specific processes of various components of TOGA COARE.

  11. Cosmic Web of Galaxies in the COMOS Field

    NASA Astrophysics Data System (ADS)

    Darvish, Behnam; Martin, Christopher D.; Mobasher, Bahram; Scoville, Nicholas; Sobral, David; COSMOS science Team

    2017-01-01

    We use a mass complete sample of galaxies with accurate photometric redshifts in the COSMOS field to estimate the density field and to extract the components of the cosmic web. The comic web extraction algorithm relies on the signs and the ratio of eigenvalues of the Hessian matrix and is enable to integrate the density field into clusters, filaments and the field. We show that at z < 0.8, the median star-formation rate in the cosmic web gradually declines from the field to clusters and this decline is especially sharp for satellite galaxies (~1 dex vs. ~0.4 dex for centrals). However, at z > 0.8, the trend flattens out. For star-forming galaxies only, the median star-formation rate declines by ~ 0.3-0.4 dex from the field to clusters for both satellites and centrals, only at z < 0.5. We argue that for satellite galaxies, the main role of the cosmic web environment is to control their star-forming/quiescent fraction, whereas for centrals, it is mainly to control their overall star-formation rate. Given these, we suggest that most satellite galaxies experience a rapid quenching mechanism as they fall from the field into clusters through the channel of filaments, whereas for central galaxies, it is mostly due to a slow quenching process. Our preliminary results highlight the importance of the large-scale cosmic web on the evolution of galaxies.

  12. KAGLVis - On-line 3D Visualisation of Earth-observing-satellite Data

    NASA Astrophysics Data System (ADS)

    Szuba, Marek; Ameri, Parinaz; Grabowski, Udo; Maatouki, Ahmad; Meyer, Jörg

    2015-04-01

    One of the goals of the Large-Scale Data Management and Analysis project is to provide a high-performance framework facilitating management of data acquired by Earth-observing satellites such as Envisat. On the client-facing facet of this framework, we strive to provide visualisation and basic analysis tool which could be used by scientists with minimal to no knowledge of the underlying infrastructure. Our tool, KAGLVis, is a JavaScript client-server Web application which leverages modern Web technologies to provide three-dimensional visualisation of satellite observables on a wide range of client systems. It takes advantage of the WebGL API to employ locally available GPU power for 3D rendering; this approach has been demonstrated to perform well even on relatively weak hardware such as integrated graphics chipsets found in modern laptop computers and with some user-interface tuning could even be usable on embedded devices such as smartphones or tablets. Data is fetched from the database back-end using a ReST API and cached locally, both in memory and using HTML5 Web Storage, to minimise network use. Computations, calculation of cloud altitude from cloud-index measurements for instance, can depending on configuration be performed on either the client or the server side. Keywords: satellite data, Envisat, visualisation, 3D graphics, Web application, WebGL, MEAN stack.

  13. Turning Access into a web-enabled secure information system for clinical trials.

    PubMed

    Dongquan Chen; Chen, Wei-Bang; Soong, Mayhue; Soong, Seng-Jaw; Orthner, Helmuth F

    2009-08-01

    Organizations that have limited resources need to conduct clinical studies in a cost-effective, but secure way. Clinical data residing in various individual databases need to be easily accessed and secured. Although widely available, digital certification, encryption, and secure web server, have not been implemented as widely, partly due to a lack of understanding of needs and concerns over issues such as cost and difficulty in implementation. The objective of this study was to test the possibility of centralizing various databases and to demonstrate ways of offering an alternative to a large-scale comprehensive and costly commercial product, especially for simple phase I and II trials, with reasonable convenience and security. We report a working procedure to transform and develop a standalone Access database into a secure Web-based secure information system. For data collection and reporting purposes, we centralized several individual databases; developed, and tested a web-based secure server using self-issued digital certificates. The system lacks audit trails. The cost of development and maintenance may hinder its wide application. The clinical trial databases scattered in various departments of an institution could be centralized into a web-enabled secure information system. The limitations such as the lack of a calendar and audit trail can be partially addressed with additional programming. The centralized Web system may provide an alternative to a comprehensive clinical trial management system.

  14. PubMed and beyond: a survey of web tools for searching biomedical literature

    PubMed Central

    Lu, Zhiyong

    2011-01-01

    The past decade has witnessed the modern advances of high-throughput technology and rapid growth of research capacity in producing large-scale biological data, both of which were concomitant with an exponential growth of biomedical literature. This wealth of scholarly knowledge is of significant importance for researchers in making scientific discoveries and healthcare professionals in managing health-related matters. However, the acquisition of such information is becoming increasingly difficult due to its large volume and rapid growth. In response, the National Center for Biotechnology Information (NCBI) is continuously making changes to its PubMed Web service for improvement. Meanwhile, different entities have devoted themselves to developing Web tools for helping users quickly and efficiently search and retrieve relevant publications. These practices, together with maturity in the field of text mining, have led to an increase in the number and quality of various Web tools that provide comparable literature search service to PubMed. In this study, we review 28 such tools, highlight their respective innovations, compare them to the PubMed system and one another, and discuss directions for future development. Furthermore, we have built a website dedicated to tracking existing systems and future advances in the field of biomedical literature search. Taken together, our work serves information seekers in choosing tools for their needs and service providers and developers in keeping current in the field. Database URL: http://www.ncbi.nlm.nih.gov/CBBresearch/Lu/search PMID:21245076

  15. Inter-decadal variability of phytoplankton biomass along the coastal West Antarctic Peninsula.

    PubMed

    Kim, Hyewon; Ducklow, Hugh W; Abele, Doris; Ruiz Barlett, Eduardo M; Buma, Anita G J; Meredith, Michael P; Rozema, Patrick D; Schofield, Oscar M; Venables, Hugh J; Schloss, Irene R

    2018-06-28

    The West Antarctic Peninsula (WAP) is a climatically sensitive region where periods of strong warming have caused significant changes in the marine ecosystem and food-web processes. Tight coupling between phytoplankton and higher trophic levels implies that the coastal WAP is a bottom-up controlled system, where changes in phytoplankton dynamics may largely impact other food-web components. Here, we analysed the inter-decadal time series of year-round chlorophyll- a (Chl) collected from three stations along the coastal WAP: Carlini Station at Potter Cove (PC) on King George Island, Palmer Station on Anvers Island and Rothera Station on Adelaide Island. There were trends towards increased phytoplankton biomass at Carlini Station (PC) and Palmer Station, while phytoplankton biomass declined significantly at Rothera Station over the studied period. The impacts of two relevant climate modes to the WAP, the El Niño-Southern Oscillation and the Southern Annular Mode, on winter and spring phytoplankton biomass appear to be different among the three sampling stations, suggesting an important role of local-scale forcing than large-scale forcing on phytoplankton dynamics at each station. The inter-annual variability of seasonal bloom progression derived from considering all three stations together captured ecologically meaningful, seasonally co-occurring bloom patterns which were primarily constrained by water-column stability strength. Our findings highlight a coupled link between phytoplankton and physical and climate dynamics along the coastal WAP, which may improve our understanding of overall WAP food-web responses to climate change and variability.This article is part of the theme issue 'The marine system of the West Antarctic Peninsula: status and strategy for progress in a region of rapid change'. © 2018 The Author(s).

  16. Using online database for landslide susceptibility assessment with an example from the Veneto Region (north-eastern Italy).

    NASA Astrophysics Data System (ADS)

    Floris, Mario; Squarzoni, Cristina; Zorzi, Luca; D'Alpaos, Andrea; Iafelice, Maria

    2010-05-01

    Landslide susceptibility maps describe landslide-prone areas by the spatial correlation between landslides and related factors, derived from different kinds of datasets: geological, geotechnical and geomechanical maps, hydrogeological maps, landslides maps, vector and raster terrain data, real-time inclinometer and pore pressure data. In the last decade, thanks to the increasing use of web-based tools for management, sharing and communication of territorial information, many Web-based Geographical Information Systems (WebGIS) were created by local governments or nations, University and Research Centres. Nowadays there is a strong proliferation of geological WebGIS or GeoBrowser, allowing free download of spatial information. There are global Cartographical Portals that provide a free download of DTM and other vector data related to the whole planet (http://www.webgis.com). At major scale, there are WebGIS regarding entire nation (http://www.agiweb.org), or specific region of a country (http://www.mrt.tas.gov.au), or single municipality (http://sitn.ne.ch/). Moreover, portals managed by local government and academic government (http://turtle.ags.gov.ab.ca/Peace_River/Site/) or by a private agency (http://www.bbt-se.com) are noteworthy. In Italy, the first national projects for the creation of WebGIS and web-based databases begun during the 1980s, and evolved, through years, to the present number of different WebGIS, which have different territorial extensions: national (Italian National Cartographical Portal, http://www.pcn.minambiente.it; E-GEO Project, http://www.egeo.unisi.it), interregional (River Tiber Basin Authority, www.abtevere.it ), and regional (Veneto Region, www.regione.veneto.it). In this way we investigated most of the Italian WebGIS in order to verify their geographic range and the availability and quality of data useful for landslide hazard analyses. We noticed a large variability of the accessing information among the different browsers. In particular, the Trento and Bolzano Provinces Geobrowsers (http://www.provincia.bz.it; http://www.territorio.provincia.tn.it) provide a large availability of data respect to the other regional and interregional WebGIS, which generally allow only the download of topographic data. Recently, the Italian Institute for Environmental Protection and Research, ISPRA (Istituto Superiore per la Protezione e la ricerca Ambientale), makes available and free usable the Italian Inventory of Landslides (IFFI Project). The inventory contains information derived from the census of all the instability phenomena in Italy, offering a base-cognitive instrument for the landslide hazard evaluation. For the landslide hazard assessment it is essential to evaluate the real effectiveness of the available data. Hence, we test the effectiveness of the web databases to evaluate the landslides susceptibility in the Euganean Hill Regional Park (185.5 km2), located at SE of Padua (Veneto Region, Italy). We used data available from three online spatial databases: Veneto Region Cartographic Portal (http://www.regione.veneto.it), for vector terrain data at 1:5000 scale; the IFFI archive (http://www.sinanet.apat.it), for information concerning landslides; and the National Cartographic Portal of the Italian Ministry of Environment (http://www.pcn.minambiente.it), for the multi-temporal orthophotos. The landslide susceptibility was evaluated using a simple probabilistic analysis considering the relationships between landslides and DEM-derived factors, such as slope, curvature and aspect. For the validation of the analysis, we made a spatial test by subdividing the study area in two sectors: training area and test area. The obtained results show that the actual no-completeness of online available spatial databases related to the Veneto Region allows only regional and medium scale (>1:25,000) susceptibility analysis. Data about lithology, land use, groundwater and others relevant factors are absent. In addition, the lack of data on the temporal evolution of the landslides permits only a spatial analysis, impeding a complete evaluation of the landslide hazard.

  17. NDEx - the Network Data Exchange, A Network Commons for Biologists | Informatics Technology for Cancer Research (ITCR)

    Cancer.gov

    Network models of biology, whether curated or derived from large-scale data analysis, are critical tools in the understanding of cancer mechanisms and in the design and personalization of therapies. The NDEx Project (Network Data Exchange) will create, deploy, and maintain an open-source, web-based software platform and public website to enable scientists, organizations, and software applications to share, store, manipulate, and publish biological networks.

  18. Advanced Visualization and Interactive Display Rapid Innovation and Discovery Evaluation Research (VISRIDER) Program Task 6: Point Cloud Visualization Techniques for Desktop and Web Platforms

    DTIC Science & Technology

    2017-04-01

    ADVANCED VISUALIZATION AND INTERACTIVE DISPLAY RAPID INNOVATION AND DISCOVERY EVALUATION RESEARCH (VISRIDER) PROGRAM TASK 6: POINT CLOUD...To) OCT 2013 – SEP 2014 4. TITLE AND SUBTITLE ADVANCED VISUALIZATION AND INTERACTIVE DISPLAY RAPID INNOVATION AND DISCOVERY EVALUATION RESEARCH...various point cloud visualization techniques for viewing large scale LiDAR datasets. Evaluate their potential use for thick client desktop platforms

  19. Explore, Visualize, and Analyze Functional Cancer Proteomic Data Using the Cancer Proteome Atlas. | Office of Cancer Genomics

    Cancer.gov

    Reverse-phase protein arrays (RPPA) represent a powerful functional proteomic approach to elucidate cancer-related molecular mechanisms and to develop novel cancer therapies. To facilitate community-based investigation of the large-scale protein expression data generated by this platform, we have developed a user-friendly, open-access bioinformatic resource, The Cancer Proteome Atlas (TCPA, http://tcpaportal.org), which contains two separate web applications.

  20. Food-web structure of seagrass communities across different spatial scales and human impacts.

    PubMed

    Coll, Marta; Schmidt, Allison; Romanuk, Tamara; Lotze, Heike K

    2011-01-01

    Seagrass beds provide important habitat for a wide range of marine species but are threatened by multiple human impacts in coastal waters. Although seagrass communities have been well-studied in the field, a quantification of their food-web structure and functioning, and how these change across space and human impacts has been lacking. Motivated by extensive field surveys and literature information, we analyzed the structural features of food webs associated with Zostera marina across 16 study sites in 3 provinces in Atlantic Canada. Our goals were to (i) quantify differences in food-web structure across local and regional scales and human impacts, (ii) assess the robustness of seagrass webs to simulated species loss, and (iii) compare food-web structure in temperate Atlantic seagrass beds with those of other aquatic ecosystems. We constructed individual food webs for each study site and cumulative webs for each province and the entire region based on presence/absence of species, and calculated 16 structural properties for each web. Our results indicate that food-web structure was similar among low impact sites across regions. With increasing human impacts associated with eutrophication, however, food-web structure show evidence of degradation as indicated by fewer trophic groups, lower maximum trophic level of the highest top predator, fewer trophic links connecting top to basal species, higher fractions of herbivores and intermediate consumers, and higher number of prey per species. These structural changes translate into functional changes with impacted sites being less robust to simulated species loss. Temperate Atlantic seagrass webs are similar to a tropical seagrass web, yet differed from other aquatic webs, suggesting consistent food-web characteristics across seagrass ecosystems in different regions. Our study illustrates that food-web structure and functioning of seagrass habitats change with human impacts and that the spatial scale of food-web analysis is critical for determining results.

  1. Food-Web Structure of Seagrass Communities across Different Spatial Scales and Human Impacts

    PubMed Central

    Coll, Marta; Schmidt, Allison; Romanuk, Tamara; Lotze, Heike K.

    2011-01-01

    Seagrass beds provide important habitat for a wide range of marine species but are threatened by multiple human impacts in coastal waters. Although seagrass communities have been well-studied in the field, a quantification of their food-web structure and functioning, and how these change across space and human impacts has been lacking. Motivated by extensive field surveys and literature information, we analyzed the structural features of food webs associated with Zostera marina across 16 study sites in 3 provinces in Atlantic Canada. Our goals were to (i) quantify differences in food-web structure across local and regional scales and human impacts, (ii) assess the robustness of seagrass webs to simulated species loss, and (iii) compare food-web structure in temperate Atlantic seagrass beds with those of other aquatic ecosystems. We constructed individual food webs for each study site and cumulative webs for each province and the entire region based on presence/absence of species, and calculated 16 structural properties for each web. Our results indicate that food-web structure was similar among low impact sites across regions. With increasing human impacts associated with eutrophication, however, food-web structure show evidence of degradation as indicated by fewer trophic groups, lower maximum trophic level of the highest top predator, fewer trophic links connecting top to basal species, higher fractions of herbivores and intermediate consumers, and higher number of prey per species. These structural changes translate into functional changes with impacted sites being less robust to simulated species loss. Temperate Atlantic seagrass webs are similar to a tropical seagrass web, yet differed from other aquatic webs, suggesting consistent food-web characteristics across seagrass ecosystems in different regions. Our study illustrates that food-web structure and functioning of seagrass habitats change with human impacts and that the spatial scale of food-web analysis is critical for determining results. PMID:21811637

  2. Using Web Maps to Analyze the Construction of Global Scale Cognitive Maps

    ERIC Educational Resources Information Center

    Pingel, Thomas J.

    2018-01-01

    Game-based Web sites and applications are changing the ways in which students learn the world map. In this study, a Web map-based digital learning tool was used as a study aid for a university-level geography course in order to examine the way in which global scale cognitive maps are constructed. A network analysis revealed that clicks were…

  3. Optimising web site designs for people with learning disabilities

    PubMed Central

    Williams, Peter; Hennig, Christian

    2015-01-01

    Much relevant internet-mediated information is inaccessible to people with learning disabilities because of difficulties in navigating the web. This paper reports on the methods undertaken to determine how information can be optimally presented for this cohort. Qualitative work is outlined where attributes relating to site layout affecting usability were elicited. A study comparing web sites of different design layouts exhibiting these attributes is discussed, with the emphasis on methodology. Eight interfaces were compared using various combinations of menu position (vertical or horizontal), text size and the absence or presence of images to determine which attributes of a site have the greatest performance impact. Study participants were also asked for their preferences, via a ‘smiley-face’ rating scale and simple interviews. ‘Acquiescence bias’ was minimised by avoiding polar (‘yes/no’) interrogatives, achieved by asking participants to compare layouts (such as horizontal versus vertical menu), with reasons coaxed from those able to articulate them. Preferred designs were for large text and images. This was the reverse of those facilitating fastest retrieval times, a discrepancy due to preferences being judged on aesthetic considerations. Design recommendations that reconcile preference and performance findings are offered. These include using a horizontal menu, juxtaposing images and text, and reducing text from sentences to phrases, thus facilitating preferred large text without increasing task times. PMID:26097431

  4. Secure web book to store structural genomics research data.

    PubMed

    Manjasetty, Babu A; Höppner, Klaus; Mueller, Uwe; Heinemann, Udo

    2003-01-01

    Recently established collaborative structural genomics programs aim at significantly accelerating the crystal structure analysis of proteins. These large-scale projects require efficient data management systems to ensure seamless collaboration between different groups of scientists working towards the same goal. Within the Berlin-based Protein Structure Factory, the synchrotron X-ray data collection and the subsequent crystal structure analysis tasks are located at BESSY, a third-generation synchrotron source. To organize file-based communication and data transfer at the BESSY site of the Protein Structure Factory, we have developed the web-based BCLIMS, the BESSY Crystallography Laboratory Information Management System. BCLIMS is a relational data management system which is powered by MySQL as the database engine and Apache HTTP as the web server. The database interface routines are written in Python programing language. The software is freely available to academic users. Here we describe the storage, retrieval and manipulation of laboratory information, mainly pertaining to the synchrotron X-ray diffraction experiments and the subsequent protein structure analysis, using BCLIMS.

  5. From PACS to Web-based ePR system with image distribution for enterprise-level filmless healthcare delivery.

    PubMed

    Huang, H K

    2011-07-01

    The concept of PACS (picture archiving and communication system) was initiated in 1982 during the SPIE medical imaging conference in New Port Beach, CA. Since then PACS has been matured to become an everyday clinical tool for image archiving, communication, display, and review. This paper follows the continuous development of PACS technology including Web-based PACS, PACS and ePR (electronic patient record), enterprise PACS to ePR with image distribution (ID). The concept of large-scale Web-based enterprise PACS and ePR with image distribution is presented along with its implementation, clinical deployment, and operation. The Hong Kong Hospital Authority's (HKHA) integration of its home-grown clinical management system (CMS) with PACS and ePR with image distribution is used as a case study. The current concept and design criteria of the HKHA enterprise integration of the CMS, PACS, and ePR-ID for filmless healthcare delivery are discussed, followed by its work-in-progress and current status.

  6. Columbia University's Informatics for Diabetes Education and Telemedicine (IDEATel) Project

    PubMed Central

    Starren, Justin; Hripcsak, George; Sengupta, Soumitra; Abbruscato, C.R.; Knudson, Paul E.; Weinstock, Ruth S.; Shea, Steven

    2002-01-01

    The Columbia University Informatics for Diabetes Education and Telemedicine IDEATel) project is a four-year demonstration project funded by the Centers for Medicare and Medicaid Services with the overall goal of evaluating the feasibility, acceptability, effectiveness, and cost-effectiveness of telemedicine. The focal point of the intervention is the home telemedicine unit (HTU), which provides four functions: synchronous videoconferencing over standard telephone lines, electronic transmission for fingerstick glucose and blood pressure readings, secure Web-based messaging and clinical data review, and access to Web-based educational materials. The HTU must be usable by elderly patients with no prior computer experience. Providing these functions through the HTU requires tight integration of six components: the HTU itself, case management software, a clinical information system, Web-based educational material, data security, and networking and telecommunications. These six components were integrated through a variety of interfaces, providing a system that works well for patients and providers. With more than 400 HTUs installed, IDEATel has demonstrated the feasibility of large-scale home telemedicine. PMID:11751801

  7. A Real-Time Web of Things Framework with Customizable Openness Considering Legacy Devices

    PubMed Central

    Zhao, Shuai; Yu, Le; Cheng, Bo

    2016-01-01

    With the development of the Internet of Things (IoT), resources and applications based on it have emerged on a large scale. However, most efforts are “silo” solutions where devices and applications are tightly coupled. Infrastructures are needed to connect sensors to the Internet, open up and break the current application silos and move to a horizontal application mode. Based on the concept of Web of Things (WoT), many infrastructures have been proposed to integrate the physical world with the Web. However, issues such as no real-time guarantee, lack of fine-grained control of data, and the absence of explicit solutions for integrating heterogeneous legacy devices, hinder their widespread and practical use. To address these issues, this paper proposes a WoT resource framework that provides the infrastructures for the customizable openness and sharing of users’ data and resources under the premise of ensuring the real-time behavior of their own applications. The proposed framework is validated by actual systems and experimental evaluations. PMID:27690038

  8. A Real-Time Web of Things Framework with Customizable Openness Considering Legacy Devices.

    PubMed

    Zhao, Shuai; Yu, Le; Cheng, Bo

    2016-09-28

    With the development of the Internet of Things (IoT), resources and applications based on it have emerged on a large scale. However, most efforts are "silo" solutions where devices and applications are tightly coupled. Infrastructures are needed to connect sensors to the Internet, open up and break the current application silos and move to a horizontal application mode. Based on the concept of Web of Things (WoT), many infrastructures have been proposed to integrate the physical world with the Web. However, issues such as no real-time guarantee, lack of fine-grained control of data, and the absence of explicit solutions for integrating heterogeneous legacy devices, hinder their widespread and practical use. To address these issues, this paper proposes a WoT resource framework that provides the infrastructures for the customizable openness and sharing of users' data and resources under the premise of ensuring the real-time behavior of their own applications. The proposed framework is validated by actual systems and experimental evaluations.

  9. A Global Overview of Male Escort Websites.

    PubMed

    Kumar, Navin; Minichiello, Victor; Scott, John; Harrington, Taylor

    2017-01-01

    This article details a preliminary dataset of global male escort sites to give insight into the scale of the online market. We conducted a content analysis of 499 Web sites and also measured traffic to these sites. Our analysis examined the structural characteristics of escort services, geographical and regulatory contexts, and resilience of such services. Results suggest that most sites are independent and not affiliated to escort agencies, and the majority cater to male escorts soliciting male clients, with a number of sites for female clientele and couples. These Web sites are dispersed globally, with Asian, European, and South American countries the major hubs in the market and a small number of large multinational sites based in the United States and Europe figuring as a major presence in markets. Although still subject to high levels of regulation in many parts of the world, the data suggest that male escorting is becoming more visible in diverse cultural contexts as measured by the number of Web sites appearing in public spaces.

  10. Building up the spin - orbit alignment of interacting galaxy pairs

    NASA Astrophysics Data System (ADS)

    Moon, Jun-Sung; Yoon, Suk-Jin

    2018-01-01

    Galaxies are not just randomly distributed throughout space. Instead, they are in alignment over a wide range of scales from the cosmic web down to a pair of galaxies. Motivated by recent findings that the spin and the orbital angular momentum vectors of galaxy pairs tend to be parallel, we here investigate the spin - orbit orientation in close pairs using the Illustris cosmological simulation. We find that since z ~ 1, the parallel alignment has become progressively stronger with time through repetitive encounters. The pair Interactions are preferentially in prograde at z = 0 (over 5 sigma significance). The prograde fraction at z = 0 is larger for the pairs influenced more heavily by each other during their evolution. We find no correlation between the spin - orbit orientation and the surrounding large-scale structure. Our results favor the scenario in which the alignment in close pairs is caused by tidal interactions later on, rather than the primordial torquing by the large-scale structures.

  11. UFO: a web server for ultra-fast functional profiling of whole genome protein sequences.

    PubMed

    Meinicke, Peter

    2009-09-02

    Functional profiling is a key technique to characterize and compare the functional potential of entire genomes. The estimation of profiles according to an assignment of sequences to functional categories is a computationally expensive task because it requires the comparison of all protein sequences from a genome with a usually large database of annotated sequences or sequence families. Based on machine learning techniques for Pfam domain detection, the UFO web server for ultra-fast functional profiling allows researchers to process large protein sequence collections instantaneously. Besides the frequencies of Pfam and GO categories, the user also obtains the sequence specific assignments to Pfam domain families. In addition, a comparison with existing genomes provides dissimilarity scores with respect to 821 reference proteomes. Considering the underlying UFO domain detection, the results on 206 test genomes indicate a high sensitivity of the approach. In comparison with current state-of-the-art HMMs, the runtime measurements show a considerable speed up in the range of four orders of magnitude. For an average size prokaryotic genome, the computation of a functional profile together with its comparison typically requires about 10 seconds of processing time. For the first time the UFO web server makes it possible to get a quick overview on the functional inventory of newly sequenced organisms. The genome scale comparison with a large number of precomputed profiles allows a first guess about functionally related organisms. The service is freely available and does not require user registration or specification of a valid email address.

  12. The build up of the correlation between halo spin and the large-scale structure

    NASA Astrophysics Data System (ADS)

    Wang, Peng; Kang, Xi

    2018-01-01

    Both simulations and observations have confirmed that the spin of haloes/galaxies is correlated with the large-scale structure (LSS) with a mass dependence such that the spin of low-mass haloes/galaxies tend to be parallel with the LSS, while that of massive haloes/galaxies tend to be perpendicular with the LSS. It is still unclear how this mass dependence is built up over time. We use N-body simulations to trace the evolution of the halo spin-LSS correlation and find that at early times the spin of all halo progenitors is parallel with the LSS. As time goes on, mass collapsing around massive halo is more isotropic, especially the recent mass accretion along the slowest collapsing direction is significant and it brings the halo spin to be perpendicular with the LSS. Adopting the fractional anisotropy (FA) parameter to describe the degree of anisotropy of the large-scale environment, we find that the spin-LSS correlation is a strong function of the environment such that a higher FA (more anisotropic environment) leads to an aligned signal, and a lower anisotropy leads to a misaligned signal. In general, our results show that the spin-LSS correlation is a combined consequence of mass flow and halo growth within the cosmic web. Our predicted environmental dependence between spin and large-scale structure can be further tested using galaxy surveys.

  13. Topology and evolution of technology innovation networks

    NASA Astrophysics Data System (ADS)

    Valverde, Sergi; Solé, Ricard V.; Bedau, Mark A.; Packard, Norman

    2007-11-01

    The web of relations linking technological innovation can be fairly described in terms of patent citations. The resulting patent citation network provides a picture of the large-scale organization of innovations and its time evolution. Here we study the patterns of change of patents registered by the U.S. Patent and Trademark Office. We show that the scaling behavior exhibited by this network is consistent with a preferential attachment mechanism together with a Weibull-shaped aging term. Such an attachment kernel is shared by scientific citation networks, thus indicating a universal type of mechanism linking ideas and designs and their evolution. The implications for evolutionary theory of innovation are discussed.

  14. Large-scale model quality assessment for improving protein tertiary structure prediction.

    PubMed

    Cao, Renzhi; Bhattacharya, Debswapna; Adhikari, Badri; Li, Jilong; Cheng, Jianlin

    2015-06-15

    Sampling structural models and ranking them are the two major challenges of protein structure prediction. Traditional protein structure prediction methods generally use one or a few quality assessment (QA) methods to select the best-predicted models, which cannot consistently select relatively better models and rank a large number of models well. Here, we develop a novel large-scale model QA method in conjunction with model clustering to rank and select protein structural models. It unprecedentedly applied 14 model QA methods to generate consensus model rankings, followed by model refinement based on model combination (i.e. averaging). Our experiment demonstrates that the large-scale model QA approach is more consistent and robust in selecting models of better quality than any individual QA method. Our method was blindly tested during the 11th Critical Assessment of Techniques for Protein Structure Prediction (CASP11) as MULTICOM group. It was officially ranked third out of all 143 human and server predictors according to the total scores of the first models predicted for 78 CASP11 protein domains and second according to the total scores of the best of the five models predicted for these domains. MULTICOM's outstanding performance in the extremely competitive 2014 CASP11 experiment proves that our large-scale QA approach together with model clustering is a promising solution to one of the two major problems in protein structure modeling. The web server is available at: http://sysbio.rnet.missouri.edu/multicom_cluster/human/. © The Author 2015. Published by Oxford University Press.

  15. Measuring Listening Comprehension Skills of 5th Grade School Students with the Help of Web Based System

    ERIC Educational Resources Information Center

    Acat, M. Bahaddin; Demiral, Hilmi; Kaya, Mehmet Fatih

    2016-01-01

    The main purpose of this study is to measure listening comprehension skills of 5th grade school students with the help of web based system. This study was conducted on 5th grade students studying at the primary schools of Eskisehir. The scale used in the process of the study is "Web Based Listening Scale". In the process of the study,…

  16. Development of a database system for mapping insertional mutations onto the mouse genome with large-scale experimental data

    PubMed Central

    2009-01-01

    Background Insertional mutagenesis is an effective method for functional genomic studies in various organisms. It can rapidly generate easily tractable mutations. A large-scale insertional mutagenesis with the piggyBac (PB) transposon is currently performed in mice at the Institute of Developmental Biology and Molecular Medicine (IDM), Fudan University in Shanghai, China. This project is carried out via collaborations among multiple groups overseeing interconnected experimental steps and generates a large volume of experimental data continuously. Therefore, the project calls for an efficient database system for recording, management, statistical analysis, and information exchange. Results This paper presents a database application called MP-PBmice (insertional mutation mapping system of PB Mutagenesis Information Center), which is developed to serve the on-going large-scale PB insertional mutagenesis project. A lightweight enterprise-level development framework Struts-Spring-Hibernate is used here to ensure constructive and flexible support to the application. The MP-PBmice database system has three major features: strict access-control, efficient workflow control, and good expandability. It supports the collaboration among different groups that enter data and exchange information on daily basis, and is capable of providing real time progress reports for the whole project. MP-PBmice can be easily adapted for other large-scale insertional mutation mapping projects and the source code of this software is freely available at http://www.idmshanghai.cn/PBmice. Conclusion MP-PBmice is a web-based application for large-scale insertional mutation mapping onto the mouse genome, implemented with the widely used framework Struts-Spring-Hibernate. This system is already in use by the on-going genome-wide PB insertional mutation mapping project at IDM, Fudan University. PMID:19958505

  17. The Computing and Data Grid Approach: Infrastructure for Distributed Science Applications

    NASA Technical Reports Server (NTRS)

    Johnston, William E.

    2002-01-01

    With the advent of Grids - infrastructure for using and managing widely distributed computing and data resources in the science environment - there is now an opportunity to provide a standard, large-scale, computing, data, instrument, and collaboration environment for science that spans many different projects and provides the required infrastructure and services in a relatively uniform and supportable way. Grid technology has evolved over the past several years to provide the services and infrastructure needed for building 'virtual' systems and organizations. We argue that Grid technology provides an excellent basis for the creation of the integrated environments that can combine the resources needed to support the large- scale science projects located at multiple laboratories and universities. We present some science case studies that indicate that a paradigm shift in the process of science will come about as a result of Grids providing transparent and secure access to advanced and integrated information and technologies infrastructure: powerful computing systems, large-scale data archives, scientific instruments, and collaboration tools. These changes will be in the form of services that can be integrated with the user's work environment, and that enable uniform and highly capable access to these computers, data, and instruments, regardless of the location or exact nature of these resources. These services will integrate transient-use resources like computing systems, scientific instruments, and data caches (e.g., as they are needed to perform a simulation or analyze data from a single experiment); persistent-use resources. such as databases, data catalogues, and archives, and; collaborators, whose involvement will continue for the lifetime of a project or longer. While we largely address large-scale science in this paper, Grids, particularly when combined with Web Services, will address a broad spectrum of science scenarios. both large and small scale.

  18. HAlign-II: efficient ultra-large multiple sequence alignment and phylogenetic tree reconstruction with distributed and parallel computing.

    PubMed

    Wan, Shixiang; Zou, Quan

    2017-01-01

    Multiple sequence alignment (MSA) plays a key role in biological sequence analyses, especially in phylogenetic tree construction. Extreme increase in next-generation sequencing results in shortage of efficient ultra-large biological sequence alignment approaches for coping with different sequence types. Distributed and parallel computing represents a crucial technique for accelerating ultra-large (e.g. files more than 1 GB) sequence analyses. Based on HAlign and Spark distributed computing system, we implement a highly cost-efficient and time-efficient HAlign-II tool to address ultra-large multiple biological sequence alignment and phylogenetic tree construction. The experiments in the DNA and protein large scale data sets, which are more than 1GB files, showed that HAlign II could save time and space. It outperformed the current software tools. HAlign-II can efficiently carry out MSA and construct phylogenetic trees with ultra-large numbers of biological sequences. HAlign-II shows extremely high memory efficiency and scales well with increases in computing resource. THAlign-II provides a user-friendly web server based on our distributed computing infrastructure. HAlign-II with open-source codes and datasets was established at http://lab.malab.cn/soft/halign.

  19. Improving a web-based employability intervention for work-disabled employees: results of a pilot economic evaluation.

    PubMed

    Noben, Cindy; Evers, Silvia; Genabeek, Joost van; Nijhuis, Frans; de Rijk, Angelique

    2017-04-01

    Purpose The purpose of this study is to improve web-based employability interventions for employees with work-related health problems for both intervention content and study design by means of a pilot economic evaluation. Methods Uptake rate analysis for the intervention elements, cost effectiveness, cost utility and subgroup analyses were conducted to identify potential content-related intervention improvements. Differences in work ability and quality-adjusted life years and overall contribution of resource items to the total costs were assessed. These were used to guide study design improvements. Results Sixty-three participants were a-select allocated to either the intervention (n = 29) or the control (n = 34) group. Uptake regarding the intervention elements ranged between 3% and 70%. Cost-effectiveness and cost-utility analyses resulted in negative effects although higher total costs. Incremental effects were marginal (work ability -0.51; QALY -0.01). Conclusions The web-based tool to enhance employability among work disabled employees requires improvements regarding targeting and intensity; outcome measures selected and collection of cost data. With respect to the studies of disability and rehabilitation, the findings and methods presented in this pilot economic evaluation could guide the assessment of future assistive "e-health" technologies. IMPLICATIONS FOR REHABILITATION The methods presented in this pilot economic evaluation have large potentials to guide the assessment of future assistive e-health technologies addressing work-disabilities. The findings show that the web-based tool requires content related improvements with respect to targeting and intensity to enhance employability among work disabled employees. The findings show that the web-based tool would benefit from improvements related to the study design by more adequately selecting and collecting both outcome measures and cost data. The burden attributable to large-scale studies and implementation issues were prevented as the outcomes of the pilot economic evaluation did not support the implementation of the web-based tool.

  20. Creation of a Web-Based GIS Server and Custom Geoprocessing Tools for Enhanced Hydrologic Applications

    NASA Astrophysics Data System (ADS)

    Welton, B.; Chouinard, K.; Sultan, M.; Becker, D.; Milewski, A.; Becker, R.

    2010-12-01

    Rising populations in the arid and semi arid parts of the World are increasing the demand for fresh water supplies worldwide. Many data sets needed for assessment of hydrologic applications across vast regions of the world are expensive, unpublished, difficult to obtain, or at varying scales which complicates their use. Fortunately, this situation is changing with the development of global remote sensing datasets and web-based platforms such as GIS Server. GIS provides a cost effective vehicle for comparing, analyzing, and querying a variety of spatial datasets as geographically referenced layers. We have recently constructed a web-based GIS, that incorporates all relevant geological, geochemical, geophysical, and remote sensing data sets that were readily used to identify reservoir types and potential well locations on local and regional scales in various tectonic settings including: (1) extensional environment (Red Sea rift), (2) transcurrent fault system (Najd Fault in the Arabian-Nubian Shield), and (3) compressional environments (Himalayas). The web-based GIS could also be used to detect spatial and temporal trends in precipitation, recharge, and runoff in large watersheds on local, regional, and continental scales. These applications were enabled through the construction of a web-based ArcGIS Server with Google Map’s interface and the development of customized geoprocessing tools. ArcGIS Server provides out-of-the-box setups that are generic in nature. This platform includes all of the standard web based GIS tools (e.g. pan, zoom, identify, search, data querying, and measurement). In addition to the standard suite of tools provided by ArcGIS Server an additional set of advanced data manipulation and display tools was also developed to allow for a more complete and customizable view of the area of interest. The most notable addition to the standard GIS Server tools is the custom on-demand geoprocessing tools (e.g., graph, statistical functions, custom raster creation, profile, TRMM). The generation of a wide range of derivative maps (e.g., buffer zone, contour map, graphs, temporal rainfall distribution maps) from various map layers (e.g., geologic maps, geophysics, satellite images) allows for more user flexibility. The use of these tools along with Google Map’s API which enables the website user to utilize high quality GeoEye 2 images provide by Google in conjunction with our data, creates a more complete image of the area being observed and allows for custom derivative maps to be created in the field and viewed immediately on the web, processes that were restricted to offline databases.

  1. Food-web structure and network theory: The role of connectance and size

    PubMed Central

    Dunne, Jennifer A.; Williams, Richard J.; Martinez, Neo D.

    2002-01-01

    Networks from a wide range of physical, biological, and social systems have been recently described as “small-world” and “scale-free.” However, studies disagree whether ecological networks called food webs possess the characteristic path lengths, clustering coefficients, and degree distributions required for membership in these classes of networks. Our analysis suggests that the disagreements are based on selective use of relatively few food webs, as well as analytical decisions that obscure important variability in the data. We analyze a broad range of 16 high-quality food webs, with 25–172 nodes, from a variety of aquatic and terrestrial ecosystems. Food webs generally have much higher complexity, measured as connectance (the fraction of all possible links that are realized in a network), and much smaller size than other networks studied, which have important implications for network topology. Our results resolve prior conflicts by demonstrating that although some food webs have small-world and scale-free structure, most do not if they exceed a relatively low level of connectance. Although food-web degree distributions do not display a universal functional form, observed distributions are systematically related to network connectance and size. Also, although food webs often lack small-world structure because of low clustering, we identify a continuum of real-world networks including food webs whose ratios of observed to random clustering coefficients increase as a power–law function of network size over 7 orders of magnitude. Although food webs are generally not small-world, scale-free networks, food-web topology is consistent with patterns found within those classes of networks. PMID:12235364

  2. Effects of microhabitat and large-scale land use on stream salamander occupancy in the coalfields of Central Appalachia

    USGS Publications Warehouse

    Sweeten, Sara E.; Ford, W. Mark

    2016-01-01

    Large-scale coal mining practices, particularly surface coal extraction and associated valley fills as well as residential wastewater discharge, are of ecological concern for aquatic systems in central Appalachia. Identifying and quantifying alterations to ecosystems along a gradient of spatial scales is a necessary first-step to aid in mitigation of negative consequences to aquatic biota. In central Appalachian headwater streams, apart from fish, salamanders are the most abundant vertebrate predator that provide a significant intermediate trophic role linking aquatic and terrestrial food webs. Stream salamander species are considered to be sensitive to aquatic stressors and environmental alterations, as past research has shown linkages among microhabitat parameters, large-scale land use such as urbanization and logging, and salamander abundances. However, there is little information examining these relationships between environmental conditions and salamander occupancy in the coalfields of central Appalachia. In the summer of 2013, 70 sites (sampled two to three times each) in the southwest Virginia coalfields were visited to collect salamanders and quantify stream and riparian microhabitat parameters. Using an information-theoretic framework, effects of microhabitat and large-scale land use on stream salamander occupancy were compared. The findings indicate that Desmognathus spp. occupancy rates are more correlated to microhabitat parameters such as canopy cover than to large-scale land uses. However, Eurycea spp. occupancy rates had a strong association with large-scale land uses, particularly recent mining and forest cover within the watershed. These findings suggest that protection of riparian habitats is an important consideration for maintaining aquatic systems in central Appalachia. If this is not possible, restoration riparian areas should follow guidelines using quick-growing tree species that are native to Appalachian riparian areas. These types of trees would rapidly establish a canopy cover, stabilize the soil, and impede invasive plant species which would, in turn, provide high-quality refuges for stream salamanders.

  3. BrainLiner: A Neuroinformatics Platform for Sharing Time-Aligned Brain-Behavior Data

    PubMed Central

    Takemiya, Makoto; Majima, Kei; Tsukamoto, Mitsuaki; Kamitani, Yukiyasu

    2016-01-01

    Data-driven neuroscience aims to find statistical relationships between brain activity and task behavior from large-scale datasets. To facilitate high-throughput data processing and modeling, we created BrainLiner as a web platform for sharing time-aligned, brain-behavior data. Using an HDF5-based data format, BrainLiner treats brain activity and data related to behavior with the same salience, aligning both behavioral and brain activity data on a common time axis. This facilitates learning the relationship between behavior and brain activity. Using a common data file format also simplifies data processing and analyses. Properties describing data are unambiguously defined using a schema, allowing machine-readable definition of data. The BrainLiner platform allows users to upload and download data, as well as to explore and search for data from the web platform. A WebGL-based data explorer can visualize highly detailed neurophysiological data from within the web browser, and a data-driven search feature allows users to search for similar time windows of data. This increases transparency, and allows for visual inspection of neural coding. BrainLiner thus provides an essential set of tools for data sharing and data-driven modeling. PMID:26858636

  4. Dynamics of pairwise motions in the Cosmic Web

    NASA Astrophysics Data System (ADS)

    Hellwing, Wojciech A.

    2016-10-01

    We present results of analysis of the dark matter (DM) pairwise velocity statistics in different Cosmic Web environments. We use the DM velocity and density field from the Millennium 2 simulation together with the NEXUS+ algorithm to segment the simulation volume into voxels uniquely identifying one of the four possible environments: nodes, filaments, walls or cosmic voids. We show that the PDFs of the mean infall velocities v 12 as well as its spatial dependence together with the perpendicular and parallel velocity dispersions bear a significant signal of the large-scale structure environment in which DM particle pairs are embedded. The pairwise flows are notably colder and have smaller mean magnitude in wall and voids, when compared to much denser environments of filaments and nodes. We discuss on our results, indicating that they are consistent with a simple theoretical predictions for pairwise motions as induced by gravitational instability mechanism. Our results indicate that the Cosmic Web elements are coherent dynamical entities rather than just temporal geometrical associations. In addition it should be possible to observationally test various Cosmic Web finding algorithms by segmenting available peculiar velocity data and studying resulting pairwise velocity statistics.

  5. Development and Initial Validation of a Multidimensional Scale Assessing Subjective Well-Being: The Well-Being Scale (WeBS).

    PubMed

    Lui, P Priscilla; Fernando, Gaithri A

    2018-02-01

    Numerous scales currently exist that assess well-being, but research on measures of well-being is still advancing. Conceptualization and measurement of subjective well-being have emphasized intrapsychic over psychosocial domains of optimal functioning, and disparate research on hedonic, eudaimonic, and psychological well-being lacks a unifying theoretical model. Lack of systematic investigations on the impact of culture on subjective well-being has also limited advancement of this field. The goals of this investigation were to (1) develop and validate a self-report measure, the Well-Being Scale (WeBS), that simultaneously assesses overall well-being and physical, financial, social, hedonic, and eudaimonic domains of this construct; (2) evaluate factor structures that underlie subjective well-being; and (3) examine the measure's psychometric properties. Three empirical studies were conducted to develop and validate the 29-item scale. The WeBS demonstrated an adequate five-factor structure in an exploratory structural equation model in Study 1. Confirmatory factor analyses showed that a bifactor structure best fit the WeBS data in Study 2 and Study 3. Overall WeBS scores and five domain-specific subscale scores demonstrated adequate to excellent internal consistency reliability and construct validity. Mean differences in overall well-being and its five subdomains are presented for different ethnic groups. The WeBS is a reliable and valid measure of multiple aspects of well-being that are considered important to different ethnocultural groups.

  6. Alignments of Dark Matter Halos with Large-scale Tidal Fields: Mass and Redshift Dependence

    NASA Astrophysics Data System (ADS)

    Chen, Sijie; Wang, Huiyuan; Mo, H. J.; Shi, Jingjing

    2016-07-01

    Large-scale tidal fields estimated directly from the distribution of dark matter halos are used to investigate how halo shapes and spin vectors are aligned with the cosmic web. The major, intermediate, and minor axes of halos are aligned with the corresponding tidal axes, and halo spin axes tend to be parallel with the intermediate axes and perpendicular to the major axes of the tidal field. The strengths of these alignments generally increase with halo mass and redshift, but the dependence is only on the peak height, ν \\equiv {δ }{{c}}/σ ({M}{{h}},z). The scaling relations of the alignment strengths with the value of ν indicate that the alignment strengths remain roughly constant when the structures within which the halos reside are still in a quasi-linear regime, but decreases as nonlinear evolution becomes more important. We also calculate the alignments in projection so that our results can be compared directly with observations. Finally, we investigate the alignments of tidal tensors on large scales, and use the results to understand alignments of halo pairs separated at various distances. Our results suggest that the coherent structure of the tidal field is the underlying reason for the alignments of halos and galaxies seen in numerical simulations and in observations.

  7. Structural analysis of behavioral networks from the Internet

    NASA Astrophysics Data System (ADS)

    Meiss, M. R.; Menczer, F.; Vespignani, A.

    2008-06-01

    In spite of the Internet's phenomenal growth and social impact, many aspects of the collective communication behavior of its users are largely unknown. Understanding the structure and dynamics of the behavioral networks that connect users with each other and with services across the Internet is key to modeling the network and designing future applications. We present a characterization of the properties of the behavioral networks generated by several million users of the Abilene (Internet2) network. Structural features of these networks offer new insights into scaling properties of network activity and ways of distinguishing particular patterns of traffic. For example, we find that the structure of the behavioral network associated with Web activity is characterized by such extreme heterogeneity as to challenge any simple attempt to model Web server traffic.

  8. Effects of spatial scale of sampling on food web structure

    PubMed Central

    Wood, Spencer A; Russell, Roly; Hanson, Dieta; Williams, Richard J; Dunne, Jennifer A

    2015-01-01

    This study asks whether the spatial scale of sampling alters structural properties of food webs and whether any differences are attributable to changes in species richness and connectance with scale. Understanding how different aspects of sampling effort affect ecological network structure is important for both fundamental ecological knowledge and the application of network analysis in conservation and management. Using a highly resolved food web for the marine intertidal ecosystem of the Sanak Archipelago in the Eastern Aleutian Islands, Alaska, we assess how commonly studied properties of network structure differ for 281 versions of the food web sampled at five levels of spatial scale representing six orders of magnitude in area spread across the archipelago. Species (S) and link (L) richness both increased by approximately one order of magnitude across the five spatial scales. Links per species (L/S) more than doubled, while connectance (C) decreased by approximately two-thirds. Fourteen commonly studied properties of network structure varied systematically with spatial scale of sampling, some increasing and others decreasing. While ecological network properties varied systematically with sampling extent, analyses using the niche model and a power-law scaling relationship indicate that for many properties, this apparent sensitivity is attributable to the increasing S and decreasing C of webs with increasing spatial scale. As long as effects of S and C are accounted for, areal sampling bias does not have a special impact on our understanding of many aspects of network structure. However, attention does need be paid to some properties such as the fraction of species in loops, which increases more than expected with greater spatial scales of sampling. PMID:26380704

  9. CmapTools: A Software Environment for Knowledge Modeling and Sharing

    NASA Technical Reports Server (NTRS)

    Canas, Alberto J.

    2004-01-01

    In an ongoing collaborative effort between a group of NASA Ames scientists and researchers at the Institute for Human and Machine Cognition (IHMC) of the University of West Florida, a new version of CmapTools has been developed that enable scientists to construct knowledge models of their domain of expertise, share them with other scientists, make them available to anybody on the Internet with access to a Web browser, and peer-review other scientists models. These software tools have been successfully used at NASA to build a large-scale multimedia on Mars and in knowledge model on Habitability Assessment. The new version of the software places emphasis on greater usability for experts constructing their own knowledge models, and support for the creation of large knowledge models with large number of supporting resources in the forms of images, videos, web pages, and other media. Additionally, the software currently allows scientists to cooperate with each other in the construction, sharing and criticizing of knowledge models. Scientists collaborating from remote distances, for example researchers at the Astrobiology Institute, can concurrently manipulate the knowledge models they are viewing without having to do this at a special videoconferencing facility.

  10. Intergalactic medium emission observations with the cosmic web imager. II. Discovery of extended, kinematically linked emission around SSA22 Lyα BLOB 2

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Christopher Martin, D.; Chang, Daphne; Matuszewski, Matt

    The intergalactic medium (IGM) is the dominant reservoir of baryons, delineates the large-scale structure of the universe at low to moderate overdensities, and provides gas from which galaxies form and evolve. Simulations of a cold-dark-matter- (CDM-) dominated universe predict that the IGM is distributed in a cosmic web of filaments and that galaxies should form along and at the intersections of these filaments. While observations of QSO absorption lines and the large-scale distribution of galaxies have confirmed the CDM paradigm, the cosmic web of IGM has never been confirmed by direct imaging. Here we report our observation of the Lyαmore » blob 2 (LAB2) in SSA22 with the Cosmic Web Imager (CWI). This is an integral field spectrograph optimized for low surface brightness, extended emission. With 22 hr of total on- and off-source exposure, CWI has revealed that LAB2 has extended Lyα emission that is organized into azimuthal zones consistent with filaments. We perform numerous tests with simulations and the data to secure the robustness of this result, which relies on data with modest signal-to-noise ratios. We have developed a smoothing algorithm that permits visualization of data cube slices along image or spectral image planes. With both raw and smoothed data cubes we demonstrate that the filaments are kinematically associated with LAB2 and display double-peaked profiles characteristic of optically thick Lyα emission. The flux is 10-20 times brighter than expected for the average emission from the IGM but is consistent with boosted fluorescence from a buried QSO or gravitation cooling radiation. Using simple emission models, we infer a baryon mass in the filaments of at least 1-4 × 10{sup 11} M {sub ☉}, and the dark halo mass is at least 2 × 10{sup 12} M {sub ☉}. The spatial-kinematic morphology is more consistent with inflow from the cosmic web than outflow from LAB2, although an outflow feature maybe present at one azimuth. LAB2 and the surrounding gas have significant and coaligned angular momentum, strengthening the case for their association.« less

  11. VORTICAL MOTIONS OF BARYONIC GAS IN THE COSMIC WEB: GROWTH HISTORY AND SCALING RELATION

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Zhu, Weishan; Feng, Long-long

    The vortical motions of the baryonic gas residing in large-scale structures are investigated by cosmological hydrodynamic simulations. Proceeding in the formation of the cosmic web, the vortical motions of baryonic matter are pumped up by baroclinity in two stages, i.e., the formation of sheets and filaments. The mean curl velocities are about <1, 1–10, 10–150, and 5–50 km s{sup −1} in voids, sheets, filaments, and knots at z = 0, respectively. The scaling of the vortical velocity of gas can be well described by the She–Leveque hierarchical turbulence model in the range of l < 0.65(1.50) h{sup −1} Mpc inmore » a simulation with a box of size 25(100) h{sup −1} Mpc. The fractal Hausdorff dimension of vortical motions, d, revealed by velocity structure functions, is ∼2.1–2.3(∼1.8–2.1). It is slightly larger than the fractal dimension of mass distribution in filaments, D{sup f} ∼ 1.9–2.2, and smaller than the fractal dimension of sheets, D{sup s} ∼ 2.4–2.7. The vortical kinetic energy of baryonic gas is mainly transported by filaments. Both scalings of mass distribution and vortical velocity increments show distinctive transitions at the turning scale of ∼0.65(1.50) h{sup −1} Mpc, which may be closely related to the characteristic radius of density filaments.« less

  12. Large Scale Data Analytics of User Behavior for Improving Content Delivery

    DTIC Science & Technology

    2014-12-01

    video streaming, web browsing To Achan and Amma iv Abstract The Internet is fast becoming the de facto content delivery network of the world...operators everywhere and they seek to de - sign and manage their networks better to improve content delivery and provide better quality of experience...Anjali, Kriti and Ruta have been great company and have let me partake in their delicious homecooked food more times than I can remember. My friends

  13. RevEcoR: an R package for the reverse ecology analysis of microbiomes.

    PubMed

    Cao, Yang; Wang, Yuanyuan; Zheng, Xiaofei; Li, Fei; Bo, Xiaochen

    2016-07-29

    All species live in complex ecosystems. The structure and complexity of a microbial community reflects not only diversity and function, but also the environment in which it occurs. However, traditional ecological methods can only be applied on a small scale and for relatively well-understood biological systems. Recently, a graph-theory-based algorithm called the reverse ecology approach has been developed that can analyze the metabolic networks of all the species in a microbial community, and predict the metabolic interface between species and their environment. Here, we present RevEcoR, an R package and a Shiny Web application that implements the reverse ecology algorithm for determining microbe-microbe interactions in microbial communities. This software allows users to obtain large-scale ecological insights into species' ecology directly from high-throughput metagenomic data. The software has great potential for facilitating the study of microbiomes. RevEcoR is open source software for the study of microbial community ecology. The RevEcoR R package is freely available under the GNU General Public License v. 2.0 at http://cran.r-project.org/web/packages/RevEcoR/ with the vignette and typical usage examples, and the interactive Shiny web application is available at http://yiluheihei.shinyapps.io/shiny-RevEcoR , or can be installed locally with the source code accessed from https://github.com/yiluheihei/shiny-RevEcoR .

  14. Reverse engineering and analysis of large genome-scale gene networks

    PubMed Central

    Aluru, Maneesha; Zola, Jaroslaw; Nettleton, Dan; Aluru, Srinivas

    2013-01-01

    Reverse engineering the whole-genome networks of complex multicellular organisms continues to remain a challenge. While simpler models easily scale to large number of genes and gene expression datasets, more accurate models are compute intensive limiting their scale of applicability. To enable fast and accurate reconstruction of large networks, we developed Tool for Inferring Network of Genes (TINGe), a parallel mutual information (MI)-based program. The novel features of our approach include: (i) B-spline-based formulation for linear-time computation of MI, (ii) a novel algorithm for direct permutation testing and (iii) development of parallel algorithms to reduce run-time and facilitate construction of large networks. We assess the quality of our method by comparison with ARACNe (Algorithm for the Reconstruction of Accurate Cellular Networks) and GeneNet and demonstrate its unique capability by reverse engineering the whole-genome network of Arabidopsis thaliana from 3137 Affymetrix ATH1 GeneChips in just 9 min on a 1024-core cluster. We further report on the development of a new software Gene Network Analyzer (GeNA) for extracting context-specific subnetworks from a given set of seed genes. Using TINGe and GeNA, we performed analysis of 241 Arabidopsis AraCyc 8.0 pathways, and the results are made available through the web. PMID:23042249

  15. Characteristics and evolution of the ecosystem of software tools supporting research in molecular biology.

    PubMed

    Pazos, Florencio; Chagoyen, Monica

    2018-01-16

    Daily work in molecular biology presently depends on a large number of computational tools. An in-depth, large-scale study of that 'ecosystem' of Web tools, its characteristics, interconnectivity, patterns of usage/citation, temporal evolution and rate of decay is crucial for understanding the forces that shape it and for informing initiatives aimed at its funding, long-term maintenance and improvement. In particular, the long-term maintenance of these tools is compromised because of their specific development model. Hundreds of published studies become irreproducible de facto, as the software tools used to conduct them become unavailable. In this study, we present a large-scale survey of >5400 publications describing Web servers within the two main bibliographic resources for disseminating new software developments in molecular biology. For all these servers, we studied their citation patterns, the subjects they address, their citation networks and the temporal evolution of these factors. We also analysed how these factors affect the availability of these servers (whether they are alive). Our results show that this ecosystem of tools is highly interconnected and adapts to the 'trendy' subjects in every moment. The servers present characteristic temporal patterns of citation/usage, and there is a worrying rate of server 'death', which is influenced by factors such as the server popularity and the institutions that hosts it. These results can inform initiatives aimed at the long-term maintenance of these resources. © The Author(s) 2018. Published by Oxford University Press. All rights reserved. For Permissions, please email: journals.permissions@oup.com.

  16. The part and the whole: voids, supervoids, and their ISW imprint

    NASA Astrophysics Data System (ADS)

    Kovács, András

    2018-04-01

    The integrated Sachs-Wolfe (ISW) imprint of extreme structures in the cosmic web probes the dynamical nature of dark energy. Looking through typical cosmic voids, no anomalous signal has been reported. On the contrary, supervoids, associated with large-scale fluctuations in the gravitational potential, have shown potentially disturbing excess signals. In this study, we used the Jubilee ISW simulation to demonstrate how the stacked signal depends on the void definition. We found that large underdensities, with at least ≈5 merged sub-voids, show a peculiar ISW imprint shape with central cold spots and surrounding hot rings, offering a natural way to define supervoids in the cosmic web. We then inspected the real-world Baryon Oscillations Spectroscopic Survey data release 12 (BOSS DR12) data using the simulated imprints as templates. The imprinted profile of BOSS supervoids appears to be more compact than in simulations, requiring an extra α ≈ 0.7 re-scaling of filter sizes. The data reveal an excess ISW-like signal with AISW ≈ 9 amplitude at the ≈2.5σ significance level, unlike previous studies that used isolated voids and reported good consistency with AISW = 1. The tension with the Jubilee-based Λ cold dark matter predictions is ≳2σ, in consistency with independent analyses of supervoids in Dark Energy Survey data. We show that such a very large enhancement of the AISW parameter hints at a possible causal relation between the cosmic microwave background Cold Spot and the Eridanus supervoid. The origin of these findings remains unclear.

  17. The potential of the internet.

    PubMed

    Coleman, Jamie J; McDowell, Sarah E

    2012-06-01

    The internet and the World Wide Web have changed the ways that we function. As technologies grow and adapt, there is a huge potential for the internet to affect drug research and development, as well as many other aspects of clinical pharmacology. We review some of the areas of interest to date and discuss some of the potential areas in which internet-based technology can be exploited. Information retrieval from the web by health-care professionals is common, and bringing evidence-based medicine to the bedside affects the care of patients. As a primary research tool the web can provide a vast array of information in generating new ideas or exploring previous research findings. This has facilitated systematic reviewing, for example. The content of the web has become a subject of research in its own right. The web is also widely used as a research facilitator, including enhancement of communication between collaborators, provision of online research tools (such as questionnaires, management of large scale multicentre trials, registration of clinical trials) and distribution of information. Problems include information overload, ignorance of early data that are not indexed in databases, difficulties in keeping web sites up to date and assessing the validity of information retrieved. Some web-based activities are viewed with suspicion, including analysis by pharmaceutical companies of drug information to facilitate direct-to-consumer advertising of novel pharmaceuticals. Use of these technologies will continue to expand in often unexpected ways. Clinical pharmacologists must embrace internet technology and include it as a key priority in their research agenda. © 2012 The Authors. British Journal of Clinical Pharmacology © 2012 The British Pharmacological Society.

  18. The semantic web in translational medicine: current applications and future directions

    PubMed Central

    Machado, Catia M.; Rebholz-Schuhmann, Dietrich; Freitas, Ana T.; Couto, Francisco M.

    2015-01-01

    Semantic web technologies offer an approach to data integration and sharing, even for resources developed independently or broadly distributed across the web. This approach is particularly suitable for scientific domains that profit from large amounts of data that reside in the public domain and that have to be exploited in combination. Translational medicine is such a domain, which in addition has to integrate private data from the clinical domain with proprietary data from the pharmaceutical domain. In this survey, we present the results of our analysis of translational medicine solutions that follow a semantic web approach. We assessed these solutions in terms of their target medical use case; the resources covered to achieve their objectives; and their use of existing semantic web resources for the purposes of data sharing, data interoperability and knowledge discovery. The semantic web technologies seem to fulfill their role in facilitating the integration and exploration of data from disparate sources, but it is also clear that simply using them is not enough. It is fundamental to reuse resources, to define mappings between resources, to share data and knowledge. All these aspects allow the instantiation of translational medicine at the semantic web-scale, thus resulting in a network of solutions that can share resources for a faster transfer of new scientific results into the clinical practice. The envisioned network of translational medicine solutions is on its way, but it still requires resolving the challenges of sharing protected data and of integrating semantic-driven technologies into the clinical practice. PMID:24197933

  19. The semantic web in translational medicine: current applications and future directions.

    PubMed

    Machado, Catia M; Rebholz-Schuhmann, Dietrich; Freitas, Ana T; Couto, Francisco M

    2015-01-01

    Semantic web technologies offer an approach to data integration and sharing, even for resources developed independently or broadly distributed across the web. This approach is particularly suitable for scientific domains that profit from large amounts of data that reside in the public domain and that have to be exploited in combination. Translational medicine is such a domain, which in addition has to integrate private data from the clinical domain with proprietary data from the pharmaceutical domain. In this survey, we present the results of our analysis of translational medicine solutions that follow a semantic web approach. We assessed these solutions in terms of their target medical use case; the resources covered to achieve their objectives; and their use of existing semantic web resources for the purposes of data sharing, data interoperability and knowledge discovery. The semantic web technologies seem to fulfill their role in facilitating the integration and exploration of data from disparate sources, but it is also clear that simply using them is not enough. It is fundamental to reuse resources, to define mappings between resources, to share data and knowledge. All these aspects allow the instantiation of translational medicine at the semantic web-scale, thus resulting in a network of solutions that can share resources for a faster transfer of new scientific results into the clinical practice. The envisioned network of translational medicine solutions is on its way, but it still requires resolving the challenges of sharing protected data and of integrating semantic-driven technologies into the clinical practice. © The Author 2013. Published by Oxford University Press.

  20. The effect of Web-based Braden Scale training on the reliability and precision of Braden Scale pressure ulcer risk assessments.

    PubMed

    Magnan, Morris A; Maklebust, Joann

    2008-01-01

    To evaluate the effect of Web-based Braden Scale training on the reliability and precision of pressure ulcer risk assessments made by registered nurses (RN) working in acute care settings. Pretest-posttest, 2-group, quasi-experimental design. Five hundred Braden Scale risk assessments were made on 102 acute care patients deemed to be at various levels of risk for pressure ulceration. Assessments were made by RNs working in acute care hospitals at 3 different medical centers where the Braden Scale was in regular daily use (2 medical centers) or new to the setting (1 medical center). The Braden Scale for Predicting Pressure Sore Risk was used to guide pressure ulcer risk assessments. A Web-based version of the Detroit Medical Center Braden Scale Computerized Training Module was used to teach nurses correct use of the Braden Scale and selection of risk-based pressure ulcer prevention interventions. In the aggregate, RN generated reliable Braden Scale pressure ulcer risk assessments 65% of the time after training. The effect of Web-based Braden Scale training on reliability and precision of assessments varied according to familiarity with the scale. With training, new users of the scale made reliable assessments 84% of the time and significantly improved precision of their assessments. The reliability and precision of Braden Scale risk assessments made by its regular users was unaffected by training. Technology-assisted Braden Scale training improved both reliability and precision of risk assessments made by new users of the scale, but had virtually no effect on the reliability or precision of risk assessments made by regular users of the instrument. Further research is needed to determine best approaches for improving reliability and precision of Braden Scale assessments made by its regular users.

  1. A Web-Based Physical Activity Intervention for Spanish-Speaking Latinas: A Costs and Cost-Effectiveness Analysis.

    PubMed

    Larsen, Britta; Marcus, Bess; Pekmezi, Dori; Hartman, Sheri; Gilmer, Todd

    2017-02-22

    Latinas report particularly low levels of physical activity and suffer from greater rates of lifestyle-related conditions such as obesity and diabetes. Interventions are needed that can increase physical activity in this growing population in a large-scale, cost-effective manner. Web-based interventions may have potential given the increase in Internet use among Latinas and the scalability of Web-based programs. To examine the costs and cost-effectiveness of a Web-based, Spanish-language physical activity intervention for Latinas compared to a wellness contact control. Healthy adult Latina women (N=205) were recruited from the community and randomly assigned to receive a Spanish-language, Web-based, individually tailored physical activity intervention (intervention group) or were given access to a website with content on wellness topics other than physical activity (control group). Physical activity was measured using the 7-Day Physical Activity Recall interview and ActiGraph accelerometers at baseline, 6 months (ie, postintervention), and 12 months (ie, maintenance phase). Costs were estimated from a payer perspective and included all features necessary to implement the intervention in a community setting, including staff time (ie, wages, benefits, and overhead), materials, hardware, website hosting, and routine website maintenance. At 6 months, the costs of running the intervention and control groups were US $17 and US $8 per person per month, respectively. These costs fell to US $12 and US $6 per person per month at 12 months, respectively. Linear interpolation showed that intervention participants increased their physical activity by 1362 total minutes at 6 months (523 minutes by accelerometer) compared to 715 minutes for control participants (186 minutes by accelerometer). At 6 months, each minute increase in physical activity for the intervention group cost US $0.08 (US $0.20 by accelerometer) compared to US $0.07 for control participants (US $0.26 by accelerometer). Incremental cost-per-minute increases associated with the intervention were US $0.08 at 6 months and US $0.04 at 12 months (US $0.16 and US $0.08 by accelerometer, respectively). Sensitivity analyses showed variations in staffing costs or intervention effectiveness yielded only modest changes in incremental costs. While the Web-based physical activity intervention was more expensive than the wellness control, both were quite low cost compared to face-to-face or mail-delivered interventions. Cost-effectiveness ranged markedly based on physical activity measure and was similar between the two conditions. Overall, the Web-based intervention was effective and low cost, suggesting a promising channel for increasing physical activity on a large scale in this at-risk population. Clinicaltrials.gov NCT01834287; https://clinicaltrials.gov/ct2/show/NCT01834287 (Archived by WebCite at http://www.webcitation.org/6nyjX9Jrh). ©Britta Larsen, Bess Marcus, Dori Pekmezi, Sheri Hartman, Todd Gilmer. Originally published in the Journal of Medical Internet Research (http://www.jmir.org), 22.02.2017.

  2. A Web-Based Physical Activity Intervention for Spanish-Speaking Latinas: A Costs and Cost-Effectiveness Analysis

    PubMed Central

    Marcus, Bess; Pekmezi, Dori; Hartman, Sheri; Gilmer, Todd

    2017-01-01

    Background Latinas report particularly low levels of physical activity and suffer from greater rates of lifestyle-related conditions such as obesity and diabetes. Interventions are needed that can increase physical activity in this growing population in a large-scale, cost-effective manner. Web-based interventions may have potential given the increase in Internet use among Latinas and the scalability of Web-based programs. Objective To examine the costs and cost-effectiveness of a Web-based, Spanish-language physical activity intervention for Latinas compared to a wellness contact control. Methods Healthy adult Latina women (N=205) were recruited from the community and randomly assigned to receive a Spanish-language, Web-based, individually tailored physical activity intervention (intervention group) or were given access to a website with content on wellness topics other than physical activity (control group). Physical activity was measured using the 7-Day Physical Activity Recall interview and ActiGraph accelerometers at baseline, 6 months (ie, postintervention), and 12 months (ie, maintenance phase). Costs were estimated from a payer perspective and included all features necessary to implement the intervention in a community setting, including staff time (ie, wages, benefits, and overhead), materials, hardware, website hosting, and routine website maintenance. Results At 6 months, the costs of running the intervention and control groups were US $17 and US $8 per person per month, respectively. These costs fell to US $12 and US $6 per person per month at 12 months, respectively. Linear interpolation showed that intervention participants increased their physical activity by 1362 total minutes at 6 months (523 minutes by accelerometer) compared to 715 minutes for control participants (186 minutes by accelerometer). At 6 months, each minute increase in physical activity for the intervention group cost US $0.08 (US $0.20 by accelerometer) compared to US $0.07 for control participants (US $0.26 by accelerometer). Incremental cost-per-minute increases associated with the intervention were US $0.08 at 6 months and US $0.04 at 12 months (US $0.16 and US $0.08 by accelerometer, respectively). Sensitivity analyses showed variations in staffing costs or intervention effectiveness yielded only modest changes in incremental costs. Conclusions While the Web-based physical activity intervention was more expensive than the wellness control, both were quite low cost compared to face-to-face or mail-delivered interventions. Cost-effectiveness ranged markedly based on physical activity measure and was similar between the two conditions. Overall, the Web-based intervention was effective and low cost, suggesting a promising channel for increasing physical activity on a large scale in this at-risk population. ClinicalTrial Clinicaltrials.gov NCT01834287; https://clinicaltrials.gov/ct2/show/NCT01834287 (Archived by WebCite at http://www.webcitation.org/6nyjX9Jrh) PMID:28228368

  3. Impacts of elevated terrestrial nutrient loads and temperature on pelagic food-web efficiency and fish production.

    PubMed

    Lefébure, R; Degerman, R; Andersson, A; Larsson, S; Eriksson, L-O; Båmstedt, U; Byström, P

    2013-05-01

    Both temperature and terrestrial organic matter have strong impacts on aquatic food-web dynamics and production. Temperature affects vital rates of all organisms, and terrestrial organic matter can act both as an energy source for lower trophic levels, while simultaneously reducing light availability for autotrophic production. As climate change predictions for the Baltic Sea and elsewhere suggest increases in both terrestrial matter runoff and increases in temperature, we studied the effects on pelagic food-web dynamics and food-web efficiency in a plausible future scenario with respect to these abiotic variables in a large-scale mesocosm experiment. Total basal (phytoplankton plus bacterial) production was slightly reduced when only increasing temperatures, but was otherwise similar across all other treatments. Separate increases in nutrient loads and temperature decreased the ratio of autotrophic:heterotrophic production, but the combined treatment of elevated temperature and terrestrial nutrient loads increased both fish production and food-web efficiency. CDOM: Chl a ratios strongly indicated that terrestrial and not autotrophic carbon was the main energy source in these food webs and our results also showed that zooplankton biomass was positively correlated with increased bacterial production. Concomitantly, biomass of the dominant calanoid copepod Acartia sp. increased as an effect of increased temperature. As the combined effects of increased temperature and terrestrial organic nutrient loads were required to increase zooplankton abundance and fish production, conclusions about effects of climate change on food-web dynamics and fish production must be based on realistic combinations of several abiotic factors. Moreover, our results question established notions on the net inefficiency of heterotrophic carbon transfer to the top of the food web. © 2013 Blackwell Publishing Ltd.

  4. An open source web interface for linking models to infrastructure system databases

    NASA Astrophysics Data System (ADS)

    Knox, S.; Mohamed, K.; Harou, J. J.; Rheinheimer, D. E.; Medellin-Azuara, J.; Meier, P.; Tilmant, A.; Rosenberg, D. E.

    2016-12-01

    Models of networked engineered resource systems such as water or energy systems are often built collaboratively with developers from different domains working at different locations. These models can be linked to large scale real world databases, and they are constantly being improved and extended. As the development and application of these models becomes more sophisticated, and the computing power required for simulations and/or optimisations increases, so has the need for online services and tools which enable the efficient development and deployment of these models. Hydra Platform is an open source, web-based data management system, which allows modellers of network-based models to remotely store network topology and associated data in a generalised manner, allowing it to serve multiple disciplines. Hydra Platform uses a web API using JSON to allow external programs (referred to as `Apps') to interact with its stored networks and perform actions such as importing data, running models, or exporting the networks to different formats. Hydra Platform supports multiple users accessing the same network and has a suite of functions for managing users and data. We present ongoing development in Hydra Platform, the Hydra Web User Interface, through which users can collaboratively manage network data and models in a web browser. The web interface allows multiple users to graphically access, edit and share their networks, run apps and view results. Through apps, which are located on the server, the web interface can give users access to external data sources and models without the need to install or configure any software. This also ensures model results can be reproduced by removing platform or version dependence. Managing data and deploying models via the web interface provides a way for multiple modellers to collaboratively manage data, deploy and monitor model runs and analyse results.

  5. Tracing the cosmic web

    NASA Astrophysics Data System (ADS)

    Libeskind, Noam I.; van de Weygaert, Rien; Cautun, Marius; Falck, Bridget; Tempel, Elmo; Abel, Tom; Alpaslan, Mehmet; Aragón-Calvo, Miguel A.; Forero-Romero, Jaime E.; Gonzalez, Roberto; Gottlöber, Stefan; Hahn, Oliver; Hellwing, Wojciech A.; Hoffman, Yehuda; Jones, Bernard J. T.; Kitaura, Francisco; Knebe, Alexander; Manti, Serena; Neyrinck, Mark; Nuza, Sebastián E.; Padilla, Nelson; Platen, Erwin; Ramachandra, Nesar; Robotham, Aaron; Saar, Enn; Shandarin, Sergei; Steinmetz, Matthias; Stoica, Radu S.; Sousbie, Thierry; Yepes, Gustavo

    2018-01-01

    The cosmic web is one of the most striking features of the distribution of galaxies and dark matter on the largest scales in the Universe. It is composed of dense regions packed full of galaxies, long filamentary bridges, flattened sheets and vast low-density voids. The study of the cosmic web has focused primarily on the identification of such features, and on understanding the environmental effects on galaxy formation and halo assembly. As such, a variety of different methods have been devised to classify the cosmic web - depending on the data at hand, be it numerical simulations, large sky surveys or other. In this paper, we bring 12 of these methods together and apply them to the same data set in order to understand how they compare. In general, these cosmic-web classifiers have been designed with different cosmological goals in mind, and to study different questions. Therefore, one would not a priori expect agreement between different techniques; however, many of these methods do converge on the identification of specific features. In this paper, we study the agreements and disparities of the different methods. For example, each method finds that knots inhabit higher density regions than filaments, etc. and that voids have the lowest densities. For a given web environment, we find a substantial overlap in the density range assigned by each web classification scheme. We also compare classifications on a halo-by-halo basis; for example, we find that 9 of 12 methods classify around a third of group-mass haloes (i.e. Mhalo ∼ 1013.5 h-1 M⊙) as being in filaments. Lastly, so that any future cosmic-web classification scheme can be compared to the 12 methods used here, we have made all the data used in this paper public.

  6. Mining the Mind Research Network: A Novel Framework for Exploring Large Scale, Heterogeneous Translational Neuroscience Research Data Sources

    PubMed Central

    Bockholt, Henry J.; Scully, Mark; Courtney, William; Rachakonda, Srinivas; Scott, Adam; Caprihan, Arvind; Fries, Jill; Kalyanam, Ravi; Segall, Judith M.; de la Garza, Raul; Lane, Susan; Calhoun, Vince D.

    2009-01-01

    A neuroinformatics (NI) system is critical to brain imaging research in order to shorten the time between study conception and results. Such a NI system is required to scale well when large numbers of subjects are studied. Further, when multiple sites participate in research projects organizational issues become increasingly difficult. Optimized NI applications mitigate these problems. Additionally, NI software enables coordination across multiple studies, leveraging advantages potentially leading to exponential research discoveries. The web-based, Mind Research Network (MRN), database system has been designed and improved through our experience with 200 research studies and 250 researchers from seven different institutions. The MRN tools permit the collection, management, reporting and efficient use of large scale, heterogeneous data sources, e.g., multiple institutions, multiple principal investigators, multiple research programs and studies, and multimodal acquisitions. We have collected and analyzed data sets on thousands of research participants and have set up a framework to automatically analyze the data, thereby making efficient, practical data mining of this vast resource possible. This paper presents a comprehensive framework for capturing and analyzing heterogeneous neuroscience research data sources that has been fully optimized for end-users to perform novel data mining. PMID:20461147

  7. Exploring the role of movement in determining the global distribution of marine biomass using a coupled hydrodynamic - Size-based ecosystem model

    NASA Astrophysics Data System (ADS)

    Watson, James R.; Stock, Charles A.; Sarmiento, Jorge L.

    2015-11-01

    Modeling the dynamics of marine populations at a global scale - from phytoplankton to fish - is necessary if we are to quantify how climate change and other broad-scale anthropogenic actions affect the supply of marine-based food. Here, we estimate the abundance and distribution of fish biomass using a simple size-based food web model coupled to simulations of global ocean physics and biogeochemistry. We focus on the spatial distribution of biomass, identifying highly productive regions - shelf seas, western boundary currents and major upwelling zones. In the absence of fishing, we estimate the total ocean fish biomass to be ∼ 2.84 ×109 tonnes, similar to previous estimates. However, this value is sensitive to the choice of parameters, and further, allowing fish to move had a profound impact on the spatial distribution of fish biomass and the structure of marine communities. In particular, when movement is implemented the viable range of large predators is greatly increased, and stunted biomass spectra characterizing large ocean regions in simulations without movement, are replaced with expanded spectra that include large predators. These results highlight the importance of considering movement in global-scale ecological models.

  8. Implications of Web Mercator and its Use in Online Mapping

    USGS Publications Warehouse

    Battersby, Sarah E.; Finn, Michael P.; Usery, E. Lynn; Yamamoto, Kristina H.

    2014-01-01

    Online interactive maps have become a popular means of communicating with spatial data. In most online mapping systems, Web Mercator has become the dominant projection. While the Mercator projection has a long history of discussion about its inappropriateness for general-purpose mapping, particularly at the global scale, and seems to have been virtually phased out for general-purpose global-scale print maps, it has seen a resurgence in popularity in Web Mercator form. This article theorizes on how Web Mercator came to be widely used for online maps and what this might mean in terms of data display, technical aspects of map generation and distribution, design, and cognition of spatial patterns. The authors emphasize details of where the projection excels and where it does not, as well as some of its advantages and disadvantages for cartographic communication, and conclude with some research directions that may help to develop better solutions to the problem of projections for general-purpose, multi-scale Web mapping.

  9. Experience versus talent shapes the structure of the Web.

    PubMed

    Kong, Joseph S; Sarshar, Nima; Roychowdhury, Vwani P

    2008-09-16

    We use sequential large-scale crawl data to empirically investigate and validate the dynamics that underlie the evolution of the structure of the web. We find that the overall structure of the web is defined by an intricate interplay between experience or entitlement of the pages (as measured by the number of inbound hyperlinks a page already has), inherent talent or fitness of the pages (as measured by the likelihood that someone visiting the page would give a hyperlink to it), and the continual high rates of birth and death of pages on the web. We find that the web is conservative in judging talent and the overall fitness distribution is exponential, showing low variability. The small variance in talent, however, is enough to lead to experience distributions with high variance: The preferential attachment mechanism amplifies these small biases and leads to heavy-tailed power-law (PL) inbound degree distributions over all pages, as well as over pages that are of the same age. The balancing act between experience and talent on the web allows newly introduced pages with novel and interesting content to grow quickly and surpass older pages. In this regard, it is much like what we observe in high-mobility and meritocratic societies: People with entitlement continue to have access to the best resources, but there is just enough screening for fitness that allows for talented winners to emerge and join the ranks of the leaders. Finally, we show that the fitness estimates have potential practical applications in ranking query results.

  10. A journey to Semantic Web query federation in the life sciences.

    PubMed

    Cheung, Kei-Hoi; Frost, H Robert; Marshall, M Scott; Prud'hommeaux, Eric; Samwald, Matthias; Zhao, Jun; Paschke, Adrian

    2009-10-01

    As interest in adopting the Semantic Web in the biomedical domain continues to grow, Semantic Web technology has been evolving and maturing. A variety of technological approaches including triplestore technologies, SPARQL endpoints, Linked Data, and Vocabulary of Interlinked Datasets have emerged in recent years. In addition to the data warehouse construction, these technological approaches can be used to support dynamic query federation. As a community effort, the BioRDF task force, within the Semantic Web for Health Care and Life Sciences Interest Group, is exploring how these emerging approaches can be utilized to execute distributed queries across different neuroscience data sources. We have created two health care and life science knowledge bases. We have explored a variety of Semantic Web approaches to describe, map, and dynamically query multiple datasets. We have demonstrated several federation approaches that integrate diverse types of information about neurons and receptors that play an important role in basic, clinical, and translational neuroscience research. Particularly, we have created a prototype receptor explorer which uses OWL mappings to provide an integrated list of receptors and executes individual queries against different SPARQL endpoints. We have also employed the AIDA Toolkit, which is directed at groups of knowledge workers who cooperatively search, annotate, interpret, and enrich large collections of heterogeneous documents from diverse locations. We have explored a tool called "FeDeRate", which enables a global SPARQL query to be decomposed into subqueries against the remote databases offering either SPARQL or SQL query interfaces. Finally, we have explored how to use the vocabulary of interlinked Datasets (voiD) to create metadata for describing datasets exposed as Linked Data URIs or SPARQL endpoints. We have demonstrated the use of a set of novel and state-of-the-art Semantic Web technologies in support of a neuroscience query federation scenario. We have identified both the strengths and weaknesses of these technologies. While Semantic Web offers a global data model including the use of Uniform Resource Identifiers (URI's), the proliferation of semantically-equivalent URI's hinders large scale data integration. Our work helps direct research and tool development, which will be of benefit to this community.

  11. A journey to Semantic Web query federation in the life sciences

    PubMed Central

    Cheung, Kei-Hoi; Frost, H Robert; Marshall, M Scott; Prud'hommeaux, Eric; Samwald, Matthias; Zhao, Jun; Paschke, Adrian

    2009-01-01

    Background As interest in adopting the Semantic Web in the biomedical domain continues to grow, Semantic Web technology has been evolving and maturing. A variety of technological approaches including triplestore technologies, SPARQL endpoints, Linked Data, and Vocabulary of Interlinked Datasets have emerged in recent years. In addition to the data warehouse construction, these technological approaches can be used to support dynamic query federation. As a community effort, the BioRDF task force, within the Semantic Web for Health Care and Life Sciences Interest Group, is exploring how these emerging approaches can be utilized to execute distributed queries across different neuroscience data sources. Methods and results We have created two health care and life science knowledge bases. We have explored a variety of Semantic Web approaches to describe, map, and dynamically query multiple datasets. We have demonstrated several federation approaches that integrate diverse types of information about neurons and receptors that play an important role in basic, clinical, and translational neuroscience research. Particularly, we have created a prototype receptor explorer which uses OWL mappings to provide an integrated list of receptors and executes individual queries against different SPARQL endpoints. We have also employed the AIDA Toolkit, which is directed at groups of knowledge workers who cooperatively search, annotate, interpret, and enrich large collections of heterogeneous documents from diverse locations. We have explored a tool called "FeDeRate", which enables a global SPARQL query to be decomposed into subqueries against the remote databases offering either SPARQL or SQL query interfaces. Finally, we have explored how to use the vocabulary of interlinked Datasets (voiD) to create metadata for describing datasets exposed as Linked Data URIs or SPARQL endpoints. Conclusion We have demonstrated the use of a set of novel and state-of-the-art Semantic Web technologies in support of a neuroscience query federation scenario. We have identified both the strengths and weaknesses of these technologies. While Semantic Web offers a global data model including the use of Uniform Resource Identifiers (URI's), the proliferation of semantically-equivalent URI's hinders large scale data integration. Our work helps direct research and tool development, which will be of benefit to this community. PMID:19796394

  12. CAS-viewer: web-based tool for splicing-guided integrative analysis of multi-omics cancer data.

    PubMed

    Han, Seonggyun; Kim, Dongwook; Kim, Youngjun; Choi, Kanghoon; Miller, Jason E; Kim, Dokyoon; Lee, Younghee

    2018-04-20

    The Cancer Genome Atlas (TCGA) project is a public resource that provides transcriptomic, DNA sequence, methylation, and clinical data for 33 cancer types. Transforming the large size and high complexity of TCGA cancer genome data into integrated knowledge can be useful to promote cancer research. Alternative splicing (AS) is a key regulatory mechanism of genes in human cancer development and in the interaction with epigenetic factors. Therefore, AS-guided integration of existing TCGA data sets will make it easier to gain insight into the genetic architecture of cancer risk and related outcomes. There are already existing tools analyzing and visualizing alternative mRNA splicing patterns for large-scale RNA-seq experiments. However, these existing web-based tools are limited to the analysis of individual TCGA data sets at a time, such as only transcriptomic information. We implemented CAS-viewer (integrative analysis of Cancer genome data based on Alternative Splicing), a web-based tool leveraging multi-cancer omics data from TCGA. It illustrates alternative mRNA splicing patterns along with methylation, miRNAs, and SNPs, and then provides an analysis tool to link differential transcript expression ratio to methylation, miRNA, and splicing regulatory elements for 33 cancer types. Moreover, one can analyze AS patterns with clinical data to identify potential transcripts associated with different survival outcome for each cancer. CAS-viewer is a web-based application for transcript isoform-driven integration of multi-omics data in multiple cancer types and will aid in the visualization and possible discovery of biomarkers for cancer by integrating multi-omics data from TCGA.

  13. Developing A Large-Scale, Collaborative, Productive Geoscience Education Network

    NASA Astrophysics Data System (ADS)

    Manduca, C. A.; Bralower, T. J.; Egger, A. E.; Fox, S.; Ledley, T. S.; Macdonald, H.; Mcconnell, D. A.; Mogk, D. W.; Tewksbury, B. J.

    2012-12-01

    Over the past 15 years, the geoscience education community has grown substantially and developed broad and deep capacity for collaboration and dissemination of ideas. While this community is best viewed as emergent from complex interactions among changing educational needs and opportunities, we highlight the role of several large projects in the development of a network within this community. In the 1990s, three NSF projects came together to build a robust web infrastructure to support the production and dissemination of on-line resources: On The Cutting Edge (OTCE), Earth Exploration Toolbook, and Starting Point: Teaching Introductory Geoscience. Along with the contemporaneous Digital Library for Earth System Education, these projects engaged geoscience educators nationwide in exploring professional development experiences that produced lasting on-line resources, collaborative authoring of resources, and models for web-based support for geoscience teaching. As a result, a culture developed in the 2000s in which geoscience educators anticipated that resources for geoscience teaching would be shared broadly and that collaborative authoring would be productive and engaging. By this time, a diverse set of examples demonstrated the power of the web infrastructure in supporting collaboration, dissemination and professional development . Building on this foundation, more recent work has expanded both the size of the network and the scope of its work. Many large research projects initiated collaborations to disseminate resources supporting educational use of their data. Research results from the rapidly expanding geoscience education research community were integrated into the Pedagogies in Action website and OTCE. Projects engaged faculty across the nation in large-scale data collection and educational research. The Climate Literacy and Energy Awareness Network and OTCE engaged community members in reviewing the expanding body of on-line resources. Building Strong Geoscience Departments sought to create the same type of shared information base that was supporting individual faculty for departments. The Teach the Earth portal and its underlying web development tools were used by NSF-funded projects in education to disseminate their results. Leveraging these funded efforts, the Climate Literacy Network has expanded this geoscience education community to include individuals broadly interested in fostering climate literacy. Most recently, the InTeGrate project is implementing inter-institutional collaborative authoring, testing and evaluation of curricular materials. While these projects represent only a fraction of the activity in geoscience education, they are important drivers in the development of a large, national, coherent geoscience education network with the ability to collaborate and disseminate information effectively. Importantly, the community is open and defined by active participation. Key mechanisms for engagement have included alignment of project activities with participants needs and goals; productive face-to-face and virtual workshops, events, and series; stipends for completion of large products; and strong supporting staff to keep projects moving and assist with product production. One measure of its success is the adoption and adaptation of resources and models by emerging projects, which results in the continued growth of the network.

  14. SureChEMBL: a large-scale, chemically annotated patent document database.

    PubMed

    Papadatos, George; Davies, Mark; Dedman, Nathan; Chambers, Jon; Gaulton, Anna; Siddle, James; Koks, Richard; Irvine, Sean A; Pettersson, Joe; Goncharoff, Nicko; Hersey, Anne; Overington, John P

    2016-01-04

    SureChEMBL is a publicly available large-scale resource containing compounds extracted from the full text, images and attachments of patent documents. The data are extracted from the patent literature according to an automated text and image-mining pipeline on a daily basis. SureChEMBL provides access to a previously unavailable, open and timely set of annotated compound-patent associations, complemented with sophisticated combined structure and keyword-based search capabilities against the compound repository and patent document corpus; given the wealth of knowledge hidden in patent documents, analysis of SureChEMBL data has immediate applications in drug discovery, medicinal chemistry and other commercial areas of chemical science. Currently, the database contains 17 million compounds extracted from 14 million patent documents. Access is available through a dedicated web-based interface and data downloads at: https://www.surechembl.org/. © The Author(s) 2015. Published by Oxford University Press on behalf of Nucleic Acids Research.

  15. SureChEMBL: a large-scale, chemically annotated patent document database

    PubMed Central

    Papadatos, George; Davies, Mark; Dedman, Nathan; Chambers, Jon; Gaulton, Anna; Siddle, James; Koks, Richard; Irvine, Sean A.; Pettersson, Joe; Goncharoff, Nicko; Hersey, Anne; Overington, John P.

    2016-01-01

    SureChEMBL is a publicly available large-scale resource containing compounds extracted from the full text, images and attachments of patent documents. The data are extracted from the patent literature according to an automated text and image-mining pipeline on a daily basis. SureChEMBL provides access to a previously unavailable, open and timely set of annotated compound-patent associations, complemented with sophisticated combined structure and keyword-based search capabilities against the compound repository and patent document corpus; given the wealth of knowledge hidden in patent documents, analysis of SureChEMBL data has immediate applications in drug discovery, medicinal chemistry and other commercial areas of chemical science. Currently, the database contains 17 million compounds extracted from 14 million patent documents. Access is available through a dedicated web-based interface and data downloads at: https://www.surechembl.org/. PMID:26582922

  16. WEbcoli: an interactive and asynchronous web application for in silico design and analysis of genome-scale E.coli model.

    PubMed

    Jung, Tae-Sung; Yeo, Hock Chuan; Reddy, Satty G; Cho, Wan-Sup; Lee, Dong-Yup

    2009-11-01

    WEbcoli is a WEb application for in silico designing, analyzing and engineering Escherichia coli metabolism. It is devised and implemented using advanced web technologies, thereby leading to enhanced usability and dynamic web accessibility. As a main feature, the WEbcoli system provides a user-friendly rich web interface, allowing users to virtually design and synthesize mutant strains derived from the genome-scale wild-type E.coli model and to customize pathways of interest through a graph editor. In addition, constraints-based flux analysis can be conducted for quantifying metabolic fluxes and charactering the physiological and metabolic states under various genetic and/or environmental conditions. WEbcoli is freely accessible at http://webcoli.org. cheld@nus.edu.sg.

  17. The effect of Web-based Braden Scale training on the reliability of Braden subscale ratings.

    PubMed

    Magnan, Morris A; Maklebust, JoAnn

    2009-01-01

    The primary purpose of this study was to evaluate the effect of Web-based Braden Scale training on the reliability of Braden Scale subscale ratings made by nurses working in acute care hospitals. A secondary purpose was to describe the distribution of reliable Braden subscale ratings before and after Web-based Braden Scale training. Secondary analysis of data from a recently completed quasi-experimental, pretest-posttest, interrater reliability study. A convenience sample of RNs working at 3 Michigan medical centers voluntarily participated in the study. RN participants included nurses who used the Braden Scale regularly at their place of employment ("regular users") as well as nurses who did not use the Braden Scale at their place of employment ("new users"). Using a pretest-posttest, quasi-experimental design, pretest interrater reliability data were collected to identify the percentage of nurses making reliable Braden subscale assessments. Nurses then completed a Web-based Braden Scale training module after which posttest interrater reliability data were collected. The reliability of nurses' Braden subscale ratings was determined by examining the level of agreement/disagreement between ratings made by an RN and an "expert" rating the same patient. In total, 381 RN-to-expert dyads were available for analysis. During both the pretest and posttest periods, the percentage of reliable subscale ratings was highest for the activity subscale, lowest for the moisture subscale, and second lowest for the nutrition subscale. With Web-based Braden Scale training, the percentage of reliable Braden subscale ratings made by new users increased for all 6 subscales with statistically significant improvements in the percentage of reliable assessments made on 3 subscales: sensory-perception, moisture, and mobility. Training had virtually no effect on the percentage of reliable subscale ratings made by regular users of the Braden Scale. With Web-based Braden Scale training the percentage of nurses making reliable ratings increased for all 6 subscales, but this was true for new users only. Additional research is needed to identify educational approaches that effectively improve and sustain the reliability of subscale ratings among regular users of the Braden Scale. Moreover, special attention needs to be given to ensuring that all nurses working with the Braden Scale have a clear understanding of the intended meanings and correct approaches to rating moisture and nutrition subscales.

  18. Screensaver: an open source lab information management system (LIMS) for high throughput screening facilities

    PubMed Central

    2010-01-01

    Background Shared-usage high throughput screening (HTS) facilities are becoming more common in academe as large-scale small molecule and genome-scale RNAi screening strategies are adopted for basic research purposes. These shared facilities require a unique informatics infrastructure that must not only provide access to and analysis of screening data, but must also manage the administrative and technical challenges associated with conducting numerous, interleaved screening efforts run by multiple independent research groups. Results We have developed Screensaver, a free, open source, web-based lab information management system (LIMS), to address the informatics needs of our small molecule and RNAi screening facility. Screensaver supports the storage and comparison of screening data sets, as well as the management of information about screens, screeners, libraries, and laboratory work requests. To our knowledge, Screensaver is one of the first applications to support the storage and analysis of data from both genome-scale RNAi screening projects and small molecule screening projects. Conclusions The informatics and administrative needs of an HTS facility may be best managed by a single, integrated, web-accessible application such as Screensaver. Screensaver has proven useful in meeting the requirements of the ICCB-Longwood/NSRB Screening Facility at Harvard Medical School, and has provided similar benefits to other HTS facilities. PMID:20482787

  19. Screensaver: an open source lab information management system (LIMS) for high throughput screening facilities.

    PubMed

    Tolopko, Andrew N; Sullivan, John P; Erickson, Sean D; Wrobel, David; Chiang, Su L; Rudnicki, Katrina; Rudnicki, Stewart; Nale, Jennifer; Selfors, Laura M; Greenhouse, Dara; Muhlich, Jeremy L; Shamu, Caroline E

    2010-05-18

    Shared-usage high throughput screening (HTS) facilities are becoming more common in academe as large-scale small molecule and genome-scale RNAi screening strategies are adopted for basic research purposes. These shared facilities require a unique informatics infrastructure that must not only provide access to and analysis of screening data, but must also manage the administrative and technical challenges associated with conducting numerous, interleaved screening efforts run by multiple independent research groups. We have developed Screensaver, a free, open source, web-based lab information management system (LIMS), to address the informatics needs of our small molecule and RNAi screening facility. Screensaver supports the storage and comparison of screening data sets, as well as the management of information about screens, screeners, libraries, and laboratory work requests. To our knowledge, Screensaver is one of the first applications to support the storage and analysis of data from both genome-scale RNAi screening projects and small molecule screening projects. The informatics and administrative needs of an HTS facility may be best managed by a single, integrated, web-accessible application such as Screensaver. Screensaver has proven useful in meeting the requirements of the ICCB-Longwood/NSRB Screening Facility at Harvard Medical School, and has provided similar benefits to other HTS facilities.

  20. Biological impacts of local vs. regional land use on a small tributary of the Seine River (France): insights from a food web approach based on stable isotopes.

    PubMed

    Hette-Tronquart, Nicolas; Oberdorff, Thierry; Tales, Evelyne; Zahm, Amandine; Belliard, Jérôme

    2017-03-23

    As part of the landscape, streams are influenced by land use. Here, we contributed to the understanding of the biological impacts of land use on streams, investigating how landscape effects vary with spatial scales (local vs. regional). We adopted a food web approach integrating both biological structure and functioning, to focus on the overall effect of land use on stream biocœnosis. We selected 17 sites of a small tributary of the Seine River (France) for their contrasted land use, and conducted a natural experiment by sampling three organic matter sources, three macroinvertebrate taxa, and most of the fish community. Using stable isotope analysis, we calculated three food web metrics evaluating two major dimensions of the trophic diversity displayed by the fish community: (i) the diversity of exploited resources and (ii) the trophic level richness. The idea was to examine whether (1) land-use effects varied according to spatial scales, (2) land use affected food webs through an effect on community structure and (3) land use affected food webs through an effect on available resources. Beside an increase in trophic diversity from upstream to downstream, our empirical data showed that food webs were influenced by land use in the riparian corridors (local scale). The effect was complex, and depended on site's position along the upstream-downstream gradient. By contrast, land use in the catchment (regional scale) did not influence stream biocœnosis. At the local scale, community structure was weakly influenced by land use, and thus played a minor role in explaining food web modifications. Our results suggested that the amount of available resources at the base of the food web was partly responsible for food web modifications. In addition, changes in biological functioning (i.e. feeding interactions) can also explain another part of the land-use effect. These results highlight the role played by the riparian corridors as a buffer zone, and advocate that riparian corridor should be at the centre of water management attention.

  1. Food-web dynamics in a large river discontinuum

    USGS Publications Warehouse

    Cross, Wyatt F.; Baxter, Colden V.; Rosi-Marshall, Emma J.; Hall, Robert O.; Kennedy, Theodore A.; Donner, Kevin C.; Kelly, Holly A. Wellard; Seegert, Sarah E.Z.; Behn, Kathrine E.; Yard, Michael D.

    2013-01-01

    Nearly all ecosystems have been altered by human activities, and most communities are now composed of interacting species that have not co-evolved. These changes may modify species interactions, energy and material flows, and food-web stability. Although structural changes to ecosystems have been widely reported, few studies have linked such changes to dynamic food-web attributes and patterns of energy flow. Moreover, there have been few tests of food-web stability theory in highly disturbed and intensely managed freshwater ecosystems. Such synthetic approaches are needed for predicting the future trajectory of ecosystems, including how they may respond to natural or anthropogenic perturbations. We constructed flow food webs at six locations along a 386-km segment of the Colorado River in Grand Canyon (Arizona, USA) for three years. We characterized food-web structure and production, trophic basis of production, energy efficiencies, and interaction-strength distributions across a spatial gradient of perturbation (i.e., distance from Glen Canyon Dam), as well as before and after an experimental flood. We found strong longitudinal patterns in food-web characteristics that strongly correlated with the spatial position of large tributaries. Above tributaries, food webs were dominated by nonnative New Zealand mudsnails (62% of production) and nonnative rainbow trout (100% of fish production). The simple structure of these food webs led to few dominant energy pathways (diatoms to few invertebrate taxa to rainbow trout), large energy inefficiencies (i.e., Below large tributaries, invertebrate production declined ∼18-fold, while fish production remained similar to upstream sites and comprised predominately native taxa (80–100% of production). Sites below large tributaries had increasingly reticulate and detritus-based food webs with a higher prevalence of omnivory, as well as interaction strength distributions more typical of theoretically stable food webs (i.e., nearly twofold higher proportion of weak interactions). Consistent with theory, downstream food webs were less responsive to the experimental flood than sites closest to the dam. We show how human-induced shifts to food-web structure can affect energy flow and interaction strengths, and we show that these changes have consequences for food-web function and response to perturbations.

  2. From projected species distribution to food-web structure under climate change.

    PubMed

    Albouy, Camille; Velez, Laure; Coll, Marta; Colloca, Francesco; Le Loc'h, François; Mouillot, David; Gravel, Dominique

    2014-03-01

    Climate change is inducing deep modifications in species geographic ranges worldwide. However, the consequences of such changes on community structure are still poorly understood, particularly the impacts on food-web properties. Here, we propose a new framework, coupling species distribution and trophic models, to predict climate change impacts on food-web structure across the Mediterranean Sea. Sea surface temperature was used to determine the fish climate niches and their future distributions. Body size was used to infer trophic interactions between fish species. Our projections reveal that 54 fish species of 256 endemic and native species included in our analysis would disappear by 2080-2099 from the Mediterranean continental shelf. The number of feeding links between fish species would decrease on 73.4% of the continental shelf. However, the connectance of the overall fish web would increase on average, from 0.26 to 0.29, mainly due to a differential loss rate of feeding links and species richness. This result masks a systematic decrease in predator generality, estimated here as the number of prey species, from 30.0 to 25.4. Therefore, our study highlights large-scale impacts of climate change on marine food-web structure with potential deep consequences on ecosystem functioning. However, these impacts will likely be highly heterogeneous in space, challenging our current understanding of climate change impact on local marine ecosystems. © 2013 John Wiley & Sons Ltd.

  3. ICM: a web server for integrated clustering of multi-dimensional biomedical data.

    PubMed

    He, Song; He, Haochen; Xu, Wenjian; Huang, Xin; Jiang, Shuai; Li, Fei; He, Fuchu; Bo, Xiaochen

    2016-07-08

    Large-scale efforts for parallel acquisition of multi-omics profiling continue to generate extensive amounts of multi-dimensional biomedical data. Thus, integrated clustering of multiple types of omics data is essential for developing individual-based treatments and precision medicine. However, while rapid progress has been made, methods for integrated clustering are lacking an intuitive web interface that facilitates the biomedical researchers without sufficient programming skills. Here, we present a web tool, named Integrated Clustering of Multi-dimensional biomedical data (ICM), that provides an interface from which to fuse, cluster and visualize multi-dimensional biomedical data and knowledge. With ICM, users can explore the heterogeneity of a disease or a biological process by identifying subgroups of patients. The results obtained can then be interactively modified by using an intuitive user interface. Researchers can also exchange the results from ICM with collaborators via a web link containing a Project ID number that will directly pull up the analysis results being shared. ICM also support incremental clustering that allows users to add new sample data into the data of a previous study to obtain a clustering result. Currently, the ICM web server is available with no login requirement and at no cost at http://biotech.bmi.ac.cn/icm/. © The Author(s) 2016. Published by Oxford University Press on behalf of Nucleic Acids Research.

  4. Applying semantic web technologies for phenome-wide scan using an electronic health record linked Biobank

    PubMed Central

    2012-01-01

    Background The ability to conduct genome-wide association studies (GWAS) has enabled new exploration of how genetic variations contribute to health and disease etiology. However, historically GWAS have been limited by inadequate sample size due to associated costs for genotyping and phenotyping of study subjects. This has prompted several academic medical centers to form “biobanks” where biospecimens linked to personal health information, typically in electronic health records (EHRs), are collected and stored on a large number of subjects. This provides tremendous opportunities to discover novel genotype-phenotype associations and foster hypotheses generation. Results In this work, we study how emerging Semantic Web technologies can be applied in conjunction with clinical and genotype data stored at the Mayo Clinic Biobank to mine the phenotype data for genetic associations. In particular, we demonstrate the role of using Resource Description Framework (RDF) for representing EHR diagnoses and procedure data, and enable federated querying via standardized Web protocols to identify subjects genotyped for Type 2 Diabetes and Hypothyroidism to discover gene-disease associations. Our study highlights the potential of Web-scale data federation techniques to execute complex queries. Conclusions This study demonstrates how Semantic Web technologies can be applied in conjunction with clinical data stored in EHRs to accurately identify subjects with specific diseases and phenotypes, and identify genotype-phenotype associations. PMID:23244446

  5. Applications of Lexical Link Analysis Web Service for Large-Scale Automation, Validation, Discovery, Visualization, and Real-Time Program-Awareness

    DTIC Science & Technology

    2012-04-30

    individualistic , Western view of human behavior. That is, it assumes that people adhere to the individualistic principles of Western culture . What happens to...people with networks rich in structural holes that live/work in environments that adhere to other principles, such as those of a collectivistic ... culture ? •http://orgtheory.wordpress.com/2007/06/19/structural-holes-in-context/ Preferential Attachment (PA) [Barabási & Albert, 1999] • The most

  6. A Web Service Implementation for Large-Scale Automation, Visualization and Real-Time Program Awareness via Lexical Link Analysis

    DTIC Science & Technology

    2011-04-30

    internal constructs f l f t th h l i l li k l i (LLA)? 3 use u or managemen , roug ex ca n ana ys s LLA Methodology Can Help! Warfighters RDTE...information, including suggestions for reducing this burden, to Washington Headquarters Services, Directorate for Information Operations and Reports...categories of interest in various spreadsheets). This year, we started to develop LLA from a demonstration to an operational capability and facilitate a

  7. Modeling the Webgraph: How Far We Are

    NASA Astrophysics Data System (ADS)

    Donato, Debora; Laura, Luigi; Leonardi, Stefano; Millozzi, Stefano

    The following sections are included: * Introduction * Preliminaries * WebBase * In-degree and out-degree * PageRank * Bipartite cliques * Strongly connected components * Stochastic models of the webgraph * Models of the webgraph * A multi-layer model * Large scale simulation * Algorithmic techniques for generating and measuring webgraphs * Data representation and multifiles * Generating webgraphs * Traversal with two bits for each node * Semi-external breadth first search * Semi-external depth first search * Computation of the SCCs * Computation of the bow-tie regions * Disjoint bipartite cliques * PageRank * Summary and outlook

  8. Visualizing and communicating climate change using the ClimateWizard: decision support and education through web-based analysis and mapping

    NASA Astrophysics Data System (ADS)

    Girvetz, E. H.; Zganjar, C.; Raber, G. T.; Maurer, E. P.; Duffy, P.

    2009-12-01

    Virtually all fields of study and parts of society—from ecological science and nature conservation, to global development, multinational corporations, and government bodies—need to know how climate change has and may impact specific locations of interest. Our ability to respond to climate change depends on having convenient tools that make past and projected climate trends available to planners, managers, scientists and the general public, at scales ranging from global to local scales. Web-mapping applications provide an effective platform for communicating climate change impacts in specific geographic areas of interest to the public. Here, we present one such application, the ClimateWizard, that allows users to analyze, visualize and explore climate change maps for specific geographic areas of interest throughout the world (http://ClimateWizard.org). Built on Web 2.0 web-services (SOAP), Google Maps mash-up, and cloud computing technologies, the ClimateWizard analyzes large databases of climate information located on remote servers to create synthesized information and useful products tailored to geographic areas of interest (e.g. maps, graphs, tables, GIS layers). We demonstrate how the ClimateWizard can be used to assess projected changes to temperature and precipitation across all states in the contiguous United States and all countries of the world using statistically downscaled general circulation models from the CMIP3 dataset. We then go on to show how ClimateWizard can be used to analyze changes to other climate related variables, such as moisture stress and water production. Finally, we discuss how this tool can be adapted to develop a wide range of web-based tools that are targeted at informing specific audiences—from scientific research and natural resource management, to K-12 and higher education—about how climate change may affect different aspects of human and natural systems.

  9. The energetic contributions of aquatic primary producers to terrestrial food webs in a mid-size river system.

    PubMed

    Kautza, Adam; Mazeika, S; Sullivan, P

    2016-03-01

    Rivers are increasingly recognized as providing nutritional subsidies (i.e., energy and nutrients) to adjacent terrestrial food webs via depredation of aquatic organisms (e.g., emergent aquatic insects, crayfish, fish) by terrestrial consumers. However, because these prey organisms assimilate energy from both aquatic (e.g., benthic algae, phytoplankton, aquatic macrophytes) and terrestrial (e.g., riparian leaf detritus) primary producers, river subsidies to terrestrial consumers represent a combination of aquatically and terrestrially derived energy. To date, the explicit contribution of energy derived from aquatic primary producers to terrestrial consumers has not been fully explored yet might be expected to be quantitatively important to terrestrial food webs. At 12 reaches along a 185-km segment of the sixth-order Scioto River system (Ohio, USA), we quantified the relative contribution of energy derived from aquatic primary producers to a suite of terrestrial riparian consumers that integrate the adjacent landscape across multiple spatial scales through their foraging activities (tetragnathid spiders, rove beetles, adult coenagrionid damselflies, riparian swallows, and raccoons). We used naturally abundant stable isotopes (13C and 15N) of periphyton, phytoplankton, macrophytes, and terrestrial vegetation to evaluate the energetic contribution of aquatic primary producers to terrestrial food webs. Shoreline tetragnathid spiders were most reliant on aquatic primary producers (50%), followed by wider-ranging raccoons (48%), damselflies (44%), and riparian swallows (41%). Of the primary producers, phytoplankton (19%) provisioned the greatest nutritional contribution to terrestrial consumers (considered collectively), followed by periphyton (14%) and macrophytes (11%). Our findings provide empirical evidence that aquatic primary producers of large streams and rivers can be a critical nutritional resource for terrestrial food webs. We also show that aquatically derived nutrition contributes to both shoreline and broader-ranging terrestrial consumers and thus may be an important landscape-scale energetic linkage between rivers and upland habitats.

  10. The bacterial composition within the Sarracenia purpurea model system: local scale differences and the relationship with the other members of the food web.

    PubMed

    Gray, Sarah M; Akob, Denise M; Green, Stefan J; Kostka, Joel E

    2012-01-01

    The leaves of the carnivorous pitcher plant, Sarracenia purpurea, contain a microscopic aquatic food web that is considered a model system in ecological research. The species identity of the intermediate and top trophic level of this food web, as well the detritivore midge, are highly similar across the native geographic range of S. purpurea and, in some cases, appear to have co-evolved with the plant. However, until recently, the identity, geographic variation, and diversity of the bacteria in the bottom trophic level of this food web have remained largely unknown. This study investigated bacterial community composition inside the leaves of S. purpurea to address: 1) variation in bacterial communities at the beginning of succession at the local scale in different areas of the plant's native geographic range (southern and mid-regional sites) and 2) the impacts of bacterial consumers and other members of the aquatic food web (i.e., insects) on bacterial community structure. Communities from six leaves (one leaf per plant) from New York and Florida study sites were analyzed using 16S ribosomal RNA gene cloning. Each pitcher within each site had a distinct community; however, there was more overlap in bacterial composition within each site than when communities were compared across sites. In contrast, the identity of protozoans and metazoans in this community were similar in species identity both within a site and between the two sites, but abundances differed. Our results indicate that, at least during the beginning of succession, there is no strong selection for bacterial taxa and that there is no core group of bacteria required by the plant to start the decomposition of trapped insects. Co-evolution between the plant and bacteria appears to not have occurred as it has for other members of this community.

  11. The Bacterial Composition within the Sarracenia purpurea Model System: Local Scale Differences and the Relationship with the Other Members of the Food Web

    PubMed Central

    Gray, Sarah M.; Akob, Denise M.; Green, Stefan J.; Kostka, Joel E.

    2012-01-01

    The leaves of the carnivorous pitcher plant, Sarracenia purpurea, contain a microscopic aquatic food web that is considered a model system in ecological research. The species identity of the intermediate and top trophic level of this food web, as well the detritivore midge, are highly similar across the native geographic range of S. purpurea and, in some cases, appear to have co-evolved with the plant. However, until recently, the identity, geographic variation, and diversity of the bacteria in the bottom trophic level of this food web have remained largely unknown. This study investigated bacterial community composition inside the leaves of S. purpurea to address: 1) variation in bacterial communities at the beginning of succession at the local scale in different areas of the plant’s native geographic range (southern and mid-regional sites) and 2) the impacts of bacterial consumers and other members of the aquatic food web (i.e., insects) on bacterial community structure. Communities from six leaves (one leaf per plant) from New York and Florida study sites were analyzed using 16S ribosomal RNA gene cloning. Each pitcher within each site had a distinct community; however, there was more overlap in bacterial composition within each site than when communities were compared across sites. In contrast, the identity of protozoans and metazoans in this community were similar in species identity both within a site and between the two sites, but abundances differed. Our results indicate that, at least during the beginning of succession, there is no strong selection for bacterial taxa and that there is no core group of bacteria required by the plant to start the decomposition of trapped insects. Co-evolution between the plant and bacteria appears to not have occurred as it has for other members of this community. PMID:23227224

  12. ‘Sciencenet’—towards a global search and share engine for all scientific knowledge

    PubMed Central

    Lütjohann, Dominic S.; Shah, Asmi H.; Christen, Michael P.; Richter, Florian; Knese, Karsten; Liebel, Urban

    2011-01-01

    Summary: Modern biological experiments create vast amounts of data which are geographically distributed. These datasets consist of petabytes of raw data and billions of documents. Yet to the best of our knowledge, a search engine technology that searches and cross-links all different data types in life sciences does not exist. We have developed a prototype distributed scientific search engine technology, ‘Sciencenet’, which facilitates rapid searching over this large data space. By ‘bringing the search engine to the data’, we do not require server farms. This platform also allows users to contribute to the search index and publish their large-scale data to support e-Science. Furthermore, a community-driven method guarantees that only scientific content is crawled and presented. Our peer-to-peer approach is sufficiently scalable for the science web without performance or capacity tradeoff. Availability and Implementation: The free to use search portal web page and the downloadable client are accessible at: http://sciencenet.kit.edu. The web portal for index administration is implemented in ASP.NET, the ‘AskMe’ experiment publisher is written in Python 2.7, and the backend ‘YaCy’ search engine is based on Java 1.6. Contact: urban.liebel@kit.edu Supplementary Material: Detailed instructions and descriptions can be found on the project homepage: http://sciencenet.kit.edu. PMID:21493657

  13. Software Framework for Development of Web-GIS Systems for Analysis of Georeferenced Geophysical Data

    NASA Astrophysics Data System (ADS)

    Okladnikov, I.; Gordov, E. P.; Titov, A. G.

    2011-12-01

    Georeferenced datasets (meteorological databases, modeling and reanalysis results, remote sensing products, etc.) are currently actively used in numerous applications including modeling, interpretation and forecast of climatic and ecosystem changes for various spatial and temporal scales. Due to inherent heterogeneity of environmental datasets as well as their size which might constitute up to tens terabytes for a single dataset at present studies in the area of climate and environmental change require a special software support. A dedicated software framework for rapid development of providing such support information-computational systems based on Web-GIS technologies has been created. The software framework consists of 3 basic parts: computational kernel developed using ITTVIS Interactive Data Language (IDL), a set of PHP-controllers run within specialized web portal, and JavaScript class library for development of typical components of web mapping application graphical user interface (GUI) based on AJAX technology. Computational kernel comprise of number of modules for datasets access, mathematical and statistical data analysis and visualization of results. Specialized web-portal consists of web-server Apache, complying OGC standards Geoserver software which is used as a base for presenting cartographical information over the Web, and a set of PHP-controllers implementing web-mapping application logic and governing computational kernel. JavaScript library aiming at graphical user interface development is based on GeoExt library combining ExtJS Framework and OpenLayers software. Based on the software framework an information-computational system for complex analysis of large georeferenced data archives was developed. Structured environmental datasets available for processing now include two editions of NCEP/NCAR Reanalysis, JMA/CRIEPI JRA-25 Reanalysis, ECMWF ERA-40 Reanalysis, ECMWF ERA Interim Reanalysis, MRI/JMA APHRODITE's Water Resources Project Reanalysis, meteorological observational data for the territory of the former USSR for the 20th century, and others. Current version of the system is already involved into a scientific research process. Particularly, recently the system was successfully used for analysis of Siberia climate changes and its impact in the region. The software framework presented allows rapid development of Web-GIS systems for geophysical data analysis thus providing specialists involved into multidisciplinary research projects with reliable and practical instruments for complex analysis of climate and ecosystems changes on global and regional scales. This work is partially supported by RFBR grants #10-07-00547, #11-05-01190, and SB RAS projects 4.31.1.5, 4.31.2.7, 4, 8, 9, 50 and 66.

  14. FusionHub: A unified web platform for annotation and visualization of gene fusion events in human cancer.

    PubMed

    Panigrahi, Priyabrata; Jere, Abhay; Anamika, Krishanpal

    2018-01-01

    Gene fusion is a chromosomal rearrangement event which plays a significant role in cancer due to the oncogenic potential of the chimeric protein generated through fusions. At present many databases are available in public domain which provides detailed information about known gene fusion events and their functional role. Existing gene fusion detection tools, based on analysis of transcriptomics data usually report a large number of fusion genes as potential candidates, which could be either known or novel or false positives. Manual annotation of these putative genes is indeed time-consuming. We have developed a web platform FusionHub, which acts as integrated search engine interfacing various fusion gene databases and simplifies large scale annotation of fusion genes in a seamless way. In addition, FusionHub provides three ways of visualizing fusion events: circular view, domain architecture view and network view. Design of potential siRNA molecules through ensemble method is another utility integrated in FusionHub that could aid in siRNA-based targeted therapy. FusionHub is freely available at https://fusionhub.persistent.co.in.

  15. Morphological statistics of the cosmic web

    NASA Astrophysics Data System (ADS)

    Shandarin, Sergei F.

    2004-07-01

    We report the first systematic study of the supercluster-void network in the ΛCDM concordance cosmology treating voids and superclusters on an equal footing. We study the dark matter density field in real space smoothed with the Ls = 5 h[minus sign]1Mpc Gaussian window. Superclusters and voids are defined as individual members of over-dense and under-dense excursion sets respectively. We determine the morphological properties of the cosmic web at a large number of dark matter density levels by computing Minkowski functionals for every supercluster and void. At the adopted smoothing scale individual superclusters totally occupy no more than about 5% of the total volume and contain no more than 20% of mass if the largest supercluster is excluded. Likewise, individual voids totally occupy no more than 14% of volume and contain no more than 4% of mass if the largest void is excluded. The genus of individual superclusters can be ˜ 5 while the genus of individual voids reaches ˜ 55, implying significant amount of substructure in superclusters and especially in voids. Large voids are typically distinctly non-spherical.

  16. Facilitating Scientific Collaboration and Education with Easy Access Web Maps Using the AGAP Antarctic Geophysical Data

    NASA Astrophysics Data System (ADS)

    Abdi, A.

    2012-12-01

    Science and science education benefit from easy access to data yet often geophysical data sets are large, complex and difficult to share. The difficulty in sharing data and imagery easily inhibits both collaboration and the use of real data in educational applications. The dissemination of data products through web maps serves a very efficient and user-friendly method for students, the public and the science community to gain insights and understanding from data. Few research groups provide direct access to their data, let alone map-based visualizations. By building upon current GIS infrastructure with web mapping technologies, like ArcGIS Server, scientific groups, institutions and agencies can enhance the value of their GIS investments. The advantages of web maps to serve data products are many; existing web-mapping technology allows complex GIS analysis to be shared across the Internet, and can be easily scaled from a few users to millions. This poster highlights the features of an interactive web map developed at the Polar Geophysics Group at the Lamont-Doherty Earth Observatory of Columbia University that provides a visual representation of, and access to, data products that resulted from the group's recently concluded AGAP project (http://pgg.ldeo.columbia.edu). The AGAP project collected more than 120,000 line km of new aerogeophysical data using two Twin Otter aircrafts. Data included ice penetrating radar, magnetometer, gravimeter and laser altimeter measurements. The web map is based upon ArcGIS Viewer for Flex, which is a configurable client application built on the ArcGIS API for Flex that works seamlessly with ArcGIS Server 10. The application can serve a variety of raster and vector file formats through the Data Interoperability for Server, which eliminates data sharing barriers across numerous file formats. The ability of the application to serve large datasets is only hindered by the availability of appropriate hardware. ArcGIS is a proprietary product, but there are a few data portals in the earth sciences that have a map interface using open access products such as MapServer and OpenLayers, the most notable being the NASA IceBridge Data Portal. Indeed, with the widespread availability of web mapping technology, the scientific community should advance towards this direction when disseminating their data.

  17. ComplexContact: a web server for inter-protein contact prediction using deep learning.

    PubMed

    Zeng, Hong; Wang, Sheng; Zhou, Tianming; Zhao, Feifeng; Li, Xiufeng; Wu, Qing; Xu, Jinbo

    2018-05-22

    ComplexContact (http://raptorx2.uchicago.edu/ComplexContact/) is a web server for sequence-based interfacial residue-residue contact prediction of a putative protein complex. Interfacial residue-residue contacts are critical for understanding how proteins form complex and interact at residue level. When receiving a pair of protein sequences, ComplexContact first searches for their sequence homologs and builds two paired multiple sequence alignments (MSA), then it applies co-evolution analysis and a CASP-winning deep learning (DL) method to predict interfacial contacts from paired MSAs and visualizes the prediction as an image. The DL method was originally developed for intra-protein contact prediction and performed the best in CASP12. Our large-scale experimental test further shows that ComplexContact greatly outperforms pure co-evolution methods for inter-protein contact prediction, regardless of the species.

  18. Relational databases: a transparent framework for encouraging biology students to think informatically.

    PubMed

    Rice, Michael; Gladstone, William; Weir, Michael

    2004-01-01

    We discuss how relational databases constitute an ideal framework for representing and analyzing large-scale genomic data sets in biology. As a case study, we describe a Drosophila splice-site database that we recently developed at Wesleyan University for use in research and teaching. The database stores data about splice sites computed by a custom algorithm using Drosophila cDNA transcripts and genomic DNA and supports a set of procedures for analyzing splice-site sequence space. A generic Web interface permits the execution of the procedures with a variety of parameter settings and also supports custom structured query language queries. Moreover, new analytical procedures can be added by updating special metatables in the database without altering the Web interface. The database provides a powerful setting for students to develop informatic thinking skills.

  19. Distribution of guidance models for cardiac resynchronization therapy in the setting of multi-center clinical trials

    NASA Astrophysics Data System (ADS)

    Rajchl, Martin; Abhari, Kamyar; Stirrat, John; Ukwatta, Eranga; Cantor, Diego; Li, Feng P.; Peters, Terry M.; White, James A.

    2014-03-01

    Multi-center trials provide the unique ability to investigate novel techniques across a range of geographical sites with sufficient statistical power, the inclusion of multiple operators determining feasibility under a wider array of clinical environments and work-flows. For this purpose, we introduce a new means of distributing pre-procedural cardiac models for image-guided interventions across a large scale multi-center trial. In this method, a single core facility is responsible for image processing, employing a novel web-based interface for model visualization and distribution. The requirements for such an interface, being WebGL-based, are minimal and well within the realms of accessibility for participating centers. We then demonstrate the accuracy of our approach using a single-center pacemaker lead implantation trial with generic planning models.

  20. Design and application of a web-based real-time personal PM2.5 exposure monitoring system.

    PubMed

    Sun, Qinghua; Zhuang, Jia; Du, Yanjun; Xu, Dandan; Li, Tiantian

    2018-06-15

    Growing demand from public health research for conduct large-scale epidemiological studies to explore health effect of PM 2.5 was well-documented. To address this need, we design a web-based real-time personal PM 2.5 exposure monitoring system (RPPM2.5 system) which can help researcher to get big data of personal PM 2.5 exposure with low-cost, low labor requirement, and low operating technical requirements. RPPM2.5 system can provide relative accurate real-time personal exposure data for individuals, researches, and decision maker. And this system has been used in a survey of PM 2.5 personal exposure level conducted in 5 cities of China and has provided mass of valuable data for epidemiological research. Copyright © 2018 Elsevier B.V. All rights reserved.

  1. Computer problem-solving coaches for introductory physics: Design and usability studies

    NASA Astrophysics Data System (ADS)

    Ryan, Qing X.; Frodermann, Evan; Heller, Kenneth; Hsu, Leonardo; Mason, Andrew

    2016-06-01

    The combination of modern computing power, the interactivity of web applications, and the flexibility of object-oriented programming may finally be sufficient to create computer coaches that can help students develop metacognitive problem-solving skills, an important competence in our rapidly changing technological society. However, no matter how effective such coaches might be, they will only be useful if they are attractive to students. We describe the design and testing of a set of web-based computer programs that act as personal coaches to students while they practice solving problems from introductory physics. The coaches are designed to supplement regular human instruction, giving students access to effective forms of practice outside class. We present results from large-scale usability tests of the computer coaches and discuss their implications for future versions of the coaches.

  2. Relational Databases: A Transparent Framework for Encouraging Biology Students To Think Informatically

    PubMed Central

    2004-01-01

    We discuss how relational databases constitute an ideal framework for representing and analyzing large-scale genomic data sets in biology. As a case study, we describe a Drosophila splice-site database that we recently developed at Wesleyan University for use in research and teaching. The database stores data about splice sites computed by a custom algorithm using Drosophila cDNA transcripts and genomic DNA and supports a set of procedures for analyzing splice-site sequence space. A generic Web interface permits the execution of the procedures with a variety of parameter settings and also supports custom structured query language queries. Moreover, new analytical procedures can be added by updating special metatables in the database without altering the Web interface. The database provides a powerful setting for students to develop informatic thinking skills. PMID:15592597

  3. Feasibility and acceptability of a child sexual abuse prevention program for childcare professionals: comparison of a web-based and in-person training.

    PubMed

    Rheingold, Alyssa A; Zajac, Kristyn; Patton, Meghan

    2012-01-01

    Recent prevention research has established the efficacy of some child sexual abuse prevention programs targeting adults; however, less is known about the feasibility of implementing such programs. The current study examines the feasibility and acceptability of a child sexual abuse prevention program for child care professionals provided in two different formats: in person and Web based. The sample consisted of 188 child care professionals from a large-scale, multisite, randomized controlled trial. Findings indicate that both in-person and online training formats are feasible to implement and acceptable to professionals. When comparing formats, the in-person format was favored in terms of comfort level and likelihood of sharing information with others. These findings have significant implications for dissemination of child sexual abuse prevention programs for child care professionals.

  4. Prototyping of Remote Experiment and Exercise Systems for an Engineering Education based on World Wide Web

    NASA Astrophysics Data System (ADS)

    Iwatsuki, Masami; Kato, Yoriyuki; Yonekawa, Akira

    State-of-the-art Internet technologies allow us to provide advanced and interactive distance education services. However, we could not help but gather students for experiments and exercises in an education for engineering because large-scale equipments and expensive software are required. On the other hand, teleoperation systems with robot manipulator or vehicle via Internet have been developed in the field of robotics. By fusing these two techniques, we can realize remote experiment and exercise systems for the engineering education based on World Wide Web. This paper presents how to construct the remote environment that allows students to take courses on experiment and exercise independently of their locations. By using the proposed system, users can exercise and practice remotely about control of a manipulator and a robot vehicle and programming of image processing.

  5. Are There Differences Depending on the Device Used to Complete a Web Survey (PC or Smartphone) for Order-by-Click Questions?

    ERIC Educational Resources Information Center

    Revilla, Melanie

    2017-01-01

    The development of web surveys has been accompanied by the emergence of new scales, taking advantages of the visual and interactive features provided by the Internet like drop-down menus, sliders, drag-and-drop, or order-by-click scales. This article focuses on the order-by-click scales, studying the comparability of the data obtained for this…

  6. Open-Source Python Tools for Deploying Interactive GIS Dashboards for a Billion Datapoints on a Laptop

    NASA Astrophysics Data System (ADS)

    Steinberg, P. D.; Bednar, J. A.; Rudiger, P.; Stevens, J. L. R.; Ball, C. E.; Christensen, S. D.; Pothina, D.

    2017-12-01

    The rich variety of software libraries available in the Python scientific ecosystem provides a flexible and powerful alternative to traditional integrated GIS (geographic information system) programs. Each such library focuses on doing a certain set of general-purpose tasks well, and Python makes it relatively simple to glue the libraries together to solve a wide range of complex, open-ended problems in Earth science. However, choosing an appropriate set of libraries can be challenging, and it is difficult to predict how much "glue code" will be needed for any particular combination of libraries and tasks. Here we present a set of libraries that have been designed to work well together to build interactive analyses and visualizations of large geographic datasets, in standard web browsers. The resulting workflows run on ordinary laptops even for billions of data points, and easily scale up to larger compute clusters when available. The declarative top-level interface used in these libraries means that even complex, fully interactive applications can be built and deployed as web services using only a few dozen lines of code, making it simple to create and share custom interactive applications even for datasets too large for most traditional GIS systems. The libraries we will cover include GeoViews (HoloViews extended for geographic applications) for declaring visualizable/plottable objects, Bokeh for building visual web applications from GeoViews objects, Datashader for rendering arbitrarily large datasets faithfully as fixed-size images, Param for specifying user-modifiable parameters that model your domain, Xarray for computing with n-dimensional array data, Dask for flexibly dispatching computational tasks across processors, and Numba for compiling array-based Python code down to fast machine code. We will show how to use the resulting workflow with static datasets and with simulators such as GSSHA or AdH, allowing you to deploy flexible, high-performance web-based dashboards for your GIS data or simulations without needing major investments in code development or maintenance.

  7. The utility of web mining for epidemiological research: studying the association between parity and cancer risk.

    PubMed

    Tourassi, Georgia; Yoon, Hong-Jun; Xu, Songhua; Han, Xuesong

    2016-05-01

    The World Wide Web has emerged as a powerful data source for epidemiological studies related to infectious disease surveillance. However, its potential for cancer-related epidemiological discoveries is largely unexplored. Using advanced web crawling and tailored information extraction procedures, the authors automatically collected and analyzed the text content of 79 394 online obituary articles published between 1998 and 2014. The collected data included 51 911 cancer (27 330 breast; 9470 lung; 6496 pancreatic; 6342 ovarian; 2273 colon) and 27 483 non-cancer cases. With the derived information, the authors replicated a case-control study design to investigate the association between parity (i.e., childbearing) and cancer risk. Age-adjusted odds ratios (ORs) with 95% confidence intervals (CIs) were calculated for each cancer type and compared to those reported in large-scale epidemiological studies. Parity was found to be associated with a significantly reduced risk of breast cancer (OR = 0.78, 95% CI, 0.75-0.82), pancreatic cancer (OR = 0.78, 95% CI, 0.72-0.83), colon cancer (OR = 0.67, 95% CI, 0.60-0.74), and ovarian cancer (OR = 0.58, 95% CI, 0.54-0.62). Marginal association was found for lung cancer risk (OR = 0.87, 95% CI, 0.81-0.92). The linear trend between increased parity and reduced cancer risk was dramatically more pronounced for breast and ovarian cancer than the other cancers included in the analysis. This large web-mining study on parity and cancer risk produced findings very similar to those reported with traditional observational studies. It may be used as a promising strategy to generate study hypotheses for guiding and prioritizing future epidemiological studies. © The Author 2015. Published by Oxford University Press on behalf of the American Medical Informatics Association. All rights reserved. For Permissions, please email: journals.permissions@oup.com.

  8. The Multi-Scale Network Landscape of Collaboration.

    PubMed

    Bae, Arram; Park, Doheum; Ahn, Yong-Yeol; Park, Juyong

    2016-01-01

    Propelled by the increasing availability of large-scale high-quality data, advanced data modeling and analysis techniques are enabling many novel and significant scientific understanding of a wide range of complex social, natural, and technological systems. These developments also provide opportunities for studying cultural systems and phenomena--which can be said to refer to all products of human creativity and way of life. An important characteristic of a cultural product is that it does not exist in isolation from others, but forms an intricate web of connections on many levels. In the creation and dissemination of cultural products and artworks in particular, collaboration and communication of ideas play an essential role, which can be captured in the heterogeneous network of the creators and practitioners of art. In this paper we propose novel methods to analyze and uncover meaningful patterns from such a network using the network of western classical musicians constructed from a large-scale comprehensive Compact Disc recordings data. We characterize the complex patterns in the network landscape of collaboration between musicians across multiple scales ranging from the macroscopic to the mesoscopic and microscopic that represent the diversity of cultural styles and the individuality of the artists.

  9. The Multi-Scale Network Landscape of Collaboration

    PubMed Central

    Ahn, Yong-Yeol; Park, Juyong

    2016-01-01

    Propelled by the increasing availability of large-scale high-quality data, advanced data modeling and analysis techniques are enabling many novel and significant scientific understanding of a wide range of complex social, natural, and technological systems. These developments also provide opportunities for studying cultural systems and phenomena—which can be said to refer to all products of human creativity and way of life. An important characteristic of a cultural product is that it does not exist in isolation from others, but forms an intricate web of connections on many levels. In the creation and dissemination of cultural products and artworks in particular, collaboration and communication of ideas play an essential role, which can be captured in the heterogeneous network of the creators and practitioners of art. In this paper we propose novel methods to analyze and uncover meaningful patterns from such a network using the network of western classical musicians constructed from a large-scale comprehensive Compact Disc recordings data. We characterize the complex patterns in the network landscape of collaboration between musicians across multiple scales ranging from the macroscopic to the mesoscopic and microscopic that represent the diversity of cultural styles and the individuality of the artists. PMID:26990088

  10. Mining large heterogeneous data sets in drug discovery.

    PubMed

    Wild, David J

    2009-10-01

    Increasingly, effective drug discovery involves the searching and data mining of large volumes of information from many sources covering the domains of chemistry, biology and pharmacology amongst others. This has led to a proliferation of databases and data sources relevant to drug discovery. This paper provides a review of the publicly-available large-scale databases relevant to drug discovery, describes the kinds of data mining approaches that can be applied to them and discusses recent work in integrative data mining that looks for associations that pan multiple sources, including the use of Semantic Web techniques. The future of mining large data sets for drug discovery requires intelligent, semantic aggregation of information from all of the data sources described in this review, along with the application of advanced methods such as intelligent agents and inference engines in client applications.

  11. Utilizing Web 2.0 Technologies for Library Web Tutorials: An Examination of Instruction on Community College Libraries' Websites Serving Large Student Bodies

    ERIC Educational Resources Information Center

    Blummer, Barbara; Kenton, Jeffrey M.

    2015-01-01

    This is the second part of a series on Web 2.0 tools available from community college libraries' Websites. The first article appeared in an earlier volume of this journal and it illustrated the wide variety of Web 2.0 tools on community college libraries' Websites serving large student bodies (Blummer and Kenton 2014). The research found many of…

  12. GeneXplorer: an interactive web application for microarray data visualization and analysis.

    PubMed

    Rees, Christian A; Demeter, Janos; Matese, John C; Botstein, David; Sherlock, Gavin

    2004-10-01

    When publishing large-scale microarray datasets, it is of great value to create supplemental websites where either the full data, or selected subsets corresponding to figures within the paper, can be browsed. We set out to create a CGI application containing many of the features of some of the existing standalone software for the visualization of clustered microarray data. We present GeneXplorer, a web application for interactive microarray data visualization and analysis in a web environment. GeneXplorer allows users to browse a microarray dataset in an intuitive fashion. It provides simple access to microarray data over the Internet and uses only HTML and JavaScript to display graphic and annotation information. It provides radar and zoom views of the data, allows display of the nearest neighbors to a gene expression vector based on their Pearson correlations and provides the ability to search gene annotation fields. The software is released under the permissive MIT Open Source license, and the complete documentation and the entire source code are freely available for download from CPAN http://search.cpan.org/dist/Microarray-GeneXplorer/.

  13. Quantifying the web browser ecosystem

    PubMed Central

    Ferdman, Sela; Minkov, Einat; Gefen, David

    2017-01-01

    Contrary to the assumption that web browsers are designed to support the user, an examination of a 900,000 distinct PCs shows that web browsers comprise a complex ecosystem with millions of addons collaborating and competing with each other. It is possible for addons to “sneak in” through third party installations or to get “kicked out” by their competitors without user involvement. This study examines that ecosystem quantitatively by constructing a large-scale graph with nodes corresponding to users, addons, and words (terms) that describe addon functionality. Analyzing addon interactions at user level using the Personalized PageRank (PPR) random walk measure shows that the graph demonstrates ecological resilience. Adapting the PPR model to analyzing the browser ecosystem at the level of addon manufacturer, the study shows that some addon companies are in symbiosis and others clash with each other as shown by analyzing the behavior of 18 prominent addon manufacturers. Results may herald insight on how other evolving internet ecosystems may behave, and suggest a methodology for measuring this behavior. Specifically, applying such a methodology could transform the addon market. PMID:28644833

  14. An introduction to web scale discovery systems.

    PubMed

    Hoy, Matthew B

    2012-01-01

    This article explores the basic principles of web-scale discovery systems and how they are being implemented in libraries. "Web scale discovery" refers to a class of products that index a vast number of resources in a wide variety formats and allow users to search for content in the physical collection, print and electronic journal collections, and other resources from a single search box. Search results are displayed in a manner similar to Internet searches, in a relevance ranked list with links to online content. The advantages and disadvantages of these systems are discussed, and a list of popular discovery products is provided. A list of library websites with discovery systems currently implemented is also provided.

  15. The Effectiveness of Lecture-Integrated, Web-Supported Case Studies in Large Group Teaching

    ERIC Educational Resources Information Center

    Azzawi, May; Dawson, Maureen M.

    2007-01-01

    The effectiveness of lecture-integrated and web-supported case studies in supporting a large and academically diverse group of undergraduate students was evaluated in the present study. Case studies and resource (web)-based learning were incorporated as two complementary interactive learning strategies into the traditional curriculum. A truncated…

  16. The beaming of subhalo accretion

    NASA Astrophysics Data System (ADS)

    Libeskind, Noam I.

    2016-10-01

    We examine the infall pattern of subhaloes onto hosts in the context of the large-scale structure. We find that the infall pattern is essentially driven by the shear tensor of the ambient velocity field. Dark matter subhaloes are preferentially accreted along the principal axis of the shear tensor which corresponds to the direction of weakest collapse. We examine the dependence of this preferential infall on subhalo mass, host halo mass and redshift. Although strongest for the most massive hosts and the most massive subhaloes at high redshift, the preferential infall of subhaloes is effectively universal in the sense that its always aligned with the axis of weakest collapse of the velocity shear tensor. It is the same shear tensor that dictates the structure of the cosmic web and hence the shear field emerges as the key factor that governs the local anisotropic pattern of structure formation. Since the small (sub-Mpc) scale is strongly correlated with the mid-range (~ 10 Mpc) scale - a scale accessible by current surveys of peculiar velocities - it follows that findings presented here open a new window into the relation between the observed large scale structure unveiled by current surveys of peculiar velocities and the preferential infall direction of the Local Group. This may shed light on the unexpected alignments of dwarf galaxies seen in the Local Group.

  17. Cosmic Web of Galaxies in the COSMOS Field: Public Catalog and Different Quenching for Centrals and Satellites

    NASA Astrophysics Data System (ADS)

    Darvish, Behnam; Mobasher, Bahram; Martin, D. Christopher; Sobral, David; Scoville, Nick; Stroe, Andra; Hemmati, Shoubaneh; Kartaltepe, Jeyhan

    2017-03-01

    We use a mass complete (log(M/{M}⊙ ) ≥slant 9.6) sample of galaxies with accurate photometric redshifts in the COSMOS field to construct the density field and the cosmic web to z = 1.2. The comic web extraction relies on the density field Hessian matrix and breaks the density field into clusters, filaments, and the field. We provide the density field and cosmic web measures to the community. We show that at z ≲ 0.8, the median star formation rate (SFR) in the cosmic web gradually declines from the field to clusters and this decline is especially sharp for satellites (˜1 dex versus ˜0.5 dex for centrals). However, at z ≳ 0.8, the trend flattens out for the overall galaxy population and satellites. For star-forming (SF) galaxies only, the median SFR is constant at z ≳ 0.5 but declines by ˜0.3-0.4 dex from the field to clusters for satellites and centrals at z ≲ 0.5. We argue that for satellites, the main role of the cosmic web environment is to control their SF fraction, whereas for centrals, it is mainly to control their overall SFR at z ≲ 0.5 and to set their fraction at z ≳ 0.5. We suggest that most satellites experience a rapid quenching mechanism as they fall from the field into clusters through filaments, whereas centrals mostly undergo a slow environmental quenching at z ≲ 0.5 and a fast mechanism at higher redshifts. Our preliminary results highlight the importance of the large-scale cosmic web on galaxy evolution.

  18. Randomized evaluation of a web based interview process for urology resident selection.

    PubMed

    Shah, Satyan K; Arora, Sanjeev; Skipper, Betty; Kalishman, Summers; Timm, T Craig; Smith, Anthony Y

    2012-04-01

    We determined whether a web based interview process for resident selection could effectively replace the traditional on-site interview. For the 2010 to 2011 match cycle, applicants to the University of New Mexico urology residency program were randomized to participate in a web based interview process via Skype or a traditional on-site interview process. Both methods included interviews with the faculty, a tour of facilities and the opportunity to ask current residents any questions. To maintain fairness the applicants were then reinterviewed via the opposite process several weeks later. We assessed comparative effectiveness, cost, convenience and satisfaction using anonymous surveys largely scored on a 5-point Likert scale. Of 39 total participants (33 applicants and 6 faculty) 95% completed the surveys. The web based interview was less costly to applicants (mean $171 vs $364, p=0.05) and required less time away from school (10% missing 1 or more days vs 30%, p=0.04) compared to traditional on-site interview. However, applicants perceived the web based interview process as less effective than traditional on-site interview, with a mean 6-item summative effectiveness score of 21.3 vs 25.6 (p=0.003). Applicants and faculty favored continuing the web based interview process in the future as an adjunct to on-site interviews. Residency interviews can be successfully conducted via the Internet. The web based interview process reduced costs and improved convenience. The findings of this study support the use of videoconferencing as an adjunct to traditional interview methods rather than as a replacement. Copyright © 2012 American Urological Association Education and Research, Inc. Published by Elsevier Inc. All rights reserved.

  19. Bioavailability of Mercury to Riverine Food Webs as a Function of Flood-Event Inundation of Channel Boundary Sediments

    NASA Astrophysics Data System (ADS)

    Singer, M. B.; Pellachini, C.; Blum, J. D.; Marvin-DiPasquale, M. C.; Donovan, P. M.

    2013-12-01

    Bioavailability of sediment-adsorbed contaminants to food webs in river corridors is typically controlled by biological, chemical, and physical factors, but understanding of their respective influences is limited due to a dearth of landscape-scale investigations of these biogeochemical links. Studies that account for the dynamics and interactions of hydrology and sediment transport in affecting the reactivity of sediment-adsorbed heavy metals such as mercury (Hg) are particularly lacking. Sequences of flood events generate complex inundation histories with banks, terraces, and floodplains that have the potential to alter local redox conditions and thereby affect the oxidation of elemental Hg0 to inorganic Hg(II), and the microbial conversion of Hg(II) to methylmercury (MeHg), potentially increasing the risk of Hg uptake into aquatic food webs. However, the probability distributions of saturation/inundation frequency and duration are typically unknown for channel boundaries along sediment transport pathways, and landscape-scale characterizations of Hg reactivity are rare along contaminated rivers. This research provides the first links between the dynamics of physical processes and biochemical processing and uptake into food webs in fluvial systems beset by large-scale mining contamination. Here we present new research on Hg-contaminated legacy terraces and banks along the Yuba River anthropogenic fan, produced by 19th C. hydraulic gold mining in Northern California. To assess the changes in Hg(II) availability for methylation and MeHg bioavailability into the food web, we combine numerical modeling of streamflow with geochemical assays of total Hg and Hg reactivity to identify hot spots of toxicity within the river corridor as a function of cycles of wetting/drying. We employ a 3D hydraulic model to route historical streamflow hydrographs from major flood events through the Yuba and Feather Rivers into the Central Valley to assess the frequency and duration of saturation/inundation of channel boundary sediments. We compare these spatiotemporal modeling results to sediment total Hg and stannous chloride ';reducible' Hg(II) concentrations (the latter as a proxy for Hg(II) availability for methylation) along this ~70 km swath of river corridor. Finally, we evaluate these potential hot spots of Hg toxicity against MeHg concentrations in local aquatic biota at several trophic levels. The research will provide the basis for new models describing the evolution of toxic substances in river corridors and may prove helpful in explaining the contribution of Hg to food webs of the San Francisco Bay-Delta as an enduring legacy of California's 19th C. Gold Rush.

  20. NaviCom: a web application to create interactive molecular network portraits using multi-level omics data.

    PubMed

    Dorel, Mathurin; Viara, Eric; Barillot, Emmanuel; Zinovyev, Andrei; Kuperstein, Inna

    2017-01-01

    Human diseases such as cancer are routinely characterized by high-throughput molecular technologies, and multi-level omics data are accumulated in public databases at increasing rate. Retrieval and visualization of these data in the context of molecular network maps can provide insights into the pattern of regulation of molecular functions reflected by an omics profile. In order to make this task easy, we developed NaviCom, a Python package and web platform for visualization of multi-level omics data on top of biological network maps. NaviCom is bridging the gap between cBioPortal, the most used resource of large-scale cancer omics data and NaviCell, a data visualization web service that contains several molecular network map collections. NaviCom proposes several standardized modes of data display on top of molecular network maps, allowing addressing specific biological questions. We illustrate how users can easily create interactive network-based cancer molecular portraits via NaviCom web interface using the maps of Atlas of Cancer Signalling Network (ACSN) and other maps. Analysis of these molecular portraits can help in formulating a scientific hypothesis on the molecular mechanisms deregulated in the studied disease. NaviCom is available at https://navicom.curie.fr. © The Author(s) 2017. Published by Oxford University Press.

  1. Semantic Web repositories for genomics data using the eXframe platform.

    PubMed

    Merrill, Emily; Corlosquet, Stéphane; Ciccarese, Paolo; Clark, Tim; Das, Sudeshna

    2014-01-01

    With the advent of inexpensive assay technologies, there has been an unprecedented growth in genomics data as well as the number of databases in which it is stored. In these databases, sample annotation using ontologies and controlled vocabularies is becoming more common. However, the annotation is rarely available as Linked Data, in a machine-readable format, or for standardized queries using SPARQL. This makes large-scale reuse, or integration with other knowledge bases very difficult. To address this challenge, we have developed the second generation of our eXframe platform, a reusable framework for creating online repositories of genomics experiments. This second generation model now publishes Semantic Web data. To accomplish this, we created an experiment model that covers provenance, citations, external links, assays, biomaterials used in the experiment, and the data collected during the process. The elements of our model are mapped to classes and properties from various established biomedical ontologies. Resource Description Framework (RDF) data is automatically produced using these mappings and indexed in an RDF store with a built-in Sparql Protocol and RDF Query Language (SPARQL) endpoint. Using the open-source eXframe software, institutions and laboratories can create Semantic Web repositories of their experiments, integrate it with heterogeneous resources and make it interoperable with the vast Semantic Web of biomedical knowledge.

  2. Tension and Bending Testing of an Integral T-Cap for Stitched Composite Airframe Joints

    NASA Technical Reports Server (NTRS)

    Lovejoy, Andrew E.; Leone, Frank A., Jr.

    2016-01-01

    The Pultruded Rod Stitched Efficient Unitized Structure (PRSEUS) is a structural concept that was developed by The Boeing Company to address the complex structural design aspects associated with a pressurized hybrid wing body aircraft configuration. An important design feature required for assembly is the integrally stitched T-cap, which provides connectivity of the corner (orthogonal) joint between adjacent panels. A series of tests were conducted on T-cap test articles, with and without a rod stiffener penetrating the T-cap web, under tension (pull-off) and bending loads. Three designs were tested, including the baseline design used in large-scale test articles. The baseline had only the manufacturing stitch row adjacent to the fillet at the base of the T-cap web. Two new designs added stitching rows to the T-cap web at either 0.5- or 1.0-inch spacing along the height of the web. Testing was conducted at NASA Langley Research Center to determine the behavior of the T-cap region resulting from the applied loading. Results show that stitching arrests the initial delamination failures so that the maximum strength capability exceeds the load at which the initial delaminations develop. However, it was seen that the added web stitching had very little effect on the initial delamination failure load, but actually decreased the initial delamination failure load for tension loading of test articles without a stiffener passing through the web. Additionally, the added web stitching only increased the maximum load capability by between 1% and 12.5%. The presence of the stiffener, however, did increase the initial and maximum loads for both tension and bending loading as compared to the stringerless baseline design. Based on the results of the few samples tested, the additional stitching in the T-cap web showed little advantage over the baseline design in terms of structural failure at the T-cap web/skin junction for the current test articles.

  3. Semantic similarity between ontologies at different scales

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Zhang, Qingpeng; Haglin, David J.

    In the past decade, existing and new knowledge and datasets has been encoded in different ontologies for semantic web and biomedical research. The size of ontologies is often very large in terms of number of concepts and relationships, which makes the analysis of ontologies and the represented knowledge graph computational and time consuming. As the ontologies of various semantic web and biomedical applications usually show explicit hierarchical structures, it is interesting to explore the trade-offs between ontological scales and preservation/precision of results when we analyze ontologies. This paper presents the first effort of examining the capability of this idea viamore » studying the relationship between scaling biomedical ontologies at different levels and the semantic similarity values. We evaluate the semantic similarity between three Gene Ontology slims (Plant, Yeast, and Candida, among which the latter two belong to the same kingdom—Fungi) using four popular measures commonly applied to biomedical ontologies (Resnik, Lin, Jiang-Conrath, and SimRel). The results of this study demonstrate that with proper selection of scaling levels and similarity measures, we can significantly reduce the size of ontologies without losing substantial detail. In particular, the performance of Jiang-Conrath and Lin are more reliable and stable than that of the other two in this experiment, as proven by (a) consistently showing that Yeast and Candida are more similar (as compared to Plant) at different scales, and (b) small deviations of the similarity values after excluding a majority of nodes from several lower scales. This study provides a deeper understanding of the application of semantic similarity to biomedical ontologies, and shed light on how to choose appropriate semantic similarity measures for biomedical engineering.« less

  4. Properties of galaxies reproduced by a hydrodynamic simulation

    NASA Astrophysics Data System (ADS)

    Vogelsberger, M.; Genel, S.; Springel, V.; Torrey, P.; Sijacki, D.; Xu, D.; Snyder, G.; Bird, S.; Nelson, D.; Hernquist, L.

    2014-05-01

    Previous simulations of the growth of cosmic structures have broadly reproduced the `cosmic web' of galaxies that we see in the Universe, but failed to create a mixed population of elliptical and spiral galaxies, because of numerical inaccuracies and incomplete physical models. Moreover, they were unable to track the small-scale evolution of gas and stars to the present epoch within a representative portion of the Universe. Here we report a simulation that starts 12 million years after the Big Bang, and traces 13 billion years of cosmic evolution with 12 billion resolution elements in a cube of 106.5 megaparsecs a side. It yields a reasonable population of ellipticals and spirals, reproduces the observed distribution of galaxies in clusters and characteristics of hydrogen on large scales, and at the same time matches the `metal' and hydrogen content of galaxies on small scales.

  5. ODMedit: uniform semantic annotation for data integration in medicine based on a public metadata repository.

    PubMed

    Dugas, Martin; Meidt, Alexandra; Neuhaus, Philipp; Storck, Michael; Varghese, Julian

    2016-06-01

    The volume and complexity of patient data - especially in personalised medicine - is steadily increasing, both regarding clinical data and genomic profiles: Typically more than 1,000 items (e.g., laboratory values, vital signs, diagnostic tests etc.) are collected per patient in clinical trials. In oncology hundreds of mutations can potentially be detected for each patient by genomic profiling. Therefore data integration from multiple sources constitutes a key challenge for medical research and healthcare. Semantic annotation of data elements can facilitate to identify matching data elements in different sources and thereby supports data integration. Millions of different annotations are required due to the semantic richness of patient data. These annotations should be uniform, i.e., two matching data elements shall contain the same annotations. However, large terminologies like SNOMED CT or UMLS don't provide uniform coding. It is proposed to develop semantic annotations of medical data elements based on a large-scale public metadata repository. To achieve uniform codes, semantic annotations shall be re-used if a matching data element is available in the metadata repository. A web-based tool called ODMedit ( https://odmeditor.uni-muenster.de/ ) was developed to create data models with uniform semantic annotations. It contains ~800,000 terms with semantic annotations which were derived from ~5,800 models from the portal of medical data models (MDM). The tool was successfully applied to manually annotate 22 forms with 292 data items from CDISC and to update 1,495 data models of the MDM portal. Uniform manual semantic annotation of data models is feasible in principle, but requires a large-scale collaborative effort due to the semantic richness of patient data. A web-based tool for these annotations is available, which is linked to a public metadata repository.

  6. Evolution of the cosmic web

    NASA Astrophysics Data System (ADS)

    Cautun, Marius; van de Weygaert, Rien; Jones, Bernard J. T.; Frenk, Carlos S.

    2014-07-01

    The cosmic web is the largest scale manifestation of the anisotropic gravitational collapse of matter. It represents the transitional stage between linear and non-linear structures and contains easily accessible information about the early phases of structure formation processes. Here we investigate the characteristics and the time evolution of morphological components. Our analysis involves the application of the NEXUS Multiscale Morphology Filter technique, predominantly its NEXUS+ version, to high resolution and large volume cosmological simulations. We quantify the cosmic web components in terms of their mass and volume content, their density distribution and halo populations. We employ new analysis techniques to determine the spatial extent of filaments and sheets, like their total length and local width. This analysis identifies clusters and filaments as the most prominent components of the web. In contrast, while voids and sheets take most of the volume, they correspond to underdense environments and are devoid of group-sized and more massive haloes. At early times the cosmos is dominated by tenuous filaments and sheets, which, during subsequent evolution, merge together, such that the present-day web is dominated by fewer, but much more massive, structures. The analysis of the mass transport between environments clearly shows how matter flows from voids into walls, and then via filaments into cluster regions, which form the nodes of the cosmic web. We also study the properties of individual filamentary branches, to find long, almost straight, filaments extending to distances larger than 100 h-1 Mpc. These constitute the bridges between massive clusters, which seem to form along approximatively straight lines.

  7. Experience versus talent shapes the structure of the Web

    PubMed Central

    Kong, Joseph S.; Sarshar, Nima; Roychowdhury, Vwani P.

    2008-01-01

    We use sequential large-scale crawl data to empirically investigate and validate the dynamics that underlie the evolution of the structure of the web. We find that the overall structure of the web is defined by an intricate interplay between experience or entitlement of the pages (as measured by the number of inbound hyperlinks a page already has), inherent talent or fitness of the pages (as measured by the likelihood that someone visiting the page would give a hyperlink to it), and the continual high rates of birth and death of pages on the web. We find that the web is conservative in judging talent and the overall fitness distribution is exponential, showing low variability. The small variance in talent, however, is enough to lead to experience distributions with high variance: The preferential attachment mechanism amplifies these small biases and leads to heavy-tailed power-law (PL) inbound degree distributions over all pages, as well as over pages that are of the same age. The balancing act between experience and talent on the web allows newly introduced pages with novel and interesting content to grow quickly and surpass older pages. In this regard, it is much like what we observe in high-mobility and meritocratic societies: People with entitlement continue to have access to the best resources, but there is just enough screening for fitness that allows for talented winners to emerge and join the ranks of the leaders. Finally, we show that the fitness estimates have potential practical applications in ranking query results. PMID:18779560

  8. Spatial and temporal variance in fatty acid and stable isotope signatures across trophic levels in large river systems

    USGS Publications Warehouse

    Fritts, Andrea; Knights, Brent C.; Lafrancois, Toben D.; Bartsch, Lynn; Vallazza, Jon; Bartsch, Michelle; Richardson, William B.; Karns, Byron N.; Bailey, Sean; Kreiling, Rebecca

    2018-01-01

    Fatty acid and stable isotope signatures allow researchers to better understand food webs, food sources, and trophic relationships. Research in marine and lentic systems has indicated that the variance of these biomarkers can exhibit substantial differences across spatial and temporal scales, but this type of analysis has not been completed for large river systems. Our objectives were to evaluate variance structures for fatty acids and stable isotopes (i.e. δ13C and δ15N) of seston, threeridge mussels, hydropsychid caddisflies, gizzard shad, and bluegill across spatial scales (10s-100s km) in large rivers of the Upper Mississippi River Basin, USA that were sampled annually for two years, and to evaluate the implications of this variance on the design and interpretation of trophic studies. The highest variance for both isotopes was present at the largest spatial scale for all taxa (except seston δ15N) indicating that these isotopic signatures are responding to factors at a larger geographic level rather than being influenced by local-scale alterations. Conversely, the highest variance for fatty acids was present at the smallest spatial scale (i.e. among individuals) for all taxa except caddisflies, indicating that the physiological and metabolic processes that influence fatty acid profiles can differ substantially between individuals at a given site. Our results highlight the need to consider the spatial partitioning of variance during sample design and analysis, as some taxa may not be suitable to assess ecological questions at larger spatial scales.

  9. HammerCloud: A Stress Testing System for Distributed Analysis

    NASA Astrophysics Data System (ADS)

    van der Ster, Daniel C.; Elmsheuser, Johannes; Úbeda García, Mario; Paladin, Massimo

    2011-12-01

    Distributed analysis of LHC data is an I/O-intensive activity which places large demands on the internal network, storage, and local disks at remote computing facilities. Commissioning and maintaining a site to provide an efficient distributed analysis service is therefore a challenge which can be aided by tools to help evaluate a variety of infrastructure designs and configurations. HammerCloud is one such tool; it is a stress testing service which is used by central operations teams, regional coordinators, and local site admins to (a) submit arbitrary number of analysis jobs to a number of sites, (b) maintain at a steady-state a predefined number of jobs running at the sites under test, (c) produce web-based reports summarizing the efficiency and performance of the sites under test, and (d) present a web-interface for historical test results to both evaluate progress and compare sites. HammerCloud was built around the distributed analysis framework Ganga, exploiting its API for grid job management. HammerCloud has been employed by the ATLAS experiment for continuous testing of many sites worldwide, and also during large scale computing challenges such as STEP'09 and UAT'09, where the scale of the tests exceeded 10,000 concurrently running and 1,000,000 total jobs over multi-day periods. In addition, HammerCloud is being adopted by the CMS experiment; the plugin structure of HammerCloud allows the execution of CMS jobs using their official tool (CRAB).

  10. Postmarket Drug Surveillance Without Trial Costs: Discovery of Adverse Drug Reactions Through Large-Scale Analysis of Web Search Queries

    PubMed Central

    Gabrilovich, Evgeniy

    2013-01-01

    Background Postmarket drug safety surveillance largely depends on spontaneous reports by patients and health care providers; hence, less common adverse drug reactions—especially those caused by long-term exposure, multidrug treatments, or those specific to special populations—often elude discovery. Objective Here we propose a low cost, fully automated method for continuous monitoring of adverse drug reactions in single drugs and in combinations thereof, and demonstrate the discovery of heretofore-unknown ones. Methods We used aggregated search data of large populations of Internet users to extract information related to drugs and adverse reactions to them, and correlated these data over time. We further extended our method to identify adverse reactions to combinations of drugs. Results We validated our method by showing high correlations of our findings with known adverse drug reactions (ADRs). However, although acute early-onset drug reactions are more likely to be reported to regulatory agencies, we show that less acute later-onset ones are better captured in Web search queries. Conclusions Our method is advantageous in identifying previously unknown adverse drug reactions. These ADRs should be considered as candidates for further scrutiny by medical regulatory authorities, for example, through phase 4 trials. PMID:23778053

  11. GENESIS SciFlo: Choreographing Interoperable Web Services on the Grid using a Semantically-Enabled Dataflow Execution Environment

    NASA Astrophysics Data System (ADS)

    Wilson, B. D.; Manipon, G.; Xing, Z.

    2007-12-01

    The General Earth Science Investigation Suite (GENESIS) project is a NASA-sponsored partnership between the Jet Propulsion Laboratory, academia, and NASA data centers to develop a new suite of Web Services tools to facilitate multi-sensor investigations in Earth System Science. The goal of GENESIS is to enable large-scale, multi-instrument atmospheric science using combined datasets from the AIRS, MODIS, MISR, and GPS sensors. Investigations include cross-comparison of spaceborne climate sensors, cloud spectral analysis, study of upper troposphere-stratosphere water transport, study of the aerosol indirect cloud effect, and global climate model validation. The challenges are to bring together very large datasets, reformat and understand the individual instrument retrievals, co-register or re-grid the retrieved physical parameters, perform computationally-intensive data fusion and data mining operations, and accumulate complex statistics over months to years of data. To meet these challenges, we have developed a Grid computing and dataflow framework, named SciFlo, in which we are deploying a set of versatile and reusable operators for data access, subsetting, registration, mining, fusion, compression, and advanced statistical analysis. SciFlo leverages remote Web Services, called via Simple Object Access Protocol (SOAP) or REST (one-line) URLs, and the Grid Computing standards (WS-* & Globus Alliance toolkits), and enables scientists to do multi- instrument Earth Science by assembling reusable Web Services and native executables into a distributed computing flow (tree of operators). The SciFlo client & server engines optimize the execution of such distributed data flows and allow the user to transparently find and use datasets and operators without worrying about the actual location of the Grid resources. In particular, SciFlo exploits the wealth of datasets accessible by OpenGIS Consortium (OGC) Web Mapping Servers & Web Coverage Servers (WMS/WCS), and by Open Data Access Protocol (OpenDAP) servers. SciFlo also publishes its own SOAP services for space/time query and subsetting of Earth Science datasets, and automated access to large datasets via lists of (FTP, HTTP, or DAP) URLs which point to on-line HDF or netCDF files. Typical distributed workflows obtain datasets by calling standard WMS/WCS servers or discovering and fetching data granules from ftp sites; invoke remote analysis operators available as SOAP services (interface described by a WSDL document); and merge results into binary containers (netCDF or HDF files) for further analysis using local executable operators. Naming conventions (HDFEOS and CF-1.0 for netCDF) are exploited to automatically understand and read on-line datasets. More interoperable conventions, and broader adoption of existing converntions, are vital if we are to "scale up" automated choreography of Web Services beyond toy applications. Recently, the ESIP Federation sponsored a collaborative activity in which several ESIP members developed some collaborative science scenarios for atmospheric and aerosol science, and then choreographed services from multiple groups into demonstration workflows using the SciFlo engine and a Business Process Execution Language (BPEL) workflow engine. We will discuss the lessons learned from this activity, the need for standardized interfaces (like WMS/WCS), the difficulty in agreeing on even simple XML formats and interfaces, the benefits of doing collaborative science analysis at the "touch of a button" once services are connected, and further collaborations that are being pursued.

  12. Evaluation of the Feasibility of Screening Patients for Early Signs of Lung Carcinoma in Web Search Logs.

    PubMed

    White, Ryen W; Horvitz, Eric

    2017-03-01

    A statistical model that predicts the appearance of strong evidence of a lung carcinoma diagnosis via analysis of large-scale anonymized logs of web search queries from millions of people across the United States. To evaluate the feasibility of screening patients at risk of lung carcinoma via analysis of signals from online search activity. We identified people who issue special queries that provide strong evidence of a recent diagnosis of lung carcinoma. We then considered patterns of symptoms expressed as searches about concerning symptoms over several months prior to the appearance of the landmark web queries. We built statistical classifiers that predict the future appearance of landmark queries based on the search log signals. This was a retrospective log analysis of the online activity of millions of web searchers seeking health-related information online. Of web searchers who queried for symptoms related to lung carcinoma, some (n = 5443 of 4 813 985) later issued queries that provide strong evidence of recent clinical diagnosis of lung carcinoma and are regarded as positive cases in our analysis. Additional evidence on the reliability of these queries as representing clinical diagnoses is based on the significant increase in follow-on searches for treatments and medications for these searchers and on the correlation between lung carcinoma incidence rates and our log-based statistics. The remaining symptom searchers (n = 4 808 542) are regarded as negative cases. Performance of the statistical model for early detection from online search behavior, for different lead times, different sets of signals, and different cohorts of searchers stratified by potential risk. The statistical classifier predicting the future appearance of landmark web queries based on search log signals identified searchers who later input queries consistent with a lung carcinoma diagnosis, with a true-positive rate ranging from 3% to 57% for false-positive rates ranging from 0.00001 to 0.001, respectively. The methods can be used to identify people at highest risk up to a year in advance of the inferred diagnosis time. The 5 factors associated with the highest relative risk (RR) were evidence of family history (RR = 7.548; 95% CI, 3.937-14.470), age (RR = 3.558; 95% CI, 3.357-3.772), radon (RR = 2.529; 95% CI, 1.137-5.624), primary location (RR = 2.463; 95% CI, 1.364-4.446), and occupation (RR = 1.969; 95% CI, 1.143-3.391). Evidence of smoking (RR = 1.646; 95% CI, 1.032-2.260) was important but not top-ranked, which was due to the difficulty of identifying smoking history from search terms. Pattern recognition based on data drawn from large-scale web search queries holds opportunity for identifying risk factors and frames new directions with early detection of lung carcinoma.

  13. Towards building high performance medical image management system for clinical trials

    NASA Astrophysics Data System (ADS)

    Wang, Fusheng; Lee, Rubao; Zhang, Xiaodong; Saltz, Joel

    2011-03-01

    Medical image based biomarkers are being established for therapeutic cancer clinical trials, where image assessment is among the essential tasks. Large scale image assessment is often performed by a large group of experts by retrieving images from a centralized image repository to workstations to markup and annotate images. In such environment, it is critical to provide a high performance image management system that supports efficient concurrent image retrievals in a distributed environment. There are several major challenges: high throughput of large scale image data over the Internet from the server for multiple concurrent client users, efficient communication protocols for transporting data, and effective management of versioning of data for audit trails. We study the major bottlenecks for such a system, propose and evaluate a solution by using a hybrid image storage with solid state drives and hard disk drives, RESTfulWeb Services based protocols for exchanging image data, and a database based versioning scheme for efficient archive of image revision history. Our experiments show promising results of our methods, and our work provides a guideline for building enterprise level high performance medical image management systems.

  14. Spatial variations in food web structures with alternative stable states: evidence from stable isotope analysis in a large eutrophic lake

    NASA Astrophysics Data System (ADS)

    Li, Yunkai; Zhang, Yuying; Xu, Jun; Zhang, Shuo

    2018-03-01

    Food web structures are well known to vary widely among ecosystems. Moreover, many food web studies of lakes have generally attempted to characterize the overall food web structure and have largely ignored internal spatial and environmental variations. In this study, we hypothesize that there is a high degree of spatial heterogeneity within an ecosystem and such heterogeneity may lead to strong variations in environmental conditions and resource availability, in turn resulting in different trophic pathways. Stable carbon and nitrogen isotopes were employed for the whole food web to describe the structure of the food web in different sub-basins within Taihu Lake. This lake is a large eutrophic freshwater lake that has been intensively managed and highly influenced by human activities for more than 50 years. The results show significant isotopic differences between basins with different environmental characteristics. Such differences likely result from isotopic baseline differences combining with a shift in food web structure. Both are related to local spatial heterogeneity in nutrient loading in waters. Such variation should be explicitly considered in future food web studies and ecosystem-based management in this lake ecosystem.

  15. How does abundance scale with body size in coupled size-structured food webs?

    PubMed

    Blanchard, Julia L; Jennings, Simon; Law, Richard; Castle, Matthew D; McCloghrie, Paul; Rochet, Marie-Joëlle; Benoît, Eric

    2009-01-01

    1. Widely observed macro-ecological patterns in log abundance vs. log body mass of organisms can be explained by simple scaling theory based on food (energy) availability across a spectrum of body sizes. The theory predicts that when food availability falls with body size (as in most aquatic food webs where larger predators eat smaller prey), the scaling between log N vs. log m is steeper than when organisms of different sizes compete for a shared unstructured resource (e.g. autotrophs, herbivores and detritivores; hereafter dubbed 'detritivores'). 2. In real communities, the mix of feeding characteristics gives rise to complex food webs. Such complexities make empirical tests of scaling predictions prone to error if: (i) the data are not disaggregated in accordance with the assumptions of the theory being tested, or (ii) the theory does not account for all of the trophic interactions within and across the communities sampled. 3. We disaggregated whole community data collected in the North Sea into predator and detritivore components and report slopes of log abundance vs. log body mass relationships. Observed slopes for fish and epifaunal predator communities (-1.2 to -2.25) were significantly steeper than those for infaunal detritivore communities (-0.56 to -0.87). 4. We present a model describing the dynamics of coupled size spectra, to explain how coupling of predator and detritivore communities affects the scaling of log N vs. log m. The model captures the trophic interactions and recycling of material that occur in many aquatic ecosystems. 5. Our simulations demonstrate that the biological processes underlying growth and mortality in the two distinct size spectra lead to patterns consistent with data. Slopes of log N vs. log m were steeper and growth rates faster for predators compared to detritivores. Size spectra were truncated when primary production was too low for predators and when detritivores experienced predation pressure. 6. The approach also allows us to assess the effects of external sources of mortality (e.g. harvesting). Removal of large predators resulted in steeper predator spectra and increases in their prey (small fish and detritivores). The model predictions are remarkably consistent with observed patterns of exploited ecosystems.

  16. The initial development of the WebMedQual scale: domain assessment of the construct of quality of health web sites.

    PubMed

    Provost, Mélanie; Koompalum, Dayin; Dong, Diane; Martin, Bradley C

    2006-01-01

    To develop a comprehensive instrument assessing quality of health-related web sites. Phase I consisted of a literature review to identify constructs thought to indicate web site quality and to identify items. During content analysis, duplicate items were eliminated and items that were not clear, meaningful, or measurable were reworded or removed. Some items were generated by the authors. Phase II: a panel consisting of six healthcare and MIS reviewers was convened to assess each item for its relevance and importance to the construct and to assess item clarity and measurement feasibility. Three hundred and eighty-four items were generated from 26 sources. The initial content analysis reduced the scale to 104 items. Four of the six expert reviewers responded; high concordance on the relevance, importance and measurement feasibility of each item was observed: 3 out of 4, or all raters agreed on 76-85% of items. Based on the panel ratings, 9 items were removed, 3 added, and 10 revised. The WebMedQual consists of 8 categories, 8 sub-categories, 95 items and 3 supplemental items to assess web site quality. The constructs are: content (19 items), authority of source (18 items), design (19 items), accessibility and availability (6 items), links (4 items), user support (9 items), confidentiality and privacy (17 items), e-commerce (6 items). The "WebMedQual" represents a first step toward a comprehensive and standard quality assessment of health web sites. This scale will allow relatively easy assessment of quality with possible numeric scoring.

  17. Robust multi-site MR data processing: iterative optimization of bias correction, tissue classification, and registration.

    PubMed

    Young Kim, Eun; Johnson, Hans J

    2013-01-01

    A robust multi-modal tool, for automated registration, bias correction, and tissue classification, has been implemented for large-scale heterogeneous multi-site longitudinal MR data analysis. This work focused on improving the an iterative optimization framework between bias-correction, registration, and tissue classification inspired from previous work. The primary contributions are robustness improvements from incorporation of following four elements: (1) utilize multi-modal and repeated scans, (2) incorporate high-deformable registration, (3) use extended set of tissue definitions, and (4) use of multi-modal aware intensity-context priors. The benefits of these enhancements were investigated by a series of experiments with both simulated brain data set (BrainWeb) and by applying to highly-heterogeneous data from a 32 site imaging study with quality assessments through the expert visual inspection. The implementation of this tool is tailored for, but not limited to, large-scale data processing with great data variation with a flexible interface. In this paper, we describe enhancements to a joint registration, bias correction, and the tissue classification, that improve the generalizability and robustness for processing multi-modal longitudinal MR scans collected at multi-sites. The tool was evaluated by using both simulated and simulated and human subject MRI images. With these enhancements, the results showed improved robustness for large-scale heterogeneous MRI processing.

  18. Revelations from the Literature: How Web-Scale Discovery Has Already Changed Us

    ERIC Educational Resources Information Center

    Richardson, Hillary A. H.

    2013-01-01

    For nearly a decade now, librarians have discussed and deliberated ways, for the sake of convenience, to integrate internet-like searching into their own catalogs to mimic what academic library patrons have been using outside the library. Discovery tools and web-scale discovery services (WSDS) attempt to provide users with a similar one-stop shop…

  19. Essential Skills and Knowledge for Troubleshooting E-Resources Access Issues in a Web-Scale Discovery Environment

    ERIC Educational Resources Information Center

    Carter, Sunshine; Traill, Stacie

    2017-01-01

    Electronic resource access troubleshooting is familiar work in most libraries. The added complexity introduced when a library implements a web-scale discovery service, however, creates a strong need for well-organized, rigorous training to enable troubleshooting staff to provide the best service possible. This article outlines strategies, tools,…

  20. A Web-Based Educational Setting Supporting Individualized Learning, Collaborative Learning and Assessment

    ERIC Educational Resources Information Center

    Gogoulou, Agoritsa; Gouli, Evangelia; Grigoriadou, Maria; Samarakou, Maria; Chinou, Dionisia

    2007-01-01

    In this paper, we present a web-based educational setting, referred to as SCALE (Supporting Collaboration and Adaptation in a Learning Environment), which aims to serve learning and assessment. SCALE enables learners to (i) work on individual and collaborative activities proposed by the environment with respect to learners' knowledge level, (ii)…

  1. Proposition and Organization of an Adaptive Learning Domain Based on Fusion from the Web

    ERIC Educational Resources Information Center

    Chaoui, Mohammed; Laskri, Mohamed Tayeb

    2013-01-01

    The Web allows self-navigated education through interaction with large amounts of Web resources. While enjoying the flexibility of Web tools, authors may suffer from research and filtering Web resources, when they face various resources formats and complex structures. An adaptation of extracted Web resources must be assured by authors, to give…

  2. A Web service-based architecture for real-time hydrologic sensor networks

    NASA Astrophysics Data System (ADS)

    Wong, B. P.; Zhao, Y.; Kerkez, B.

    2014-12-01

    Recent advances in web services and cloud computing provide new means by which to process and respond to real-time data. This is particularly true of platforms built for the Internet of Things (IoT). These enterprise-scale platforms have been designed to exploit the IP-connectivity of sensors and actuators, providing a robust means by which to route real-time data feeds and respond to events of interest. While powerful and scalable, these platforms have yet to be adopted by the hydrologic community, where the value of real-time data impacts both scientists and decision makers. We discuss the use of one such IoT platform for the purpose of large-scale hydrologic measurements, showing how rapid deployment and ease-of-use allows scientists to focus on their experiment rather than software development. The platform is hardware agnostic, requiring only IP-connectivity of field devices to capture, store, process, and visualize data in real-time. We demonstrate the benefits of real-time data through a real-world use case by showing how our architecture enables the remote control of sensor nodes, thereby permitting the nodes to adaptively change sampling strategies to capture major hydrologic events of interest.

  3. The case for electron re-acceleration at galaxy cluster shocks

    NASA Astrophysics Data System (ADS)

    van Weeren, Reinout J.; Andrade-Santos, Felipe; Dawson, William A.; Golovich, Nathan; Lal, Dharam V.; Kang, Hyesung; Ryu, Dongsu; Brìggen, Marcus; Ogrean, Georgiana A.; Forman, William R.; Jones, Christine; Placco, Vinicius M.; Santucci, Rafael M.; Wittman, David; Jee, M. James; Kraft, Ralph P.; Sobral, David; Stroe, Andra; Fogarty, Kevin

    2017-01-01

    On the largest scales, the Universe consists of voids and filaments making up the cosmic web. Galaxy clusters are located at the knots in this web, at the intersection of filaments. Clusters grow through accretion from these large-scale filaments and by mergers with other clusters and groups. In a growing number of galaxy clusters, elongated Mpc-sized radio sources have been found1,2 . Also known as radio relics, these regions of diffuse radio emission are thought to trace relativistic electrons in the intracluster plasma accelerated by low-Mach-number shocks generated by cluster-cluster merger events 3 . A long-standing problem is how low-Mach-number shocks can accelerate electrons so efficiently to explain the observed radio relics. Here, we report the discovery of a direct connection between a radio relic and a radio galaxy in the merging galaxy cluster Abell 3411-3412 by combining radio, X-ray and optical observations. This discovery indicates that fossil relativistic electrons from active galactic nuclei are re-accelerated at cluster shocks. It also implies that radio galaxies play an important role in governing the non-thermal component of the intracluster medium in merging clusters.

  4. The case for electron re-acceleration at galaxy cluster shocks

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    van Weeren, Reinout J.; Andrade-Santos, Felipe; Dawson, William A.

    On the largest scales, the Universe consists of voids and filaments making up the cosmic web. Galaxy clusters are located at the knots in this web, at the intersection of filaments. Clusters grow through accretion from these large-scale filaments and by mergers with other clusters and groups. In a growing number of galaxy clusters, elongated Mpc-sized radio sources have been found. Also known as radio relics, these regions of diffuse radio emission are thought to trace relativistic electrons in the intracluster plasma accelerated by low-Mach-number shocks generated by cluster–cluster merger events. A long-standing problem is how low-Mach-number shocks can acceleratemore » electrons so efficiently to explain the observed radio relics. Here, we report the discovery of a direct connection between a radio relic and a radio galaxy in the merging galaxy cluster Abell 3411–3412 by combining radio, X-ray and optical observations. This discovery indicates that fossil relativistic electrons from active galactic nuclei are re-accelerated at cluster shocks. Lastly, it also implies that radio galaxies play an important role in governing the non-thermal component of the intracluster medium in merging clusters.« less

  5. The case for electron re-acceleration at galaxy cluster shocks

    DOE PAGES

    van Weeren, Reinout J.; Andrade-Santos, Felipe; Dawson, William A.; ...

    2017-01-04

    On the largest scales, the Universe consists of voids and filaments making up the cosmic web. Galaxy clusters are located at the knots in this web, at the intersection of filaments. Clusters grow through accretion from these large-scale filaments and by mergers with other clusters and groups. In a growing number of galaxy clusters, elongated Mpc-sized radio sources have been found. Also known as radio relics, these regions of diffuse radio emission are thought to trace relativistic electrons in the intracluster plasma accelerated by low-Mach-number shocks generated by cluster–cluster merger events. A long-standing problem is how low-Mach-number shocks can acceleratemore » electrons so efficiently to explain the observed radio relics. Here, we report the discovery of a direct connection between a radio relic and a radio galaxy in the merging galaxy cluster Abell 3411–3412 by combining radio, X-ray and optical observations. This discovery indicates that fossil relativistic electrons from active galactic nuclei are re-accelerated at cluster shocks. Lastly, it also implies that radio galaxies play an important role in governing the non-thermal component of the intracluster medium in merging clusters.« less

  6. Practical Issues of Wireless Mobile Devices Usage with Downlink Optimization

    NASA Astrophysics Data System (ADS)

    Krejcar, Ondrej; Janckulik, Dalibor; Motalova, Leona

    Mobile device makers produce tens of new complex mobile devices per year to put users a special mobile device with a possibility to do anything, anywhere, anytime. These devices can operate full scale applications with nearly the same comfort as their desktop equivalents only with several limitations. One of such limitation is insufficient download on wireless connectivity in case of the large multimedia files. Main area of paper is in a possibilities description of solving this problem as well as the test of several new mobile devices along with server interface tests and common software descriptions. New devices have a full scale of wireless connectivity which can be used not only to communication with outer land. Several such possibilities of use are described. Mobile users will have also an online connection to internet all time powered on. Internet is mainly the web pages but the web services use is still accelerate up. The paper deal also with a possibility of maximum user amounts to have a connection at same time to current server type. At last the new kind of database access - Linq technology is compare to ADO.NET in response time meaning.

  7. Assessing the equivalence of Web-based and paper-and-pencil questionnaires using differential item and test functioning (DIF and DTF) analysis: a case of the Four-Dimensional Symptom Questionnaire (4DSQ).

    PubMed

    Terluin, Berend; Brouwers, Evelien P M; Marchand, Miquelle A G; de Vet, Henrica C W

    2018-05-01

    Many paper-and-pencil (P&P) questionnaires have been migrated to electronic platforms. Differential item and test functioning (DIF and DTF) analysis constitutes a superior research design to assess measurement equivalence across modes of administration. The purpose of this study was to demonstrate an item response theory (IRT)-based DIF and DTF analysis to assess the measurement equivalence of a Web-based version and the original P&P format of the Four-Dimensional Symptom Questionnaire (4DSQ), measuring distress, depression, anxiety, and somatization. The P&P group (n = 2031) and the Web group (n = 958) consisted of primary care psychology clients. Unidimensionality and local independence of the 4DSQ scales were examined using IRT and Yen's Q3. Bifactor modeling was used to assess the scales' essential unidimensionality. Measurement equivalence was assessed using IRT-based DIF analysis using a 3-stage approach: linking on the latent mean and variance, selection of anchor items, and DIF testing using the Wald test. DTF was evaluated by comparing expected scale scores as a function of the latent trait. The 4DSQ scales proved to be essentially unidimensional in both modalities. Five items, belonging to the distress and somatization scales, displayed small amounts of DIF. DTF analysis revealed that the impact of DIF on the scale level was negligible. IRT-based DIF and DTF analysis is demonstrated as a way to assess the equivalence of Web-based and P&P questionnaire modalities. Data obtained with the Web-based 4DSQ are equivalent to data obtained with the P&P version.

  8. Interactive effects of fire and large herbivores on web-building spiders.

    PubMed

    Foster, C N; Barton, P S; Wood, J T; Lindenmayer, D B

    2015-09-01

    Altered disturbance regimes are a major driver of biodiversity loss worldwide. Maintaining or re-creating natural disturbance regimes is therefore the focus of many conservation programmes. A key challenge, however, is to understand how co-occurring disturbances interact to affect biodiversity. We experimentally tested for the interactive effects of prescribed fire and large macropod herbivores on the web-building spider assemblage of a eucalypt forest understorey and investigated the role of vegetation in mediating these effects using path analysis. Fire had strong negative effects on the density of web-building spiders, which were partly mediated by effects on vegetation structure, while negative effects of large herbivores on web density were not related to changes in vegetation. Fire amplified the effects of large herbivores on spiders, both via vegetation-mediated pathways and by increasing herbivore activity. The importance of vegetation-mediated pathways and fire-herbivore interactions differed for web density and richness and also differed between web types. Our results demonstrate that for some groups of web-building spiders, the effects of co-occurring disturbance drivers may be mostly additive, whereas for other groups, interactions between drivers can amplify disturbance effects. In our study system, the use of prescribed fire in the presence of high densities of herbivores could lead to reduced densities and altered composition of web-building spiders, with potential cascading effects through the arthropod food web. Our study highlights the importance of considering both the independent and interactive effects of disturbances, as well as the mechanisms driving their effects, in the management of disturbance regimes.

  9. Comparative analysis of marine ecosystems: workshop on predator-prey interactions.

    PubMed

    Bailey, Kevin M; Ciannelli, Lorenzo; Hunsicker, Mary; Rindorf, Anna; Neuenfeldt, Stefan; Möllmann, Christian; Guichard, Frederic; Huse, Geir

    2010-10-23

    Climate and human influences on marine ecosystems are largely manifested by changes in predator-prey interactions. It follows that ecosystem-based management of the world's oceans requires a better understanding of food web relationships. An international workshop on predator-prey interactions in marine ecosystems was held at the Oregon State University, Corvallis, OR, USA on 16-18 March 2010. The meeting brought together scientists from diverse fields of expertise including theoretical ecology, animal behaviour, fish and seabird ecology, statistics, fisheries science and ecosystem modelling. The goals of the workshop were to critically examine the methods of scaling-up predator-prey interactions from local observations to systems, the role of shifting ecological processes with scale changes, and the complexity and organizational structure in trophic interactions.

  10. Web Spam, Social Propaganda and the Evolution of Search Engine Rankings

    NASA Astrophysics Data System (ADS)

    Metaxas, Panagiotis Takis

    Search Engines have greatly influenced the way we experience the web. Since the early days of the web, users have been relying on them to get informed and make decisions. When the web was relatively small, web directories were built and maintained using human experts to screen and categorize pages according to their characteristics. By the mid 1990's, however, it was apparent that the human expert model of categorizing web pages does not scale. The first search engines appeared and they have been evolving ever since, taking over the role that web directories used to play.

  11. Enabling Cross-Platform Clinical Decision Support through Web-Based Decision Support in Commercial Electronic Health Record Systems: Proposal and Evaluation of Initial Prototype Implementations

    PubMed Central

    Zhang, Mingyuan; Velasco, Ferdinand T.; Musser, R. Clayton; Kawamoto, Kensaku

    2013-01-01

    Enabling clinical decision support (CDS) across multiple electronic health record (EHR) systems has been a desired but largely unattained aim of clinical informatics, especially in commercial EHR systems. A potential opportunity for enabling such scalable CDS is to leverage vendor-supported, Web-based CDS development platforms along with vendor-supported application programming interfaces (APIs). Here, we propose a potential staged approach for enabling such scalable CDS, starting with the use of custom EHR APIs and moving towards standardized EHR APIs to facilitate interoperability. We analyzed three commercial EHR systems for their capabilities to support the proposed approach, and we implemented prototypes in all three systems. Based on these analyses and prototype implementations, we conclude that the approach proposed is feasible, already supported by several major commercial EHR vendors, and potentially capable of enabling cross-platform CDS at scale. PMID:24551426

  12. Improving PHENIX search with Solr, Nutch and Drupal.

    NASA Astrophysics Data System (ADS)

    Morrison, Dave; Sourikova, Irina

    2012-12-01

    During its 20 years of R&D, construction and operation the PHENIX experiment at the Relativistic Heavy Ion Collider (RHIC) has accumulated large amounts of proprietary collaboration data that is hosted on many servers around the world and is not open for commercial search engines for indexing and searching. The legacy search infrastructure did not scale well with the fast growing PHENIX document base and produced results inadequate in both precision and recall. After considering the possible alternatives that would provide an aggregated, fast, full text search of a variety of data sources and file formats we decided to use Nutch [1] as a web crawler and Solr [2] as a search engine. To present XML-based Solr search results in a user-friendly format we use Drupal [3] as a web interface to Solr. We describe the experience of building a federated search for a heterogeneous collection of 10 million PHENIX documents with Nutch, Solr and Drupal.

  13. A Proposal for a Thesaurus for Web Services in Solar Radiation

    NASA Technical Reports Server (NTRS)

    Gschwind, Benoit; Menard, Lionel; Ranchin, Thierry; Wald, Lucien; Stackhouse, Paul W., Jr.

    2007-01-01

    Metadata are necessary to discover, describe and exchange any type of information, resource and service at a large scale. A significant amount of effort has been made in the field of geography and environment to establish standards. Efforts still remain to address more specific domains such as renewable energies. This communication focuses on solar energy and more specifically on aspects in solar radiation that relate to geography and meteorology. A thesaurus in solar radiation is proposed for the keys elements in solar radiation namely time, space and radiation types. The importance of time-series in solar radiation is outlined and attributes of the key elements are discussed. An XML schema for encoding metadata is proposed. The exploitation of such a schema in web services is discussed. This proposal is a first attempt at establishing a thesaurus for describing data and applications in solar radiation.

  14. VizieR Online Data Catalog: Horizon MareNostrum cosmological run (Gay+, 2010)

    NASA Astrophysics Data System (ADS)

    Gay, C.; Pichon, C.; Le Borgne, D.; Teyssier, R.; Sousbie, T.; Devriendt, J.

    2010-11-01

    The correlation between the large-scale distribution of galaxies and their spectroscopic properties at z=1.5 is investigated using the Horizon MareNostrum cosmological run. We have extracted a large sample of 105 galaxies from this large hydrodynamical simulation featuring standard galaxy formation physics. Spectral synthesis is applied to these single stellar populations to generate spectra and colours for all galaxies. We use the skeleton as a tracer of the cosmic web and study how our galaxy catalogue depends on the distance to the skeleton. We show that galaxies closer to the skeleton tend to be redder but that the effect is mostly due to the proximity of large haloes at the nodes of the skeleton, rather than the filaments themselves. The virtual catalogues (spectroscopical properties of the MareNostrum galaxies at various redshifts) are available online at http://www.iap.fr/users/pichon/MareNostrum/catalogues. (7 data files).

  15. Comparison Between a Self-Administered and Supervised Version of a Web-Based Cognitive Test Battery: Results From the NutriNet-Santé Cohort Study.

    PubMed

    Assmann, Karen E; Bailet, Marion; Lecoffre, Amandine C; Galan, Pilar; Hercberg, Serge; Amieva, Hélène; Kesse-Guyot, Emmanuelle

    2016-04-05

    Dementia is a major public health problem, and repeated cognitive data from large epidemiological studies could help to develop efficient measures of early prevention. Data collection by self-administered online tools could drastically reduce the logistical and financial burden of such large-scale investigations. In this context, it is important to obtain data concerning the comparability of such new online tools with traditional, supervised modes of cognitive assessment. Our objective was to compare self-administration of the Web-based NutriNet-Santé cognitive test battery (NutriCog) with administration by a neuropsychologist. The test battery included four tests, measuring, among others aspects, psychomotor speed, attention, executive function, episodic memory, working memory, and associative memory. Both versions of the cognitive battery were completed by 189 volunteers (either self-administered version first, n=99, or supervised version first, n=90). Subjects also completed a satisfaction questionnaire. Concordance was assessed by Spearman correlation. Agreement between both versions varied according to the investigated cognitive task and outcome variable. Spearman correlations ranged between .42 and .73. Moreover, a majority of participants responded that they "absolutely" or "rather" agreed that the duration of the self-administered battery was acceptable (184/185, 99.5%), that the tasks were amusing (162/185, 87.6%), that the instructions were sufficiently detailed (168/185; 90.8%) and understandable (164/185, 88.7%), and that they had overall enjoyed the test battery (182/185, 98.4%). The self-administered version of the Web-based NutriCog cognitive test battery provided similar information as the supervised version. Thus, integrating repeated cognitive evaluations into large cohorts via the implementation of self-administered online versions of traditional test batteries appears to be feasible.

  16. Portals for Real-Time Earthquake Data and Forecasting: Challenge and Promise (Invited)

    NASA Astrophysics Data System (ADS)

    Rundle, J. B.; Holliday, J. R.; Graves, W. R.; Feltstykket, R.; Donnellan, A.; Glasscoe, M. T.

    2013-12-01

    Earthquake forecasts have been computed by a variety of countries world-wide for over two decades. For the most part, forecasts have been computed for insurance, reinsurance and underwriters of catastrophe bonds. However, recent events clearly demonstrate that mitigating personal risk is becoming the responsibility of individual members of the public. Open access to a variety of web-based forecasts, tools, utilities and information is therefore required. Portals for data and forecasts present particular challenges, and require the development of both apps and the client/server architecture to deliver the basic information in real time. The basic forecast model we consider is the Natural Time Weibull (NTW) method (JBR et al., Phys. Rev. E, 86, 021106, 2012). This model uses small earthquakes (';seismicity-based models') to forecast the occurrence of large earthquakes, via data-mining algorithms combined with the ANSS earthquake catalog. This method computes large earthquake probabilities using the number of small earthquakes that have occurred in a region since the last large earthquake. Localizing these forecasts in space so that global forecasts can be computed in real time presents special algorithmic challenges, which we describe in this talk. Using 25 years of data from the ANSS California-Nevada catalog of earthquakes, we compute real-time global forecasts at a grid scale of 0.1o. We analyze and monitor the performance of these models using the standard tests, which include the Reliability/Attributes and Receiver Operating Characteristic (ROC) tests. It is clear from much of the analysis that data quality is a major limitation on the accurate computation of earthquake probabilities. We discuss the challenges of serving up these datasets over the web on web-based platforms such as those at www.quakesim.org , www.e-decider.org , and www.openhazards.com.

  17. Compression-based aggregation model for medical web services.

    PubMed

    Al-Shammary, Dhiah; Khalil, Ibrahim

    2010-01-01

    Many organizations such as hospitals have adopted Cloud Web services in applying their network services to avoid investing heavily computing infrastructure. SOAP (Simple Object Access Protocol) is the basic communication protocol of Cloud Web services that is XML based protocol. Generally,Web services often suffer congestions and bottlenecks as a result of the high network traffic that is caused by the large XML overhead size. At the same time, the massive load on Cloud Web services in terms of the large demand of client requests has resulted in the same problem. In this paper, two XML-aware aggregation techniques that are based on exploiting the compression concepts are proposed in order to aggregate the medical Web messages and achieve higher message size reduction.

  18. Development of web-GIS system for analysis of georeferenced geophysical data

    NASA Astrophysics Data System (ADS)

    Okladnikov, I.; Gordov, E. P.; Titov, A. G.; Bogomolov, V. Y.; Genina, E.; Martynova, Y.; Shulgina, T. M.

    2012-12-01

    Georeferenced datasets (meteorological databases, modeling and reanalysis results, remote sensing products, etc.) are currently actively used in numerous applications including modeling, interpretation and forecast of climatic and ecosystem changes for various spatial and temporal scales. Due to inherent heterogeneity of environmental datasets as well as their huge size which might constitute up to tens terabytes for a single dataset at present studies in the area of climate and environmental change require a special software support. A dedicated web-GIS information-computational system for analysis of georeferenced climatological and meteorological data has been created. The information-computational system consists of 4 basic parts: computational kernel developed using GNU Data Language (GDL), a set of PHP-controllers run within specialized web-portal, JavaScript class libraries for development of typical components of web mapping application graphical user interface (GUI) based on AJAX technology, and an archive of geophysical datasets. Computational kernel comprises of a number of dedicated modules for querying and extraction of data, mathematical and statistical data analysis, visualization, and preparing output files in geoTIFF and netCDF format containing processing results. Specialized web-portal consists of a web-server Apache, complying OGC standards Geoserver software which is used as a base for presenting cartographical information over the Web, and a set of PHP-controllers implementing web-mapping application logic and governing computational kernel. JavaScript libraries aiming at graphical user interface development are based on GeoExt library combining ExtJS Framework and OpenLayers software. The archive of geophysical data consists of a number of structured environmental datasets represented by data files in netCDF, HDF, GRIB, ESRI Shapefile formats. For processing by the system are available: two editions of NCEP/NCAR Reanalysis, JMA/CRIEPI JRA-25 Reanalysis, ECMWF ERA-40 Reanalysis, ECMWF ERA Interim Reanalysis, MRI/JMA APHRODITE's Water Resources Project Reanalysis, DWD Global Precipitation Climatology Centre's data, GMAO Modern Era-Retrospective analysis for Research and Applications, meteorological observational data for the territory of the former USSR for the 20th century, results of modeling by global and regional climatological models, and others. The system is already involved into a scientific research process. Particularly, recently the system was successfully used for analysis of Siberia climate changes and its impact in the region. The Web-GIS information-computational system for geophysical data analysis provides specialists involved into multidisciplinary research projects with reliable and practical instruments for complex analysis of climate and ecosystems changes on global and regional scales. Using it even unskilled user without specific knowledge can perform computational processing and visualization of large meteorological, climatological and satellite monitoring datasets through unified web-interface in a common graphical web-browser. This work is partially supported by the Ministry of education and science of the Russian Federation (contract #07.514.114044), projects IV.31.1.5, IV.31.2.7, RFBR grants #10-07-00547a, #11-05-01190a, and integrated project SB RAS #131.

  19. Interacting Science through Web Quests

    ERIC Educational Resources Information Center

    Unal, Ahmet; Karakus, Melek Altiparmak

    2016-01-01

    The purpose of this paper is to examine the effects of WebQuests on elementary students' science achievement, attitude towards science and attitude towards web supported education in teaching 7th grade subjects (Ecosystems, Solar System). With regard to this research, "Science Achievement Test," "Attitude towards Science Scale"…

  20. A Query Integrator and Manager for the Query Web

    PubMed Central

    Brinkley, James F.; Detwiler, Landon T.

    2012-01-01

    We introduce two concepts: the Query Web as a layer of interconnected queries over the document web and the semantic web, and a Query Web Integrator and Manager (QI) that enables the Query Web to evolve. QI permits users to write, save and reuse queries over any web accessible source, including other queries saved in other installations of QI. The saved queries may be in any language (e.g. SPARQL, XQuery); the only condition for interconnection is that the queries return their results in some form of XML. This condition allows queries to chain off each other, and to be written in whatever language is appropriate for the task. We illustrate the potential use of QI for several biomedical use cases, including ontology view generation using a combination of graph-based and logical approaches, value set generation for clinical data management, image annotation using terminology obtained from an ontology web service, ontology-driven brain imaging data integration, small-scale clinical data integration, and wider-scale clinical data integration. Such use cases illustrate the current range of applications of QI and lead us to speculate about the potential evolution from smaller groups of interconnected queries into a larger query network that layers over the document and semantic web. The resulting Query Web could greatly aid researchers and others who now have to manually navigate through multiple information sources in order to answer specific questions. PMID:22531831

  1. Experience in using commercial clouds in CMS

    NASA Astrophysics Data System (ADS)

    Bauerdick, L.; Bockelman, B.; Dykstra, D.; Fuess, S.; Garzoglio, G.; Girone, M.; Gutsche, O.; Holzman, B.; Hufnagel, D.; Kim, H.; Kennedy, R.; Mason, D.; Spentzouris, P.; Timm, S.; Tiradani, A.; Vaandering, E.; CMS Collaboration

    2017-10-01

    Historically high energy physics computing has been performed on large purpose-built computing systems. In the beginning there were single site computing facilities, which evolved into the Worldwide LHC Computing Grid (WLCG) used today. The vast majority of the WLCG resources are used for LHC computing and the resources are scheduled to be continuously used throughout the year. In the last several years there has been an explosion in capacity and capability of commercial and academic computing clouds. Cloud resources are highly virtualized and intended to be able to be flexibly deployed for a variety of computing tasks. There is a growing interest amongst the cloud providers to demonstrate the capability to perform large scale scientific computing. In this presentation we will discuss results from the CMS experiment using the Fermilab HEPCloud Facility, which utilized both local Fermilab resources and Amazon Web Services (AWS). The goal was to work with AWS through a matching grant to demonstrate a sustained scale approximately equal to half of the worldwide processing resources available to CMS. We will discuss the planning and technical challenges involved in organizing the most IO intensive CMS workflows on a large-scale set of virtualized resource provisioned by the Fermilab HEPCloud. We will describe the data handling and data management challenges. Also, we will discuss the economic issues and cost and operational efficiency comparison to our dedicated resources. At the end we will consider the changes in the working model of HEP computing in a domain with the availability of large scale resources scheduled at peak times.

  2. Experience in using commercial clouds in CMS

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Bauerdick, L.; Bockelman, B.; Dykstra, D.

    Historically high energy physics computing has been performed on large purposebuilt computing systems. In the beginning there were single site computing facilities, which evolved into the Worldwide LHC Computing Grid (WLCG) used today. The vast majority of the WLCG resources are used for LHC computing and the resources are scheduled to be continuously used throughout the year. In the last several years there has been an explosion in capacity and capability of commercial and academic computing clouds. Cloud resources are highly virtualized and intended to be able to be flexibly deployed for a variety of computing tasks. There is amore » growing interest amongst the cloud providers to demonstrate the capability to perform large scale scientific computing. In this presentation we will discuss results from the CMS experiment using the Fermilab HEPCloud Facility, which utilized both local Fermilab resources and Amazon Web Services (AWS). The goal was to work with AWS through a matching grant to demonstrate a sustained scale approximately equal to half of the worldwide processing resources available to CMS. We will discuss the planning and technical challenges involved in organizing the most IO intensive CMS workflows on a large-scale set of virtualized resource provisioned by the Fermilab HEPCloud. We will describe the data handling and data management challenges. Also, we will discuss the economic issues and cost and operational efficiency comparison to our dedicated resources. At the end we will consider the changes in the working model of HEP computing in a domain with the availability of large scale resources scheduled at peak times.« less

  3. Applications of Lexical Link Analysis Web Service for Large-Scale Automation, Validation, Discovery, Visualization, and Real-Time Program Awareness

    DTIC Science & Technology

    2012-10-23

    Quantum Intelligence, Inc. She was principal investigator (PI) for six contracts awarded by the DoD Small Business Innovation Research (SBIR) Program. She...with at OSD? I hope you don’t mind if I indulge in a little ‘stream of consciousness ’ musing about where LLA could really add value. One of the...implemented by Quantum Intelligence, Inc. (QI, 2001–2012). The unique contribution of this architecture is to leverage a peer-to-peer agent network

  4. SEEDisCs: How Clusters Form and Galaxies Transform in the Cosmic Web

    NASA Astrophysics Data System (ADS)

    Jablonka, P.

    2017-08-01

    This presentation introduces a new survey, the Spatial Extended EDisCS Survey (SEEDisCS), which aims at understanding how clusters assemble and the level at which galaxies are preprocessed before falling on the cluster cores. I focus on the changes in galaxy properties in the cluster large scale environments, and how we can get constraints on the timescale of star formation quenching. I also discuss new ALMA CO observations, which trace the fate of the galaxy cold gas content along the infalling paths towards the cluster cores.

  5. Participatory visualization with Wordle.

    PubMed

    Viégas, Fernanda B; Wattenberg, Martin; Feinberg, Jonathan

    2009-01-01

    We discuss the design and usage of "Wordle," a web-based tool for visualizing text. Wordle creates tag-cloud-like displays that give careful attention to typography, color, and composition. We describe the algorithms used to balance various aesthetic criteria and create the distinctive Wordle layouts. We then present the results of a study of Wordle usage, based both on spontaneous behaviour observed in the wild, and on a large-scale survey of Wordle users. The results suggest that Wordles have become a kind of medium of expression, and that a "participatory culture" has arisen around them.

  6. An Approach to Function Annotation for Proteins of Unknown Function (PUFs) in the Transcriptome of Indian Mulberry.

    PubMed

    Dhanyalakshmi, K H; Naika, Mahantesha B N; Sajeevan, R S; Mathew, Oommen K; Shafi, K Mohamed; Sowdhamini, Ramanathan; N Nataraja, Karaba

    2016-01-01

    The modern sequencing technologies are generating large volumes of information at the transcriptome and genome level. Translation of this information into a biological meaning is far behind the race due to which a significant portion of proteins discovered remain as proteins of unknown function (PUFs). Attempts to uncover the functional significance of PUFs are limited due to lack of easy and high throughput functional annotation tools. Here, we report an approach to assign putative functions to PUFs, identified in the transcriptome of mulberry, a perennial tree commonly cultivated as host of silkworm. We utilized the mulberry PUFs generated from leaf tissues exposed to drought stress at whole plant level. A sequence and structure based computational analysis predicted the probable function of the PUFs. For rapid and easy annotation of PUFs, we developed an automated pipeline by integrating diverse bioinformatics tools, designated as PUFs Annotation Server (PUFAS), which also provides a web service API (Application Programming Interface) for a large-scale analysis up to a genome. The expression analysis of three selected PUFs annotated by the pipeline revealed abiotic stress responsiveness of the genes, and hence their potential role in stress acclimation pathways. The automated pipeline developed here could be extended to assign functions to PUFs from any organism in general. PUFAS web server is available at http://caps.ncbs.res.in/pufas/ and the web service is accessible at http://capservices.ncbs.res.in/help/pufas.

  7. Pfarao: a web application for protein family analysis customized for cytoskeletal and motor proteins (CyMoBase).

    PubMed

    Odronitz, Florian; Kollmar, Martin

    2006-11-29

    Annotation of protein sequences of eukaryotic organisms is crucial for the understanding of their function in the cell. Manual annotation is still by far the most accurate way to correctly predict genes. The classification of protein sequences, their phylogenetic relation and the assignment of function involves information from various sources. This often leads to a collection of heterogeneous data, which is hard to track. Cytoskeletal and motor proteins consist of large and diverse superfamilies comprising up to several dozen members per organism. Up to date there is no integrated tool available to assist in the manual large-scale comparative genomic analysis of protein families. Pfarao (Protein Family Application for Retrieval, Analysis and Organisation) is a database driven online working environment for the analysis of manually annotated protein sequences and their relationship. Currently, the system can store and interrelate a wide range of information about protein sequences, species, phylogenetic relations and sequencing projects as well as links to literature and domain predictions. Sequences can be imported from multiple sequence alignments that are generated during the annotation process. A web interface allows to conveniently browse the database and to compile tabular and graphical summaries of its content. We implemented a protein sequence-centric web application to store, organize, interrelate, and present heterogeneous data that is generated in manual genome annotation and comparative genomics. The application has been developed for the analysis of cytoskeletal and motor proteins (CyMoBase) but can easily be adapted for any protein.

  8. Mining a Web Citation Database for Author Co-Citation Analysis.

    ERIC Educational Resources Information Center

    He, Yulan; Hui, Siu Cheung

    2002-01-01

    Proposes a mining process to automate author co-citation analysis based on the Web Citation Database, a data warehouse for storing citation indices of Web publications. Describes the use of agglomerative hierarchical clustering for author clustering and multidimensional scaling for displaying author cluster maps, and explains PubSearch, a…

  9. Rainbow: a tool for large-scale whole-genome sequencing data analysis using cloud computing.

    PubMed

    Zhao, Shanrong; Prenger, Kurt; Smith, Lance; Messina, Thomas; Fan, Hongtao; Jaeger, Edward; Stephens, Susan

    2013-06-27

    Technical improvements have decreased sequencing costs and, as a result, the size and number of genomic datasets have increased rapidly. Because of the lower cost, large amounts of sequence data are now being produced by small to midsize research groups. Crossbow is a software tool that can detect single nucleotide polymorphisms (SNPs) in whole-genome sequencing (WGS) data from a single subject; however, Crossbow has a number of limitations when applied to multiple subjects from large-scale WGS projects. The data storage and CPU resources that are required for large-scale whole genome sequencing data analyses are too large for many core facilities and individual laboratories to provide. To help meet these challenges, we have developed Rainbow, a cloud-based software package that can assist in the automation of large-scale WGS data analyses. Here, we evaluated the performance of Rainbow by analyzing 44 different whole-genome-sequenced subjects. Rainbow has the capacity to process genomic data from more than 500 subjects in two weeks using cloud computing provided by the Amazon Web Service. The time includes the import and export of the data using Amazon Import/Export service. The average cost of processing a single sample in the cloud was less than 120 US dollars. Compared with Crossbow, the main improvements incorporated into Rainbow include the ability: (1) to handle BAM as well as FASTQ input files; (2) to split large sequence files for better load balance downstream; (3) to log the running metrics in data processing and monitoring multiple Amazon Elastic Compute Cloud (EC2) instances; and (4) to merge SOAPsnp outputs for multiple individuals into a single file to facilitate downstream genome-wide association studies. Rainbow is a scalable, cost-effective, and open-source tool for large-scale WGS data analysis. For human WGS data sequenced by either the Illumina HiSeq 2000 or HiSeq 2500 platforms, Rainbow can be used straight out of the box. Rainbow is available for third-party implementation and use, and can be downloaded from http://s3.amazonaws.com/jnj_rainbow/index.html.

  10. Rainbow: a tool for large-scale whole-genome sequencing data analysis using cloud computing

    PubMed Central

    2013-01-01

    Background Technical improvements have decreased sequencing costs and, as a result, the size and number of genomic datasets have increased rapidly. Because of the lower cost, large amounts of sequence data are now being produced by small to midsize research groups. Crossbow is a software tool that can detect single nucleotide polymorphisms (SNPs) in whole-genome sequencing (WGS) data from a single subject; however, Crossbow has a number of limitations when applied to multiple subjects from large-scale WGS projects. The data storage and CPU resources that are required for large-scale whole genome sequencing data analyses are too large for many core facilities and individual laboratories to provide. To help meet these challenges, we have developed Rainbow, a cloud-based software package that can assist in the automation of large-scale WGS data analyses. Results Here, we evaluated the performance of Rainbow by analyzing 44 different whole-genome-sequenced subjects. Rainbow has the capacity to process genomic data from more than 500 subjects in two weeks using cloud computing provided by the Amazon Web Service. The time includes the import and export of the data using Amazon Import/Export service. The average cost of processing a single sample in the cloud was less than 120 US dollars. Compared with Crossbow, the main improvements incorporated into Rainbow include the ability: (1) to handle BAM as well as FASTQ input files; (2) to split large sequence files for better load balance downstream; (3) to log the running metrics in data processing and monitoring multiple Amazon Elastic Compute Cloud (EC2) instances; and (4) to merge SOAPsnp outputs for multiple individuals into a single file to facilitate downstream genome-wide association studies. Conclusions Rainbow is a scalable, cost-effective, and open-source tool for large-scale WGS data analysis. For human WGS data sequenced by either the Illumina HiSeq 2000 or HiSeq 2500 platforms, Rainbow can be used straight out of the box. Rainbow is available for third-party implementation and use, and can be downloaded from http://s3.amazonaws.com/jnj_rainbow/index.html. PMID:23802613

  11. Development of an automated large-scale protein-crystallization and monitoring system for high-throughput protein-structure analyses.

    PubMed

    Hiraki, Masahiko; Kato, Ryuichi; Nagai, Minoru; Satoh, Tadashi; Hirano, Satoshi; Ihara, Kentaro; Kudo, Norio; Nagae, Masamichi; Kobayashi, Masanori; Inoue, Michio; Uejima, Tamami; Oda, Shunichiro; Chavas, Leonard M G; Akutsu, Masato; Yamada, Yusuke; Kawasaki, Masato; Matsugaki, Naohiro; Igarashi, Noriyuki; Suzuki, Mamoru; Wakatsuki, Soichi

    2006-09-01

    Protein crystallization remains one of the bottlenecks in crystallographic analysis of macromolecules. An automated large-scale protein-crystallization system named PXS has been developed consisting of the following subsystems, which proceed in parallel under unified control software: dispensing precipitants and protein solutions, sealing crystallization plates, carrying robot, incubators, observation system and image-storage server. A sitting-drop crystallization plate specialized for PXS has also been designed and developed. PXS can set up 7680 drops for vapour diffusion per hour, which includes time for replenishing supplies such as disposable tips and crystallization plates. Images of the crystallization drops are automatically recorded according to a preprogrammed schedule and can be viewed by users remotely using web-based browser software. A number of protein crystals were successfully produced and several protein structures could be determined directly from crystals grown by PXS. In other cases, X-ray quality crystals were obtained by further optimization by manual screening based on the conditions found by PXS.

  12. Collaborative mining and interpretation of large-scale data for biomedical research insights.

    PubMed

    Tsiliki, Georgia; Karacapilidis, Nikos; Christodoulou, Spyros; Tzagarakis, Manolis

    2014-01-01

    Biomedical research becomes increasingly interdisciplinary and collaborative in nature. Researchers need to efficiently and effectively collaborate and make decisions by meaningfully assembling, mining and analyzing available large-scale volumes of complex multi-faceted data residing in different sources. In line with related research directives revealing that, in spite of the recent advances in data mining and computational analysis, humans can easily detect patterns which computer algorithms may have difficulty in finding, this paper reports on the practical use of an innovative web-based collaboration support platform in a biomedical research context. Arguing that dealing with data-intensive and cognitively complex settings is not a technical problem alone, the proposed platform adopts a hybrid approach that builds on the synergy between machine and human intelligence to facilitate the underlying sense-making and decision making processes. User experience shows that the platform enables more informed and quicker decisions, by displaying the aggregated information according to their needs, while also exploiting the associated human intelligence.

  13. Collaborative Mining and Interpretation of Large-Scale Data for Biomedical Research Insights

    PubMed Central

    Tsiliki, Georgia; Karacapilidis, Nikos; Christodoulou, Spyros; Tzagarakis, Manolis

    2014-01-01

    Biomedical research becomes increasingly interdisciplinary and collaborative in nature. Researchers need to efficiently and effectively collaborate and make decisions by meaningfully assembling, mining and analyzing available large-scale volumes of complex multi-faceted data residing in different sources. In line with related research directives revealing that, in spite of the recent advances in data mining and computational analysis, humans can easily detect patterns which computer algorithms may have difficulty in finding, this paper reports on the practical use of an innovative web-based collaboration support platform in a biomedical research context. Arguing that dealing with data-intensive and cognitively complex settings is not a technical problem alone, the proposed platform adopts a hybrid approach that builds on the synergy between machine and human intelligence to facilitate the underlying sense-making and decision making processes. User experience shows that the platform enables more informed and quicker decisions, by displaying the aggregated information according to their needs, while also exploiting the associated human intelligence. PMID:25268270

  14. Distributed 3D Information Visualization - Towards Integration of the Dynamic 3D Graphics and Web Services

    NASA Astrophysics Data System (ADS)

    Vucinic, Dean; Deen, Danny; Oanta, Emil; Batarilo, Zvonimir; Lacor, Chris

    This paper focuses on visualization and manipulation of graphical content in distributed network environments. The developed graphical middleware and 3D desktop prototypes were specialized for situational awareness. This research was done in the LArge Scale COllaborative decision support Technology (LASCOT) project, which explored and combined software technologies to support human-centred decision support system for crisis management (earthquake, tsunami, flooding, airplane or oil-tanker incidents, chemical, radio-active or other pollutants spreading, etc.). The performed state-of-the-art review did not identify any publicly available large scale distributed application of this kind. Existing proprietary solutions rely on the conventional technologies and 2D representations. Our challenge was to apply the "latest" available technologies, such Java3D, X3D and SOAP, compatible with average computer graphics hardware. The selected technologies are integrated and we demonstrate: the flow of data, which originates from heterogeneous data sources; interoperability across different operating systems and 3D visual representations to enhance the end-users interactions.

  15. Iron enrichment stimulates toxic diatom production in high-nitrate, low-chlorophyll areas

    PubMed Central

    Trick, Charles G.; Bill, Brian D.; Cochlan, William P.; Wells, Mark L.; Trainer, Vera L.; Pickell, Lisa D.

    2010-01-01

    Oceanic high-nitrate, low-chlorophyll environments have been highlighted for potential large-scale iron fertilizations to help mitigate global climate change. Controversy surrounds these initiatives, both in the degree of carbon removal and magnitude of ecosystem impacts. Previous open ocean enrichment experiments have shown that iron additions stimulate growth of the toxigenic diatom genus Pseudonitzschia. Most Pseudonitzschia species in coastal waters produce the neurotoxin domoic acid (DA), with their blooms causing detrimental marine ecosystem impacts, but oceanic Pseudonitzschia species are considered nontoxic. Here we demonstrate that the sparse oceanic Pseudonitzschia community at the high-nitrate, low-chlorophyll Ocean Station PAPA (50° N, 145° W) produces approximately 200 pg DA L−1 in response to iron addition, that DA alters phytoplankton community structure to benefit Pseudonitzschia, and that oceanic cell isolates are toxic. Given the negative effects of DA in coastal food webs, these findings raise serious concern over the net benefit and sustainability of large-scale iron fertilizations. PMID:20231473

  16. Web-Based Assessment of Mental Well-Being in Early Adolescence: A Reliability Study.

    PubMed

    Hamann, Christoph; Schultze-Lutter, Frauke; Tarokh, Leila

    2016-06-15

    The ever-increasing use of the Internet among adolescents represents an emerging opportunity for researchers to gain access to larger samples, which can be queried over several years longitudinally. Among adolescents, young adolescents (ages 11 to 13 years) are of particular interest to clinicians as this is a transitional stage, during which depressive and anxiety symptoms often emerge. However, it remains unclear whether these youngest adolescents can accurately answer questions about their mental well-being using a Web-based platform. The aim of the study was to examine the accuracy of responses obtained from Web-based questionnaires by comparing Web-based with paper-and-pencil versions of depression and anxiety questionnaires. The primary outcome was the score on the depression and anxiety questionnaires under two conditions: (1) paper-and-pencil and (2) Web-based versions. Twenty-eight adolescents (aged 11-13 years, mean age 12.78 years and SD 0.78; 18 females, 64%) were randomly assigned to complete either the paper-and-pencil or the Web-based questionnaire first. Intraclass correlation coefficients (ICCs) were calculated to measure intrarater reliability. Intraclass correlation coefficients were calculated separately for depression (Children's Depression Inventory, CDI) and anxiety (Spence Children's Anxiety Scale, SCAS) questionnaires. On average, it took participants 17 minutes (SD 6) to answer 116 questions online. Intraclass correlation coefficient analysis revealed high intrarater reliability when comparing Web-based with paper-and-pencil responses for both CDI (ICC=.88; P<.001) and the SCAS (ICC=.95; P<.001). According to published criteria, both of these values are in the "almost perfect" category indicating the highest degree of reliability. The results of the study show an excellent reliability of Web-based assessment in 11- to 13-year-old children as compared with the standard paper-pencil assessment. Furthermore, we found that Web-based assessments with young adolescents are highly feasible, with all enrolled participants completing the Web-based form. As early adolescence is a time of remarkable social and behavioral changes, these findings open up new avenues for researchers from diverse fields who are interested in studying large samples of young adolescents over time.

  17. Popularity versus similarity in growing networks

    NASA Astrophysics Data System (ADS)

    Krioukov, Dmitri; Papadopoulos, Fragkiskos; Kitsak, Maksim; Serrano, Mariangeles; Boguna, Marian

    2012-02-01

    Preferential attachment is a powerful mechanism explaining the emergence of scaling in growing networks. If new connections are established preferentially to more popular nodes in a network, then the network is scale-free. Here we show that not only popularity but also similarity is a strong force shaping the network structure and dynamics. We develop a framework where new connections, instead of preferring popular nodes, optimize certain trade-offs between popularity and similarity. The framework admits a geometric interpretation, in which preferential attachment emerges from local optimization processes. As opposed to preferential attachment, the optimization framework accurately describes large-scale evolution of technological (Internet), social (web of trust), and biological (E.coli metabolic) networks, predicting the probability of new links in them with a remarkable precision. The developed framework can thus be used for predicting new links in evolving networks, and provides a different perspective on preferential attachment as an emergent phenomenon.

  18. Properties of galaxies reproduced by a hydrodynamic simulation.

    PubMed

    Vogelsberger, M; Genel, S; Springel, V; Torrey, P; Sijacki, D; Xu, D; Snyder, G; Bird, S; Nelson, D; Hernquist, L

    2014-05-08

    Previous simulations of the growth of cosmic structures have broadly reproduced the 'cosmic web' of galaxies that we see in the Universe, but failed to create a mixed population of elliptical and spiral galaxies, because of numerical inaccuracies and incomplete physical models. Moreover, they were unable to track the small-scale evolution of gas and stars to the present epoch within a representative portion of the Universe. Here we report a simulation that starts 12 million years after the Big Bang, and traces 13 billion years of cosmic evolution with 12 billion resolution elements in a cube of 106.5 megaparsecs a side. It yields a reasonable population of ellipticals and spirals, reproduces the observed distribution of galaxies in clusters and characteristics of hydrogen on large scales, and at the same time matches the 'metal' and hydrogen content of galaxies on small scales.

  19. PhosphOrtholog: a web-based tool for cross-species mapping of orthologous protein post-translational modifications.

    PubMed

    Chaudhuri, Rima; Sadrieh, Arash; Hoffman, Nolan J; Parker, Benjamin L; Humphrey, Sean J; Stöckli, Jacqueline; Hill, Adam P; James, David E; Yang, Jean Yee Hwa

    2015-08-19

    Most biological processes are influenced by protein post-translational modifications (PTMs). Identifying novel PTM sites in different organisms, including humans and model organisms, has expedited our understanding of key signal transduction mechanisms. However, with increasing availability of deep, quantitative datasets in diverse species, there is a growing need for tools to facilitate cross-species comparison of PTM data. This is particularly important because functionally important modification sites are more likely to be evolutionarily conserved; yet cross-species comparison of PTMs is difficult since they often lie in structurally disordered protein domains. Current tools that address this can only map known PTMs between species based on known orthologous phosphosites, and do not enable the cross-species mapping of newly identified modification sites. Here, we addressed this by developing a web-based software tool, PhosphOrtholog ( www.phosphortholog.com ) that accurately maps protein modification sites between different species. This facilitates the comparison of datasets derived from multiple species, and should be a valuable tool for the proteomics community. Here we describe PhosphOrtholog, a web-based application for mapping known and novel orthologous PTM sites from experimental data obtained from different species. PhosphOrtholog is the only generic and automated tool that enables cross-species comparison of large-scale PTM datasets without relying on existing PTM databases. This is achieved through pairwise sequence alignment of orthologous protein residues. To demonstrate its utility we apply it to two sets of human and rat muscle phosphoproteomes generated following insulin and exercise stimulation, respectively, and one publicly available mouse phosphoproteome following cellular stress revealing high mapping and coverage efficiency. Although coverage statistics are dataset dependent, PhosphOrtholog increased the number of cross-species mapped sites in all our example data sets by more than double when compared to those recovered using existing resources such as PhosphoSitePlus. PhosphOrtholog is the first tool that enables mapping of thousands of novel and known protein phosphorylation sites across species, accessible through an easy-to-use web interface. Identification of conserved PTMs across species from large-scale experimental data increases our knowledgebase of functional PTM sites. Moreover, PhosphOrtholog is generic being applicable to other PTM datasets such as acetylation, ubiquitination and methylation.

  20. CoryneRegNet 4.0 – A reference database for corynebacterial gene regulatory networks

    PubMed Central

    Baumbach, Jan

    2007-01-01

    Background Detailed information on DNA-binding transcription factors (the key players in the regulation of gene expression) and on transcriptional regulatory interactions of microorganisms deduced from literature-derived knowledge, computer predictions and global DNA microarray hybridization experiments, has opened the way for the genome-wide analysis of transcriptional regulatory networks. The large-scale reconstruction of these networks allows the in silico analysis of cell behavior in response to changing environmental conditions. We previously published CoryneRegNet, an ontology-based data warehouse of corynebacterial transcription factors and regulatory networks. Initially, it was designed to provide methods for the analysis and visualization of the gene regulatory network of Corynebacterium glutamicum. Results Now we introduce CoryneRegNet release 4.0, which integrates data on the gene regulatory networks of 4 corynebacteria, 2 mycobacteria and the model organism Escherichia coli K12. As the previous versions, CoryneRegNet provides a web-based user interface to access the database content, to allow various queries, and to support the reconstruction, analysis and visualization of regulatory networks at different hierarchical levels. In this article, we present the further improved database content of CoryneRegNet along with novel analysis features. The network visualization feature GraphVis now allows the inter-species comparisons of reconstructed gene regulatory networks and the projection of gene expression levels onto that networks. Therefore, we added stimulon data directly into the database, but also provide Web Service access to the DNA microarray analysis platform EMMA. Additionally, CoryneRegNet now provides a SOAP based Web Service server, which can easily be consumed by other bioinformatics software systems. Stimulons (imported from the database, or uploaded by the user) can be analyzed in the context of known transcriptional regulatory networks to predict putative contradictions or further gene regulatory interactions. Furthermore, it integrates protein clusters by means of heuristically solving the weighted graph cluster editing problem. In addition, it provides Web Service based access to up to date gene annotation data from GenDB. Conclusion The release 4.0 of CoryneRegNet is a comprehensive system for the integrated analysis of procaryotic gene regulatory networks. It is a versatile systems biology platform to support the efficient and large-scale analysis of transcriptional regulation of gene expression in microorganisms. It is publicly available at . PMID:17986320

  1. NaviCell: a web-based environment for navigation, curation and maintenance of large molecular interaction maps

    PubMed Central

    2013-01-01

    Background Molecular biology knowledge can be formalized and systematically represented in a computer-readable form as a comprehensive map of molecular interactions. There exist an increasing number of maps of molecular interactions containing detailed and step-wise description of various cell mechanisms. It is difficult to explore these large maps, to organize discussion of their content and to maintain them. Several efforts were recently made to combine these capabilities together in one environment, and NaviCell is one of them. Results NaviCell is a web-based environment for exploiting large maps of molecular interactions, created in CellDesigner, allowing their easy exploration, curation and maintenance. It is characterized by a combination of three essential features: (1) efficient map browsing based on Google Maps; (2) semantic zooming for viewing different levels of details or of abstraction of the map and (3) integrated web-based blog for collecting community feedback. NaviCell can be easily used by experts in the field of molecular biology for studying molecular entities of interest in the context of signaling pathways and crosstalk between pathways within a global signaling network. NaviCell allows both exploration of detailed molecular mechanisms represented on the map and a more abstract view of the map up to a top-level modular representation. NaviCell greatly facilitates curation, maintenance and updating the comprehensive maps of molecular interactions in an interactive and user-friendly fashion due to an imbedded blogging system. Conclusions NaviCell provides user-friendly exploration of large-scale maps of molecular interactions, thanks to Google Maps and WordPress interfaces, with which many users are already familiar. Semantic zooming which is used for navigating geographical maps is adopted for molecular maps in NaviCell, making any level of visualization readable. In addition, NaviCell provides a framework for community-based curation of maps. PMID:24099179

  2. NaviCell: a web-based environment for navigation, curation and maintenance of large molecular interaction maps.

    PubMed

    Kuperstein, Inna; Cohen, David P A; Pook, Stuart; Viara, Eric; Calzone, Laurence; Barillot, Emmanuel; Zinovyev, Andrei

    2013-10-07

    Molecular biology knowledge can be formalized and systematically represented in a computer-readable form as a comprehensive map of molecular interactions. There exist an increasing number of maps of molecular interactions containing detailed and step-wise description of various cell mechanisms. It is difficult to explore these large maps, to organize discussion of their content and to maintain them. Several efforts were recently made to combine these capabilities together in one environment, and NaviCell is one of them. NaviCell is a web-based environment for exploiting large maps of molecular interactions, created in CellDesigner, allowing their easy exploration, curation and maintenance. It is characterized by a combination of three essential features: (1) efficient map browsing based on Google Maps; (2) semantic zooming for viewing different levels of details or of abstraction of the map and (3) integrated web-based blog for collecting community feedback. NaviCell can be easily used by experts in the field of molecular biology for studying molecular entities of interest in the context of signaling pathways and crosstalk between pathways within a global signaling network. NaviCell allows both exploration of detailed molecular mechanisms represented on the map and a more abstract view of the map up to a top-level modular representation. NaviCell greatly facilitates curation, maintenance and updating the comprehensive maps of molecular interactions in an interactive and user-friendly fashion due to an imbedded blogging system. NaviCell provides user-friendly exploration of large-scale maps of molecular interactions, thanks to Google Maps and WordPress interfaces, with which many users are already familiar. Semantic zooming which is used for navigating geographical maps is adopted for molecular maps in NaviCell, making any level of visualization readable. In addition, NaviCell provides a framework for community-based curation of maps.

  3. BiNChE: a web tool and library for chemical enrichment analysis based on the ChEBI ontology.

    PubMed

    Moreno, Pablo; Beisken, Stephan; Harsha, Bhavana; Muthukrishnan, Venkatesh; Tudose, Ilinca; Dekker, Adriano; Dornfeldt, Stefanie; Taruttis, Franziska; Grosse, Ivo; Hastings, Janna; Neumann, Steffen; Steinbeck, Christoph

    2015-02-21

    Ontology-based enrichment analysis aids in the interpretation and understanding of large-scale biological data. Ontologies are hierarchies of biologically relevant groupings. Using ontology annotations, which link ontology classes to biological entities, enrichment analysis methods assess whether there is a significant over or under representation of entities for ontology classes. While many tools exist that run enrichment analysis for protein sets annotated with the Gene Ontology, there are only a few that can be used for small molecules enrichment analysis. We describe BiNChE, an enrichment analysis tool for small molecules based on the ChEBI Ontology. BiNChE displays an interactive graph that can be exported as a high-resolution image or in network formats. The tool provides plain, weighted and fragment analysis based on either the ChEBI Role Ontology or the ChEBI Structural Ontology. BiNChE aids in the exploration of large sets of small molecules produced within Metabolomics or other Systems Biology research contexts. The open-source tool provides easy and highly interactive web access to enrichment analysis with the ChEBI ontology tool and is additionally available as a standalone library.

  4. WHAM!: a web-based visualization suite for user-defined analysis of metagenomic shotgun sequencing data.

    PubMed

    Devlin, Joseph C; Battaglia, Thomas; Blaser, Martin J; Ruggles, Kelly V

    2018-06-25

    Exploration of large data sets, such as shotgun metagenomic sequence or expression data, by biomedical experts and medical professionals remains as a major bottleneck in the scientific discovery process. Although tools for this purpose exist for 16S ribosomal RNA sequencing analysis, there is a growing but still insufficient number of user-friendly interactive visualization workflows for easy data exploration and figure generation. The development of such platforms for this purpose is necessary to accelerate and streamline microbiome laboratory research. We developed the Workflow Hub for Automated Metagenomic Exploration (WHAM!) as a web-based interactive tool capable of user-directed data visualization and statistical analysis of annotated shotgun metagenomic and metatranscriptomic data sets. WHAM! includes exploratory and hypothesis-based gene and taxa search modules for visualizing differences in microbial taxa and gene family expression across experimental groups, and for creating publication quality figures without the need for command line interface or in-house bioinformatics. WHAM! is an interactive and customizable tool for downstream metagenomic and metatranscriptomic analysis providing a user-friendly interface allowing for easy data exploration by microbiome and ecological experts to facilitate discovery in multi-dimensional and large-scale data sets.

  5. Designing next-generation platforms for evaluating scientific output: what scientists can learn from the social web

    PubMed Central

    Yarkoni, Tal

    2012-01-01

    Traditional pre-publication peer review of scientific output is a slow, inefficient, and unreliable process. Efforts to replace or supplement traditional evaluation models with open evaluation platforms that leverage advances in information technology are slowly gaining traction, but remain in the early stages of design and implementation. Here I discuss a number of considerations relevant to the development of such platforms. I focus particular attention on three core elements that next-generation evaluation platforms should strive to emphasize, including (1) open and transparent access to accumulated evaluation data, (2) personalized and highly customizable performance metrics, and (3) appropriate short-term incentivization of the userbase. Because all of these elements have already been successfully implemented on a large scale in hundreds of existing social web applications, I argue that development of new scientific evaluation platforms should proceed largely by adapting existing techniques rather than engineering entirely new evaluation mechanisms. Successful implementation of open evaluation platforms has the potential to substantially advance both the pace and the quality of scientific publication and evaluation, and the scientific community has a vested interest in shifting toward such models as soon as possible. PMID:23060783

  6. Ergatis: a web interface and scalable software system for bioinformatics workflows

    PubMed Central

    Orvis, Joshua; Crabtree, Jonathan; Galens, Kevin; Gussman, Aaron; Inman, Jason M.; Lee, Eduardo; Nampally, Sreenath; Riley, David; Sundaram, Jaideep P.; Felix, Victor; Whitty, Brett; Mahurkar, Anup; Wortman, Jennifer; White, Owen; Angiuoli, Samuel V.

    2010-01-01

    Motivation: The growth of sequence data has been accompanied by an increasing need to analyze data on distributed computer clusters. The use of these systems for routine analysis requires scalable and robust software for data management of large datasets. Software is also needed to simplify data management and make large-scale bioinformatics analysis accessible and reproducible to a wide class of target users. Results: We have developed a workflow management system named Ergatis that enables users to build, execute and monitor pipelines for computational analysis of genomics data. Ergatis contains preconfigured components and template pipelines for a number of common bioinformatics tasks such as prokaryotic genome annotation and genome comparisons. Outputs from many of these components can be loaded into a Chado relational database. Ergatis was designed to be accessible to a broad class of users and provides a user friendly, web-based interface. Ergatis supports high-throughput batch processing on distributed compute clusters and has been used for data management in a number of genome annotation and comparative genomics projects. Availability: Ergatis is an open-source project and is freely available at http://ergatis.sourceforge.net Contact: jorvis@users.sourceforge.net PMID:20413634

  7. Bounce Back Now! Protocol of a population-based randomized controlled trial to examine the efficacy of a Web-based intervention with disaster-affected families.

    PubMed

    Ruggiero, Kenneth J; Davidson, Tatiana M; McCauley, Jenna; Gros, Kirstin Stauffacher; Welsh, Kyleen; Price, Matthew; Resnick, Heidi S; Danielson, Carla Kmett; Soltis, Kathryn; Galea, Sandro; Kilpatrick, Dean G; Saunders, Benjamin E; Nissenboim, Josh; Muzzy, Wendy; Fleeman, Anna; Amstadter, Ananda B

    2015-01-01

    Disasters have far-reaching and potentially long-lasting effects on youth and families. Research has consistently shown a clear increase in the prevalence of several mental health disorders after disasters, including depression and posttraumatic stress disorder. Widely accessible evidence-based interventions are needed to address this unmet need for youth and families, who are underrepresented in disaster research. Rapid growth in Internet and Smartphone access, as well as several Web based evaluation studies with various adult populations has shown that Web-based interventions are likely to be feasible in this context and can improve clinical outcomes. Such interventions also are generally cost-effective, can be targeted or personalized, and can easily be integrated in a stepped care approach to screening and intervention delivery. This is a protocol paper that describes an innovative study design in which we evaluate a self-help Web-based resource, Bounce Back Now, with a population-based sample of disaster affected adolescents and families. The paper includes description and justification for sampling selection and procedures, selection of assessment measures and methods, design of the intervention, and statistical evaluation of critical outcomes. Unique features of this study design include the use of address-based sampling to recruit a population-based sample of disaster-affected adolescents and parents, telephone and Web-based assessments, and development and evaluation of a highly individualized Web intervention for adolescents. Challenges related to large-scale evaluation of technology-delivered interventions with high-risk samples in time-sensitive research are discussed, as well as implications for future research and practice. Published by Elsevier Inc.

  8. Bounce Back Now! Protocol of a Population-Based Randomized Controlled Trial to Examine the Efficacy of a Web-based Intervention with Disaster-Affected Families

    PubMed Central

    Ruggiero, Kenneth J.; Davidson, Tatiana M.; McCauley, Jenna; Gros, Kirstin Stauffacher; Welsh, Kyleen; Price, Matthew; Resnick, Heidi S.; Danielson, Carla Kmett; Soltis, Kathryn; Galea, Sandro; Kilpatrick, Dean G.; Saunders, Benjamin E.; Nissenboim, Josh; Muzzy, Wendy; Fleeman, Anna; Amstadter, Ananda B.

    2014-01-01

    Disasters have far-reaching and potentially long-lasting effects on youth and families. Research has consistently shown a clear increase in the prevalence of several mental health disorders after disasters, including depression and posttraumatic stress disorder. Widely accessible evidence-based interventions are needed to address this unmet need for youth and families, who are underrepresented in disaster research. Rapid growth in Internet and Smartphone access, as well as several web based evaluation studies with various adult populations has shown that web-based interventions are likely to be feasible in this context and can improve clinical outcomes. Such interventions also are generally cost-effective, can be targeted or personalized, and can easily be integrated in a stepped care approach to screening and intervention delivery. This is a protocol paper that describes an innovative study design in which we evaluate a self-help web-based resource, Bounce Back Now, with a population-based sample of disaster affected adolescents and families. The paper includes description and justification for sampling selection and procedures, selection of assessment measures and methods, design of the intervention, and statistical evaluation of critical outcomes. Unique features of this study design include the use of address-based sampling to recruit a population-based sample of disaster-affected adolescents and parents, telephone and web-based assessments, and development and evaluation of a highly individualized web intervention for adolescents. Challenges related to large-scale evaluation of technology-delivered interventions with high-risk samples in time-sensitive research are discussed, as well as implications for future research and practice. PMID:25478956

  9. Web-based education in anesthesiology: a critical overview.

    PubMed

    Doyle, D John

    2008-12-01

    The purpose of this review is to discuss the rise of web-based educational resources available to the anesthesiology community. Recent developments of particular importance include the growth of 'Web 2.0' resources, the development of the concepts of 'open access' and 'information philanthropy', and the expansion of web-based medical simulation software products.In addition, peer review of online educational resources has now come of age. The worldwide web has made available a large variety of valuable medical information and education resources only dreamed of two decades ago. To a large extent,these developments represent a shift in the focus of medical education resources to emphasize free access to materials and to encourage collaborative development efforts.

  10. How Public Is the Web?: Robots, Access, and Scholarly Communication.

    ERIC Educational Resources Information Center

    Snyder, Herbert; Rosenbaum, Howard

    1998-01-01

    Examines the use of Robot Exclusion Protocol (REP) to restrict the access of search engine robots to 10 major United States university Web sites. An analysis of Web site searching and interviews with Web server administrators shows that the decision to use this procedure is largely technical and is typically made by the Web server administrator.…

  11. Web 2.0 and internet social networking: a new tool for disaster management?--lessons from Taiwan.

    PubMed

    Huang, Cheng-Min; Chan, Edward; Hyder, Adnan A

    2010-10-06

    Internet social networking tools and the emerging web 2.0 technologies are providing a new way for web users and health workers in information sharing and knowledge dissemination. Based on the characters of immediate, two-way and large scale of impact, the internet social networking tools have been utilized as a solution in emergency response during disasters. This paper highlights the use of internet social networking in disaster emergency response and public health management of disasters by focusing on a case study of the typhoon Morakot disaster in Taiwan. In the case of typhoon disaster in Taiwan, internet social networking and mobile technology were found to be helpful for community residents, professional emergency rescuers, and government agencies in gathering and disseminating real-time information, regarding volunteer recruitment and relief supplies allocation. We noted that if internet tools are to be integrated in the development of emergency response system, the accessibility, accuracy, validity, feasibility, privacy and the scalability of itself should be carefully considered especially in the effort of applying it in resource poor settings. This paper seeks to promote an internet-based emergency response system by integrating internet social networking and information communication technology into central government disaster management system. Web-based networking provides two-way communication which establishes a reliable and accessible tunnel for proximal and distal users in disaster preparedness and management.

  12. Semantic Web repositories for genomics data using the eXframe platform

    PubMed Central

    2014-01-01

    Background With the advent of inexpensive assay technologies, there has been an unprecedented growth in genomics data as well as the number of databases in which it is stored. In these databases, sample annotation using ontologies and controlled vocabularies is becoming more common. However, the annotation is rarely available as Linked Data, in a machine-readable format, or for standardized queries using SPARQL. This makes large-scale reuse, or integration with other knowledge bases very difficult. Methods To address this challenge, we have developed the second generation of our eXframe platform, a reusable framework for creating online repositories of genomics experiments. This second generation model now publishes Semantic Web data. To accomplish this, we created an experiment model that covers provenance, citations, external links, assays, biomaterials used in the experiment, and the data collected during the process. The elements of our model are mapped to classes and properties from various established biomedical ontologies. Resource Description Framework (RDF) data is automatically produced using these mappings and indexed in an RDF store with a built-in Sparql Protocol and RDF Query Language (SPARQL) endpoint. Conclusions Using the open-source eXframe software, institutions and laboratories can create Semantic Web repositories of their experiments, integrate it with heterogeneous resources and make it interoperable with the vast Semantic Web of biomedical knowledge. PMID:25093072

  13. Autonomous system for Web-based microarray image analysis.

    PubMed

    Bozinov, Daniel

    2003-12-01

    Software-based feature extraction from DNA microarray images still requires human intervention on various levels. Manual adjustment of grid and metagrid parameters, precise alignment of superimposed grid templates and gene spots, or simply identification of large-scale artifacts have to be performed beforehand to reliably analyze DNA signals and correctly quantify their expression values. Ideally, a Web-based system with input solely confined to a single microarray image and a data table as output containing measurements for all gene spots would directly transform raw image data into abstracted gene expression tables. Sophisticated algorithms with advanced procedures for iterative correction function can overcome imminent challenges in image processing. Herein is introduced an integrated software system with a Java-based interface on the client side that allows for decentralized access and furthermore enables the scientist to instantly employ the most updated software version at any given time. This software tool is extended from PixClust as used in Extractiff incorporated with Java Web Start deployment technology. Ultimately, this setup is destined for high-throughput pipelines in genome-wide medical diagnostics labs or microarray core facilities aimed at providing fully automated service to its users.

  14. Trophic disruption: a meta-analysis of how habitat fragmentation affects resource consumption in terrestrial arthropod systems.

    PubMed

    Martinson, Holly M; Fagan, William F

    2014-09-01

    Habitat fragmentation is a complex process that affects ecological systems in diverse ways, altering everything from population persistence to ecosystem function. Despite widespread recognition that habitat fragmentation can influence food web interactions, consensus on the factors underlying variation in the impacts of fragmentation across systems remains elusive. In this study, we conduct a systematic review and meta-analysis to quantify the effects of habitat fragmentation and spatial habitat structure on resource consumption in terrestrial arthropod food webs. Across 419 studies, we found a negative overall effect of fragmentation on resource consumption. Variation in effect size was extensive but predictable. Specifically, resource consumption was reduced on small, isolated habitat fragments, higher at patch edges, and neutral with respect to landscape-scale spatial variables. In general, resource consumption increased in fragmented settings for habitat generalist consumers but decreased for specialist consumers. Our study demonstrates widespread disruption of trophic interactions in fragmented habitats and describes variation among studies that is largely predictable based on the ecological traits of the interacting species. We highlight future prospects for understanding how changes in spatial habitat structure may influence trophic modules and food webs. © 2014 John Wiley & Sons Ltd/CNRS.

  15. Intuitive web-based experimental design for high-throughput biomedical data.

    PubMed

    Friedrich, Andreas; Kenar, Erhan; Kohlbacher, Oliver; Nahnsen, Sven

    2015-01-01

    Big data bioinformatics aims at drawing biological conclusions from huge and complex biological datasets. Added value from the analysis of big data, however, is only possible if the data is accompanied by accurate metadata annotation. Particularly in high-throughput experiments intelligent approaches are needed to keep track of the experimental design, including the conditions that are studied as well as information that might be interesting for failure analysis or further experiments in the future. In addition to the management of this information, means for an integrated design and interfaces for structured data annotation are urgently needed by researchers. Here, we propose a factor-based experimental design approach that enables scientists to easily create large-scale experiments with the help of a web-based system. We present a novel implementation of a web-based interface allowing the collection of arbitrary metadata. To exchange and edit information we provide a spreadsheet-based, humanly readable format. Subsequently, sample sheets with identifiers and metainformation for data generation facilities can be created. Data files created after measurement of the samples can be uploaded to a datastore, where they are automatically linked to the previously created experimental design model.

  16. Visualization on the Web of 20 Years of Crop Rotation and Wildlife Co-Evolutions

    NASA Astrophysics Data System (ADS)

    Plumejeaud-Perreau, Christine; Poitevin, Cyril; Bretagnolle, Vincent

    2018-05-01

    The accumulation of evidences of the effects of intensive agricultural practices against wildlife fauna and flora, and biodiversity in general, has been largely published in scientific papers (Tildman, 1999). However, data serving as sup-port to their conclusions are often kept hidden behind research institutions. This paper presents a data visualization sys-tem opened on the Web allowing citizens to get a comprehensive access to data issued from such kind of research institution, collected for more than 20 years. The Web Information System has been thought in order to ease the comparison of data issues from various databases describing the same object, the agricultural landscape, at different scales and through different observation devices. An interactive visualization is proposed in order to check co-evolution of fauna and flora together with agricultural practices. It mixes aerial orthoimagery produced since 1950 with vectorial data showing the evolutions of agricultural parcels with those of a few sentinel species such as the Montagu's harrier. This is made through a composition of maps, charts and time lines, and specific tools for comparison. A particular concern is given to the observation effort bias in order to show meaningful statistical aggregates.

  17. ChlamyCyc: an integrative systems biology database and web-portal for Chlamydomonas reinhardtii.

    PubMed

    May, Patrick; Christian, Jan-Ole; Kempa, Stefan; Walther, Dirk

    2009-05-04

    The unicellular green alga Chlamydomonas reinhardtii is an important eukaryotic model organism for the study of photosynthesis and plant growth. In the era of modern high-throughput technologies there is an imperative need to integrate large-scale data sets from high-throughput experimental techniques using computational methods and database resources to provide comprehensive information about the molecular and cellular organization of a single organism. In the framework of the German Systems Biology initiative GoFORSYS, a pathway database and web-portal for Chlamydomonas (ChlamyCyc) was established, which currently features about 250 metabolic pathways with associated genes, enzymes, and compound information. ChlamyCyc was assembled using an integrative approach combining the recently published genome sequence, bioinformatics methods, and experimental data from metabolomics and proteomics experiments. We analyzed and integrated a combination of primary and secondary database resources, such as existing genome annotations from JGI, EST collections, orthology information, and MapMan classification. ChlamyCyc provides a curated and integrated systems biology repository that will enable and assist in systematic studies of fundamental cellular processes in Chlamydomonas. The ChlamyCyc database and web-portal is freely available under http://chlamycyc.mpimp-golm.mpg.de.

  18. DBMap: a TreeMap-based framework for data navigation and visualization of brain research registry

    NASA Astrophysics Data System (ADS)

    Zhang, Ming; Zhang, Hong; Tjandra, Donny; Wong, Stephen T. C.

    2003-05-01

    The purpose of this study is to investigate and apply a new, intuitive and space-conscious visualization framework to facilitate efficient data presentation and exploration of large-scale data warehouses. We have implemented the DBMap framework for the UCSF Brain Research Registry. Such a novel utility would facilitate medical specialists and clinical researchers in better exploring and evaluating a number of attributes organized in the brain research registry. The current UCSF Brain Research Registry consists of a federation of disease-oriented database modules, including Epilepsy, Brain Tumor, Intracerebral Hemorrphage, and CJD (Creuzfeld-Jacob disease). These database modules organize large volumes of imaging and non-imaging data to support Web-based clinical research. While the data warehouse supports general information retrieval and analysis, there lacks an effective way to visualize and present the voluminous and complex data stored. This study investigates whether the TreeMap algorithm can be adapted to display and navigate categorical biomedical data warehouse or registry. TreeMap is a space constrained graphical representation of large hierarchical data sets, mapped to a matrix of rectangles, whose size and color represent interested database fields. It allows the display of a large amount of numerical and categorical information in limited real estate of computer screen with an intuitive user interface. The paper will describe, DBMap, the proposed new data visualization framework for large biomedical databases. Built upon XML, Java and JDBC technologies, the prototype system includes a set of software modules that reside in the application server tier and provide interface to backend database tier and front-end Web tier of the brain registry.

  19. EarthServer: Use of Rasdaman as a data store for use in visualisation of complex EO data

    NASA Astrophysics Data System (ADS)

    Clements, Oliver; Walker, Peter; Grant, Mike

    2013-04-01

    The European Commission FP7 project EarthServer is establishing open access and ad-hoc analytics on extreme-size Earth Science data, based on and extending cutting-edge Array Database technology. EarthServer is built around the Rasdaman Raster Data Manager which extends standard relational database systems with the ability to store and retrieve multi-dimensional raster data of unlimited size through an SQL style query language. Rasdaman facilitates visualisation of data by providing several Open Geospatial Consortium (OGC) standard interfaces through its web services wrapper, Petascope. These include the well established standards, Web Coverage Service (WCS) and Web Map Service (WMS) as well as the emerging standard, Web Coverage Processing Service (WCPS). The WCPS standard allows the running of ad-hoc queries on the data stored within Rasdaman, creating an infrastructure where users are not restricted by bandwidth when manipulating or querying huge datasets. Here we will show that the use of EarthServer technologies and infrastructure allows access and visualisation of massive scale data through a web client with only marginal bandwidth use as opposed to the current mechanism of copying huge amounts of data to create visualisations locally. For example if a user wanted to generate a plot of global average chlorophyll for a complete decade time series they would only have to download the result instead of Terabytes of data. Firstly we will present a brief overview of the capabilities of Rasdaman and the WCPS query language to introduce the ways in which it is used in a visualisation tool chain. We will show that there are several ways in which WCPS can be utilised to create both standard and novel web based visualisations. An example of a standard visualisation is the production of traditional 2d plots, allowing users the ability to plot data products easily. However, the query language allows the creation of novel/custom products, which can then immediately be plotted with the same system. For more complex multi-spectral data, WCPS allows the user to explore novel combinations of bands in standard band-ratio algorithms through a web browser with dynamic updating of the resultant image. To visualise very large datasets Rasdaman has the capability to dynamically scale a dataset or query result so that it can be appraised quickly for use in later unscaled queries. All of these techniques are accessible through a web based GIS interface increasing the number of potential users of the system. Lastly we will show the advances in dynamic web based 3D visualisations being explored within the EarthServer project. By utilising the emerging declarative 3D web standard X3DOM as a tool to visualise the results of WCPS queries we introduce several possible benefits, including quick appraisal of data for outliers or anomalous data points and visualisation of the uncertainty of data alongside the actual data values.

  20. InterMine Webservices for Phytozome (Rev2)

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Carlson, Joseph; Goodstein, David; Rokhsar, Dan

    2014-07-10

    A datawarehousing framework for information provides a useful infrastructure for providers and users of genomic data. For providers, the infrastructure give them a consistent mechanism for extracting raw data. While for the users, the web services supported by the software allows them to make complex, and often unique, queries of the data. Previously, phytozome.net used BioMart to provide the infrastructure. As the complexity, scale and diversity of the dataset as grown, we decided to implement an InterMine web service on our servers. This change was largely motivated by the ability to have a more complex table structure and richer webmore » reporting mechanism than BioMart. For InterMine to achieve its more complex database schema it requires an XML description of the data and an appropriate loader. Unlimited one-to-many and many-to-many relationship between the tables can be enabled in the schema. We have implemented support for:1.) Genomes and annotations for the data in Phytozome. This set is the 48 organisms currently stored in a back end CHADO datastore. The data loaders are modified versions of the CHADO data adapters from FlyMine. 2.) Interproscan results from all proteins in the Phytozome database. 3.) Clusters of proteins into a grouped heirarchically by similarity. 4.) Cufflinks results from tissue-specific RNA-Seq data of Phytozome organisms. 5.) Diversity data (GATK and SnpEFF results) from a set of individual organism. The last two datatypes are new in this implementation of our web services. We anticipate that the scale of these data will increase considerably in the near future.« less

Top