ERIC Educational Resources Information Center
Wang, Kening; Mulvenon, Sean W.; Stegman, Charles; Anderson, Travis
2008-01-01
Google Maps API (Application Programming Interface), released in late June 2005 by Google, is an amazing technology that allows users to embed Google Maps in their own Web pages with JavaScript. Google Maps API has accelerated the development of new Google Maps based applications. This article reports a Web-based interactive mapping system…
Boulos, Maged N Kamel
2005-01-01
This eye-opener article aims at introducing the health GIS community to the emerging online consumer geoinformatics services from Google and Microsoft (MSN), and their potential utility in creating custom online interactive health maps. Using the programmable interfaces provided by Google and MSN, we created three interactive demonstrator maps of England's Strategic Health Authorities. These can be browsed online at – Google Maps API (Application Programming Interface) version, – Google Earth KML (Keyhole Markup Language) version, and – MSN Virtual Earth Map Control version. Google and MSN's worldwide distribution of "free" geospatial tools, imagery, and maps is to be commended as a significant step towards the ultimate "wikification" of maps and GIS. A discussion is provided of these emerging online mapping trends, their expected future implications and development directions, and associated individual privacy, national security and copyrights issues. Although ESRI have announced their planned response to Google (and MSN), it remains to be seen how their envisaged plans will materialize and compare to the offerings from Google and MSN, and also how Google and MSN mapping tools will further evolve in the near future. PMID:16176577
How to Display Hazards and other Scientific Data Using Google Maps
NASA Astrophysics Data System (ADS)
Venezky, D. Y.; Fee, J. M.
2007-12-01
The U.S. Geological Survey's (USGS) Volcano Hazard Program (VHP) is launching a map-based interface to display hazards information using the Google® Map API (Application Program Interface). Map-based interfaces provide a synoptic view of data, making patterns easier to detect and allowing users to quickly ascertain where hazards are in relation to major population and infrastructure centers. Several map-based interfaces are now simple to run on a web server, providing ideal platforms for sharing information with colleagues, emergency managers, and the public. There are three main steps to making data accessible on a map-based interface; formatting the input data, plotting the data on the map, and customizing the user interface. The presentation, "Creating Geospatial RSS and ATOM feeds for Map-based Interfaces" (Fee and Venezky, this session), reviews key features for map input data. Join us for this presentation on how to plot data in a geographic context and then format the display with images, custom markers, and links to external data. Examples will show how the VHP Volcano Status Map was created and how to plot a field trip with driving directions.
A Land-Use-Planning Simulation Using Google Earth
ERIC Educational Resources Information Center
Bodzin, Alec M.; Cirucci, Lori
2009-01-01
Google Earth (GE) is proving to be a valuable tool in the science classroom for understanding the environment and making responsible environmental decisions (Bodzin 2008). GE provides learners with a dynamic mapping experience using a simple interface with a limited range of functions. This interface makes geospatial analysis accessible and…
Using Google Maps to Access USGS Volcano Hazards Information
NASA Astrophysics Data System (ADS)
Venezky, D. Y.; Snedigar, S.; Guffanti, M.; Bailey, J. E.; Wall, B. G.
2006-12-01
The U.S. Geological Survey (USGS) Volcano Hazard Program (VHP) is revising the information architecture of our website to provide data within a geospatial context for emergency managers, educators, landowners in volcanic areas, researchers, and the general public. Using a map-based interface for displaying hazard information provides a synoptic view of volcanic activity along with the ability to quickly ascertain where hazards are in relation to major population and infrastructure centers. At the same time, the map interface provides a gateway for educators and the public to find information about volcanoes in their geographic context. A plethora of data visualization solutions are available that are flexible, customizable, and can be run on individual websites. We are currently using a Google map interface because it can be accessed immediately from a website (a downloadable viewer is not required), and it provides simple features for moving around and zooming within the large map area that encompasses U.S. volcanism. A text interface will also be available. The new VHP website will serve as a portal to information for each volcano the USGS monitors with icons for alert levels and aviation color codes. When a volcano is clicked, a window will provide additional information including links to maps, images, and real-time data, thereby connecting information from individual observatories, the Smithsonian Institution, and our partner universities. In addition to the VHP home page, many observatories and partners have detailed graphical interfaces to data and images that include the activity pages for the Alaska Volcano Observatory, the Smithsonian Google Earth files, and Yellowstone Volcano Observatory pictures and data. Users with varied requests such as raw data, scientific papers, images, or brief overviews expect to be able to quickly access information for their specialized needs. Over the next few years we will be gathering, cleansing, reorganizing, and posting data in multiple formats to meet these needs.
Learning to Map the Earth and Planets using a Google Earth - based Multi-student Game
NASA Astrophysics Data System (ADS)
De Paor, D. G.; Wild, S. C.; Dordevic, M.
2011-12-01
We report on progress in developing an interactive geological and geophysical mapping game employing the Google Earth, Google Moon, and Goole Mars virtual globes. Working in groups of four, students represent themselves on the Google Earth surface by selecting an avatar. One of the group drives to each field stop in a model vehicle using game-like controls. When they arrive at a field stop and get out of their field vehicle, students can control their own avatars' movements independently and can communicate with one another by text message. They are geo-fenced and receive automatic messages if they wander off target. Individual movements are logged and stored in a MySQL database for later analysis. Students collaborate on mapping decisions and submit a report to their instructor through a Javascript interface to the Google Earth API. Unlike real mapping, students are not restricted by geographic access and can engage in comparative mapping on different planets. Using newly developed techniques, they can also explore and map the sub-surface down to the core-mantle boundary. Virtual specimens created with a 3D scanner, Gigapan images of outcrops, and COLLADA models of mantle structures such as subducted lithospheric slabs all contribute to an engaging learning experience.
MaRGEE: Move and Rotate Google Earth Elements
NASA Astrophysics Data System (ADS)
Dordevic, Mladen M.; Whitmeyer, Steven J.
2015-12-01
Google Earth is recognized as a highly effective visualization tool for geospatial information. However, there remain serious limitations that have hindered its acceptance as a tool for research and education in the geosciences. One significant limitation is the inability to translate or rotate geometrical elements on the Google Earth virtual globe. Here we present a new JavaScript web application to "Move and Rotate Google Earth Elements" (MaRGEE). MaRGEE includes tools to simplify, translate, and rotate elements, add intermediate steps to a transposition, and batch process multiple transpositions. The transposition algorithm uses spherical geometry calculations, such as the haversine formula, to accurately reposition groups of points, paths, and polygons on the Google Earth globe without distortion. Due to the imminent deprecation of the Google Earth API and browser plugin, MaRGEE uses a Google Maps interface to facilitate and illustrate the transpositions. However, the inherent spatial distortions that result from the Google Maps Web Mercator projection are not apparent once the transposed elements are saved as a KML file and opened in Google Earth. Potential applications of the MaRGEE toolkit include tectonic reconstructions, the movements of glaciers or thrust sheets, and time-based animations of other large- and small-scale geologic processes.
Google Sky: A Digital View of the Night Sky
NASA Astrophysics Data System (ADS)
Connolly, A. Scranton, R.; Ornduff, T.
2008-11-01
From its inception Astronomy has been a visual science, from careful observations of the sky using the naked eye, to the use of telescopes and photographs to map the distribution of stars and galaxies, to the current era of digital cameras that can image the sky over many decades of the electromagnetic spectrum. Sky in Google Earth (http://earth.google.com) and Google Sky (http://www.google.com/sky) continue this tradition, providing an intuitive visual interface to some of the largest astronomical imaging surveys of the sky. Streaming multi-color imagery, catalogs, time domain data, as well as annotating interesting astronomical sources and events with placemarks, podcasts and videos, Sky provides a panchromatic view of the universe accessible to anyone with a computer. Beyond a simple exploration of the sky Google Sky enables users to create and share content with others around the world. With an open interface available on Linux, Mac OS X and Windows, and translations of the content into over 20 different languages we present Sky as the embodiment of a virtual telescope for discovery and sharing the excitement of astronomy and science as a whole.
What Google Maps can do for biomedical data dissemination: examples and a design study.
Jianu, Radu; Laidlaw, David H
2013-05-04
Biologists often need to assess whether unfamiliar datasets warrant the time investment required for more detailed exploration. Basing such assessments on brief descriptions provided by data publishers is unwieldy for large datasets that contain insights dependent on specific scientific questions. Alternatively, using complex software systems for a preliminary analysis may be deemed as too time consuming in itself, especially for unfamiliar data types and formats. This may lead to wasted analysis time and discarding of potentially useful data. We present an exploration of design opportunities that the Google Maps interface offers to biomedical data visualization. In particular, we focus on synergies between visualization techniques and Google Maps that facilitate the development of biological visualizations which have both low-overhead and sufficient expressivity to support the exploration of data at multiple scales. The methods we explore rely on displaying pre-rendered visualizations of biological data in browsers, with sparse yet powerful interactions, by using the Google Maps API. We structure our discussion around five visualizations: a gene co-regulation visualization, a heatmap viewer, a genome browser, a protein interaction network, and a planar visualization of white matter in the brain. Feedback from collaborative work with domain experts suggests that our Google Maps visualizations offer multiple, scale-dependent perspectives and can be particularly helpful for unfamiliar datasets due to their accessibility. We also find that users, particularly those less experienced with computer use, are attracted by the familiarity of the Google Maps API. Our five implementations introduce design elements that can benefit visualization developers. We describe a low-overhead approach that lets biologists access readily analyzed views of unfamiliar scientific datasets. We rely on pre-computed visualizations prepared by data experts, accompanied by sparse and intuitive interactions, and distributed via the familiar Google Maps framework. Our contributions are an evaluation demonstrating the validity and opportunities of this approach, a set of design guidelines benefiting those wanting to create such visualizations, and five concrete example visualizations.
What google maps can do for biomedical data dissemination: examples and a design study
2013-01-01
Background Biologists often need to assess whether unfamiliar datasets warrant the time investment required for more detailed exploration. Basing such assessments on brief descriptions provided by data publishers is unwieldy for large datasets that contain insights dependent on specific scientific questions. Alternatively, using complex software systems for a preliminary analysis may be deemed as too time consuming in itself, especially for unfamiliar data types and formats. This may lead to wasted analysis time and discarding of potentially useful data. Results We present an exploration of design opportunities that the Google Maps interface offers to biomedical data visualization. In particular, we focus on synergies between visualization techniques and Google Maps that facilitate the development of biological visualizations which have both low-overhead and sufficient expressivity to support the exploration of data at multiple scales. The methods we explore rely on displaying pre-rendered visualizations of biological data in browsers, with sparse yet powerful interactions, by using the Google Maps API. We structure our discussion around five visualizations: a gene co-regulation visualization, a heatmap viewer, a genome browser, a protein interaction network, and a planar visualization of white matter in the brain. Feedback from collaborative work with domain experts suggests that our Google Maps visualizations offer multiple, scale-dependent perspectives and can be particularly helpful for unfamiliar datasets due to their accessibility. We also find that users, particularly those less experienced with computer use, are attracted by the familiarity of the Google Maps API. Our five implementations introduce design elements that can benefit visualization developers. Conclusions We describe a low-overhead approach that lets biologists access readily analyzed views of unfamiliar scientific datasets. We rely on pre-computed visualizations prepared by data experts, accompanied by sparse and intuitive interactions, and distributed via the familiar Google Maps framework. Our contributions are an evaluation demonstrating the validity and opportunities of this approach, a set of design guidelines benefiting those wanting to create such visualizations, and five concrete example visualizations. PMID:23642009
Werts, Joshua D; Mikhailova, Elena A; Post, Christopher J; Sharp, Julia L
2012-04-01
Volunteered geographic information and social networking in a WebGIS has the potential to increase public participation in soil and water conservation, promote environmental awareness and change, and provide timely data that may be otherwise unavailable to policymakers in soil and water conservation management. The objectives of this study were: (1) to develop a framework for combining current technologies, computing advances, data sources, and social media; and (2) develop and test an online web mapping interface. The mapping interface integrates Microsoft Silverlight, Bing Maps, ArcGIS Server, Google Picasa Web Albums Data API, RSS, Google Analytics, and Facebook to create a rich user experience. The website allows the public to upload photos and attributes of their own subdivisions or sites they have identified and explore other submissions. The website was made available to the public in early February 2011 at http://www.AbandonedDevelopments.com and evaluated for its potential long-term success in a pilot study.
NASA Astrophysics Data System (ADS)
Werts, Joshua D.; Mikhailova, Elena A.; Post, Christopher J.; Sharp, Julia L.
2012-04-01
Volunteered geographic information and social networking in a WebGIS has the potential to increase public participation in soil and water conservation, promote environmental awareness and change, and provide timely data that may be otherwise unavailable to policymakers in soil and water conservation management. The objectives of this study were: (1) to develop a framework for combining current technologies, computing advances, data sources, and social media; and (2) develop and test an online web mapping interface. The mapping interface integrates Microsoft Silverlight, Bing Maps, ArcGIS Server, Google Picasa Web Albums Data API, RSS, Google Analytics, and Facebook to create a rich user experience. The website allows the public to upload photos and attributes of their own subdivisions or sites they have identified and explore other submissions. The website was made available to the public in early February 2011 at http://www.AbandonedDevelopments.com and evaluated for its potential long-term success in a pilot study.
Web GIS in practice V: 3-D interactive and real-time mapping in Second Life
Boulos, Maged N Kamel; Burden, David
2007-01-01
This paper describes technologies from Daden Limited for geographically mapping and accessing live news stories/feeds, as well as other real-time, real-world data feeds (e.g., Google Earth KML feeds and GeoRSS feeds) in the 3-D virtual world of Second Life, by plotting and updating the corresponding Earth location points on a globe or some other suitable form (in-world), and further linking those points to relevant information and resources. This approach enables users to visualise, interact with, and even walk or fly through, the plotted data in 3-D. Users can also do the reverse: put pins on a map in the virtual world, and then view the data points on the Web in Google Maps or Google Earth. The technologies presented thus serve as a bridge between mirror worlds like Google Earth and virtual worlds like Second Life. We explore the geo-data display potential of virtual worlds and their likely convergence with mirror worlds in the context of the future 3-D Internet or Metaverse, and reflect on the potential of such technologies and their future possibilities, e.g. their use to develop emergency/public health virtual situation rooms to effectively manage emergencies and disasters in real time. The paper also covers some of the issues associated with these technologies, namely user interface accessibility and individual privacy. PMID:18042275
Application based on ArcObject inquiry and Google maps demonstration to real estate database
NASA Astrophysics Data System (ADS)
Hwang, JinTsong
2007-06-01
Real estate industry in Taiwan has been flourishing in recent years. To acquire various and abundant information of real estate for sale is the same goal for the consumers and the brokerages. Therefore, before looking at the property, it is important to get all pertinent information possible. Not only this beneficial for the real estate agent as they can provide the sellers with the most information, thereby solidifying the interest of the buyer, but may also save time and the cost of manpower were something out of place. Most of the brokerage sites are aware of utilizes Internet as form of media for publicity however; the contents are limited to specific property itself and the functions of query are mostly just provided searching by condition. This paper proposes a query interface on website which gives function of zone query by spatial analysis for non-GIS users, developing a user-friendly interface with ArcObject in VB6, and query by condition. The inquiry results can show on the web page which is embedded functions of Google Maps and the UrMap API on it. In addition, the demonstration of inquiry results will give the multimedia present way which includes hyperlink to Google Earth with surrounding of the property, the Virtual Reality scene of house, panorama of interior of building and so on. Therefore, the website provides extra spatial solution for query and demonstration abundant information of real estate in two-dimensional and three-dimensional types of view.
Mandel, Jacob E; Morel-Ovalle, Louis; Boas, Franz E; Ziv, Etay; Yarmohammadi, Hooman; Deipolyi, Amy; Mohabir, Heeralall R; Erinjeri, Joseph P
2018-02-20
The purpose of this study is to determine whether a custom Google Maps application can optimize site selection when scheduling outpatient interventional radiology (IR) procedures within a multi-site hospital system. The Google Maps for Business Application Programming Interface (API) was used to develop an internal web application that uses real-time traffic data to determine estimated travel time (ETT; minutes) and estimated travel distance (ETD; miles) from a patient's home to each a nearby IR facility in our hospital system. Hypothetical patient home addresses based on the 33 cities comprising our institution's catchment area were used to determine the optimal IR site for hypothetical patients traveling from each city based on real-time traffic conditions. For 10/33 (30%) cities, there was discordance between the optimal IR site based on ETT and the optimal IR site based on ETD at non-rush hour time or rush hour time. By choosing to travel to an IR site based on ETT rather than ETD, patients from discordant cities were predicted to save an average of 7.29 min during non-rush hour (p = 0.03), and 28.80 min during rush hour (p < 0.001). Using a custom Google Maps application to schedule outpatients for IR procedures can effectively reduce patient travel time when more than one location providing IR procedures is available within the same hospital system.
Secure and Resilient Cloud Computing for the Department of Defense
2015-11-16
platform as a service (PaaS), and software as a service ( SaaS )—that target system administrators, developers, and end-users respectively (see Table 2...interfaces (API) and services Medium Amazon Elastic MapReduce, MathWorks Cloud, Red Hat OpenShift SaaS Full-fledged applications Low Google gMail
NASA Astrophysics Data System (ADS)
Minnett, R. C.; Koppers, A. A.; Staudigel, D.; Staudigel, H.
2008-12-01
EarthRef.org is comprehensive and convenient resource for Earth Science reference data and models. It encompasses four main portals: the Geochemical Earth Reference Model (GERM), the Magnetics Information Consortium (MagIC), the Seamount Biogeosciences Network (SBN), and the Enduring Resources for Earth Science Education (ERESE). Their underlying databases are publically available and the scientific community has contributed widely and is urged to continue to do so. However, the net result is a vast and largely heterogeneous warehouse of geospatial data ranging from carefully prepared maps of seamounts to geochemical data/metadata, daily reports from seagoing expeditions, large volumes of raw and processed multibeam data, images of paleomagnetic sampling sites, etc. This presents a considerable obstacle for integrating other rich media content, such as videos, images, data files, cruise tracks, and interoperable database results, without overwhelming the web user. The four EarthRef.org portals clearly lend themselves to a more intuitive user interface and has, therefore, been an invaluable test bed for the design and implementation of FlashMap, a versatile KML-driven geospatial browser written for reliability and speed in Adobe Flash. FlashMap allows layers of content to be loaded and displayed over a streaming high-resolution map which can be zoomed and panned similarly to Google Maps and Google Earth. Many organizations, from National Geographic to the USGS, have begun using Google Earth software to display geospatial content. However, Google Earth, as a desktop application, does not integrate cleanly with existing websites requiring the user to navigate away from the browser and focus on a separate application and Google Maps, written in Java Script, does not scale up reliably to large datasets. FlashMap remedies these problems as a web-based application that allows for seamless integration of the real-time display power of Google Earth and the flexibility of the web without losing scalability and control of the base maps. Our Flash-based application is fully compatible with KML (Keyhole Markup Language) 2.2, the most recent iteration of KML, allowing users with existing Google Earth KML files to effortlessly display their geospatial content embedded in a web page. As a test case for FlashMap, the annual Iron-Oxidizing Microbial Observatory (FeMO) dive cruise to the Loihi Seamount, in conjunction with data available from ongoing and published FeMO laboratory studies, showcases the flexibility of this single web-based application. With a KML 2.2 compatible web-service providing the content, any database can display results in FlashMap. The user can then hide and show multiple layers of content, potentially from several data sources, and rapidly digest a vast quantity of information to narrow the search results. This flexibility gives experienced users the ability to drill down to exactly the record they are looking for (SERC at Carleton College's educational application of FlashMap at http://serc.carleton.edu/sp/erese/activities/22223.html) and allows users familiar with Google Earth the ability to load and view geospatial data content within a browser from any computer with an internet connection.
Ingress in Geography: Portals to Academic Success?
ERIC Educational Resources Information Center
Davis, Michael
2017-01-01
Niantic Labs has developed an augmented virtual reality mobile app game called Ingress in which agents must seek out and control locations for their designated factions. The app uses the Google Maps interface along with GPS to enhance a geocaching-like experience with elements of other classical games such as capture-the-flag. This study aims to…
Immunochromatographic diagnostic test analysis using Google Glass.
Feng, Steve; Caire, Romain; Cortazar, Bingen; Turan, Mehmet; Wong, Andrew; Ozcan, Aydogan
2014-03-25
We demonstrate a Google Glass-based rapid diagnostic test (RDT) reader platform capable of qualitative and quantitative measurements of various lateral flow immunochromatographic assays and similar biomedical diagnostics tests. Using a custom-written Glass application and without any external hardware attachments, one or more RDTs labeled with Quick Response (QR) code identifiers are simultaneously imaged using the built-in camera of the Google Glass that is based on a hands-free and voice-controlled interface and digitally transmitted to a server for digital processing. The acquired JPEG images are automatically processed to locate all the RDTs and, for each RDT, to produce a quantitative diagnostic result, which is returned to the Google Glass (i.e., the user) and also stored on a central server along with the RDT image, QR code, and other related information (e.g., demographic data). The same server also provides a dynamic spatiotemporal map and real-time statistics for uploaded RDT results accessible through Internet browsers. We tested this Google Glass-based diagnostic platform using qualitative (i.e., yes/no) human immunodeficiency virus (HIV) and quantitative prostate-specific antigen (PSA) tests. For the quantitative RDTs, we measured activated tests at various concentrations ranging from 0 to 200 ng/mL for free and total PSA. This wearable RDT reader platform running on Google Glass combines a hands-free sensing and image capture interface with powerful servers running our custom image processing codes, and it can be quite useful for real-time spatiotemporal tracking of various diseases and personal medical conditions, providing a valuable tool for epidemiology and mobile health.
Immunochromatographic Diagnostic Test Analysis Using Google Glass
2014-01-01
We demonstrate a Google Glass-based rapid diagnostic test (RDT) reader platform capable of qualitative and quantitative measurements of various lateral flow immunochromatographic assays and similar biomedical diagnostics tests. Using a custom-written Glass application and without any external hardware attachments, one or more RDTs labeled with Quick Response (QR) code identifiers are simultaneously imaged using the built-in camera of the Google Glass that is based on a hands-free and voice-controlled interface and digitally transmitted to a server for digital processing. The acquired JPEG images are automatically processed to locate all the RDTs and, for each RDT, to produce a quantitative diagnostic result, which is returned to the Google Glass (i.e., the user) and also stored on a central server along with the RDT image, QR code, and other related information (e.g., demographic data). The same server also provides a dynamic spatiotemporal map and real-time statistics for uploaded RDT results accessible through Internet browsers. We tested this Google Glass-based diagnostic platform using qualitative (i.e., yes/no) human immunodeficiency virus (HIV) and quantitative prostate-specific antigen (PSA) tests. For the quantitative RDTs, we measured activated tests at various concentrations ranging from 0 to 200 ng/mL for free and total PSA. This wearable RDT reader platform running on Google Glass combines a hands-free sensing and image capture interface with powerful servers running our custom image processing codes, and it can be quite useful for real-time spatiotemporal tracking of various diseases and personal medical conditions, providing a valuable tool for epidemiology and mobile health. PMID:24571349
ERIC Educational Resources Information Center
Jacobsen, Mikael
2008-01-01
Librarians use online mapping services such as Google Maps, MapQuest, Yahoo Maps, and others to check traffic conditions, find local businesses, and provide directions. However, few libraries are using one of Google Maps most outstanding applications, My Maps, for the creation of enhanced and interactive multimedia maps. My Maps is a simple and…
NASA Astrophysics Data System (ADS)
Dimitrova, L. L.; Haines, M.; Holt, W. E.; Schultz, R. A.; Richard, G.; Haines, A. J.
2006-12-01
Interactive maps of surface-breaking faults and stress models on Mars provide important tools to engage undergraduate students, educators, and scientists with current geological and geophysical research. We have developed a map based on the Google Maps API -- an Internet based tool combining DHTML and AJAX, -- which allows very large maps to be viewed over the World Wide Web. Typically, small portions of the maps are downloaded as needed, rather than the entire image at once. This set-up enables relatively fast access for users with low bandwidth. Furthermore, Google Maps provides an extensible interactive interface making it ideal for visualizing multiple data sets at the user's choice. The Google Maps API works primarily with data referenced to latitudes and longitudes, which is then mapped in Mercator projection only. We have developed utilities for general cylindrical coordinate systems by converting these coordinates into equivalent Mercator projection before including them on the map. The MARTIAN project is available at http://rock.geo.sunysb.edu/~holt/Mars/MARTIAN/. We begin with an introduction to the Martian surface using a topography model. Faults from several datasets are classified by type (extension vs. compression) and by time epoch. Deviatoric stresses due to gravitational potential energy differences, calculated from the topography and crustal thickness, can be overlain. Several quantitative measures for the fit of the stress field to the faults are also included. We provide introductory text and exercises spanning a range of topics: how are faults identified, what stress is and how it relates to faults, what gravitational potential energy is and how variations in it produce stress, how the models are created, and how these models can be evaluated and interpreted. The MARTIAN tool is used at Stony Brook University in GEO 310: Introduction to Geophysics, a class geared towards junior and senior geosciences majors. Although this project is in its early stages, high school and college teachers, as well as researchers have expressed interest in using and extending these tools for visualizing and interacting with data on Earth and other planetary bodies.
Web GIS in practice X: a Microsoft Kinect natural user interface for Google Earth navigation
2011-01-01
This paper covers the use of depth sensors such as Microsoft Kinect and ASUS Xtion to provide a natural user interface (NUI) for controlling 3-D (three-dimensional) virtual globes such as Google Earth (including its Street View mode), Bing Maps 3D, and NASA World Wind. The paper introduces the Microsoft Kinect device, briefly describing how it works (the underlying technology by PrimeSense), as well as its market uptake and application potential beyond its original intended purpose as a home entertainment and video game controller. The different software drivers available for connecting the Kinect device to a PC (Personal Computer) are also covered, and their comparative pros and cons briefly discussed. We survey a number of approaches and application examples for controlling 3-D virtual globes using the Kinect sensor, then describe Kinoogle, a Kinect interface for natural interaction with Google Earth, developed by students at Texas A&M University. Readers interested in trying out the application on their own hardware can download a Zip archive (included with the manuscript as additional files 1, 2, &3) that contains a 'Kinnogle installation package for Windows PCs'. Finally, we discuss some usability aspects of Kinoogle and similar NUIs for controlling 3-D virtual globes (including possible future improvements), and propose a number of unique, practical 'use scenarios' where such NUIs could prove useful in navigating a 3-D virtual globe, compared to conventional mouse/3-D mouse and keyboard-based interfaces. PMID:21791054
Web GIS in practice X: a Microsoft Kinect natural user interface for Google Earth navigation.
Boulos, Maged N Kamel; Blanchard, Bryan J; Walker, Cory; Montero, Julio; Tripathy, Aalap; Gutierrez-Osuna, Ricardo
2011-07-26
This paper covers the use of depth sensors such as Microsoft Kinect and ASUS Xtion to provide a natural user interface (NUI) for controlling 3-D (three-dimensional) virtual globes such as Google Earth (including its Street View mode), Bing Maps 3D, and NASA World Wind. The paper introduces the Microsoft Kinect device, briefly describing how it works (the underlying technology by PrimeSense), as well as its market uptake and application potential beyond its original intended purpose as a home entertainment and video game controller. The different software drivers available for connecting the Kinect device to a PC (Personal Computer) are also covered, and their comparative pros and cons briefly discussed. We survey a number of approaches and application examples for controlling 3-D virtual globes using the Kinect sensor, then describe Kinoogle, a Kinect interface for natural interaction with Google Earth, developed by students at Texas A&M University. Readers interested in trying out the application on their own hardware can download a Zip archive (included with the manuscript as additional files 1, 2, &3) that contains a 'Kinnogle installation package for Windows PCs'. Finally, we discuss some usability aspects of Kinoogle and similar NUIs for controlling 3-D virtual globes (including possible future improvements), and propose a number of unique, practical 'use scenarios' where such NUIs could prove useful in navigating a 3-D virtual globe, compared to conventional mouse/3-D mouse and keyboard-based interfaces.
Displaying R spatial statistics on Google dynamic maps with web applications created by Rwui
2012-01-01
Background The R project includes a large variety of packages designed for spatial statistics. Google dynamic maps provide web based access to global maps and satellite imagery. We describe a method for displaying directly the spatial output from an R script on to a Google dynamic map. Methods This is achieved by creating a Java based web application which runs the R script and then displays the results on the dynamic map. In order to make this method easy to implement by those unfamiliar with programming Java based web applications, we have added the method to the options available in the R Web User Interface (Rwui) application. Rwui is an established web application for creating web applications for running R scripts. A feature of Rwui is that all the code for the web application being created is generated automatically so that someone with no knowledge of web programming can make a fully functional web application for running an R script in a matter of minutes. Results Rwui can now be used to create web applications that will display the results from an R script on a Google dynamic map. Results may be displayed as discrete markers and/or as continuous overlays. In addition, users of the web application may select regions of interest on the dynamic map with mouse clicks and the coordinates of the region of interest will automatically be made available for use by the R script. Conclusions This method of displaying R output on dynamic maps is designed to be of use in a number of areas. Firstly it allows statisticians, working in R and developing methods in spatial statistics, to easily visualise the results of applying their methods to real world data. Secondly, it allows researchers who are using R to study health geographics data, to display their results directly onto dynamic maps. Thirdly, by creating a web application for running an R script, a statistician can enable users entirely unfamiliar with R to run R coded statistical analyses of health geographics data. Fourthly, we envisage an educational role for such applications. PMID:22998945
Displaying R spatial statistics on Google dynamic maps with web applications created by Rwui.
Newton, Richard; Deonarine, Andrew; Wernisch, Lorenz
2012-09-24
The R project includes a large variety of packages designed for spatial statistics. Google dynamic maps provide web based access to global maps and satellite imagery. We describe a method for displaying directly the spatial output from an R script on to a Google dynamic map. This is achieved by creating a Java based web application which runs the R script and then displays the results on the dynamic map. In order to make this method easy to implement by those unfamiliar with programming Java based web applications, we have added the method to the options available in the R Web User Interface (Rwui) application. Rwui is an established web application for creating web applications for running R scripts. A feature of Rwui is that all the code for the web application being created is generated automatically so that someone with no knowledge of web programming can make a fully functional web application for running an R script in a matter of minutes. Rwui can now be used to create web applications that will display the results from an R script on a Google dynamic map. Results may be displayed as discrete markers and/or as continuous overlays. In addition, users of the web application may select regions of interest on the dynamic map with mouse clicks and the coordinates of the region of interest will automatically be made available for use by the R script. This method of displaying R output on dynamic maps is designed to be of use in a number of areas. Firstly it allows statisticians, working in R and developing methods in spatial statistics, to easily visualise the results of applying their methods to real world data. Secondly, it allows researchers who are using R to study health geographics data, to display their results directly onto dynamic maps. Thirdly, by creating a web application for running an R script, a statistician can enable users entirely unfamiliar with R to run R coded statistical analyses of health geographics data. Fourthly, we envisage an educational role for such applications.
Using Google Earth as an innovative tool for community mapping.
Lefer, Theodore B; Anderson, Matthew R; Fornari, Alice; Lambert, Anastasia; Fletcher, Jason; Baquero, Maria
2008-01-01
Maps are used to track diseases and illustrate the social context of health problems. However, commercial mapping software requires special training. This article illustrates how nonspecialists used Google Earth, a free program, to create community maps. The Bronx, New York, is characterized by high levels of obesity and diabetes. Residents and medical students measured the variety and quality of food and exercise sources around a residency training clinic and a student-run free clinic, using Google Earth to create maps with minimal assistance. Locations were identified using street addresses or simply by pointing to them on a map. Maps can be shared via e-mail, viewed online with Google Earth or Google Maps, and the data can be incorporated into other mapping software.
NASA Astrophysics Data System (ADS)
Triantafyllou, Antoine; Bastin, Christophe; Watlet, Arnaud
2016-04-01
GIS software suites are today's essential tools to gather and visualise geological data, to apply spatial and temporal analysis and in fine, to create and share interactive maps for further geosciences' investigations. For these purposes, we developed GeolOkit: an open-source, freeware and lightweight software, written in Python, a high-level, cross-platform programming language. GeolOkit software is accessible through a graphical user interface, designed to run in parallel with Google Earth. It is a super user-friendly toolbox that allows 'geo-users' to import their raw data (e.g. GPS, sample locations, structural data, field pictures, maps), to use fast data analysis tools and to plot these one into Google Earth environment using KML code. This workflow requires no need of any third party software, except Google Earth itself. GeolOkit comes with large number of geosciences' labels, symbols, colours and placemarks and may process : (i) multi-points data, (ii) contours via several interpolations methods, (iii) discrete planar and linear structural data in 2D or 3D supporting large range of structures input format, (iv) clustered stereonets and rose diagram, (v) drawn cross-sections as vertical sections, (vi) georeferenced maps and vectors, (vii) field pictures using either geo-tracking metadata from a camera built-in GPS module, or the same-day track of an external GPS. We are looking for you to discover all the functionalities of GeolOkit software. As this project is under development, we are definitely looking to discussions regarding your proper needs, your ideas and contributions to GeolOkit project.
Railroad track inspection interface demonstration : final report.
DOT National Transportation Integrated Search
2016-01-01
This project developed a track data user interface utilizing the Google Glass optical display device. The interface allows the user : to recall data stored remotely and view the data on the Google Glass. The technical effort required developing a com...
Boeker, Martin; Vach, Werner; Motschall, Edith
2013-10-26
Recent research indicates a high recall in Google Scholar searches for systematic reviews. These reports raised high expectations of Google Scholar as a unified and easy to use search interface. However, studies on the coverage of Google Scholar rarely used the search interface in a realistic approach but instead merely checked for the existence of gold standard references. In addition, the severe limitations of the Google Search interface must be taken into consideration when comparing with professional literature retrieval tools.The objectives of this work are to measure the relative recall and precision of searches with Google Scholar under conditions which are derived from structured search procedures conventional in scientific literature retrieval; and to provide an overview of current advantages and disadvantages of the Google Scholar search interface in scientific literature retrieval. General and MEDLINE-specific search strategies were retrieved from 14 Cochrane systematic reviews. Cochrane systematic review search strategies were translated to Google Scholar search expression as good as possible under consideration of the original search semantics. The references of the included studies from the Cochrane reviews were checked for their inclusion in the result sets of the Google Scholar searches. Relative recall and precision were calculated. We investigated Cochrane reviews with a number of included references between 11 and 70 with a total of 396 references. The Google Scholar searches resulted in sets between 4,320 and 67,800 and a total of 291,190 hits. The relative recall of the Google Scholar searches had a minimum of 76.2% and a maximum of 100% (7 searches). The precision of the Google Scholar searches had a minimum of 0.05% and a maximum of 0.92%. The overall relative recall for all searches was 92.9%, the overall precision was 0.13%. The reported relative recall must be interpreted with care. It is a quality indicator of Google Scholar confined to an experimental setting which is unavailable in systematic retrieval due to the severe limitations of the Google Scholar search interface. Currently, Google Scholar does not provide necessary elements for systematic scientific literature retrieval such as tools for incremental query optimization, export of a large number of references, a visual search builder or a history function. Google Scholar is not ready as a professional searching tool for tasks where structured retrieval methodology is necessary.
2013-01-01
Background Recent research indicates a high recall in Google Scholar searches for systematic reviews. These reports raised high expectations of Google Scholar as a unified and easy to use search interface. However, studies on the coverage of Google Scholar rarely used the search interface in a realistic approach but instead merely checked for the existence of gold standard references. In addition, the severe limitations of the Google Search interface must be taken into consideration when comparing with professional literature retrieval tools. The objectives of this work are to measure the relative recall and precision of searches with Google Scholar under conditions which are derived from structured search procedures conventional in scientific literature retrieval; and to provide an overview of current advantages and disadvantages of the Google Scholar search interface in scientific literature retrieval. Methods General and MEDLINE-specific search strategies were retrieved from 14 Cochrane systematic reviews. Cochrane systematic review search strategies were translated to Google Scholar search expression as good as possible under consideration of the original search semantics. The references of the included studies from the Cochrane reviews were checked for their inclusion in the result sets of the Google Scholar searches. Relative recall and precision were calculated. Results We investigated Cochrane reviews with a number of included references between 11 and 70 with a total of 396 references. The Google Scholar searches resulted in sets between 4,320 and 67,800 and a total of 291,190 hits. The relative recall of the Google Scholar searches had a minimum of 76.2% and a maximum of 100% (7 searches). The precision of the Google Scholar searches had a minimum of 0.05% and a maximum of 0.92%. The overall relative recall for all searches was 92.9%, the overall precision was 0.13%. Conclusion The reported relative recall must be interpreted with care. It is a quality indicator of Google Scholar confined to an experimental setting which is unavailable in systematic retrieval due to the severe limitations of the Google Scholar search interface. Currently, Google Scholar does not provide necessary elements for systematic scientific literature retrieval such as tools for incremental query optimization, export of a large number of references, a visual search builder or a history function. Google Scholar is not ready as a professional searching tool for tasks where structured retrieval methodology is necessary. PMID:24160679
Integrating Radar Image Data with Google Maps
NASA Technical Reports Server (NTRS)
Chapman, Bruce D.; Gibas, Sarah
2010-01-01
A public Web site has been developed as a method for displaying the multitude of radar imagery collected by NASA s Airborne Synthetic Aperture Radar (AIRSAR) instrument during its 16-year mission. Utilizing NASA s internal AIRSAR site, the new Web site features more sophisticated visualization tools that enable the general public to have access to these images. The site was originally maintained at NASA on six computers: one that held the Oracle database, two that took care of the software for the interactive map, and three that were for the Web site itself. Several tasks were involved in moving this complicated setup to just one computer. First, the AIRSAR database was migrated from Oracle to MySQL. Then the back-end of the AIRSAR Web site was updated in order to access the MySQL database. To do this, a few of the scripts needed to be modified; specifically three Perl scripts that query that database. The database connections were then updated from Oracle to MySQL, numerous syntax errors were corrected, and a query was implemented that replaced one of the stored Oracle procedures. Lastly, the interactive map was designed, implemented, and tested so that users could easily browse and access the radar imagery through the Google Maps interface.
In Pursuit of Agile Acquisition: Are We There Yet?
2013-03-01
digital mapping capabilities like Google , 71Microsoft,72 and Wikimapia,73 are readily obtainable in the commercial marketplace. This knowledge...Fox. Defense Acquisition Reform, 14. 69 Ibid., 8. 70 XBRADTC, “Army Acquisition Woes,” Bring the Heat Bring the Stupid , entry posted May 1, 2011...https://xbradtc.wordpress.com/2011/05/01/Army-acquisition-woes/ (accessed on December 5, 2012). 71 Google Maps, http://maps.google.com/maps (accessed
ANTP Protocol Suite Software Implementation Architecture in Python
2011-06-03
a popular platform of networking programming, an area in which C has traditionally dominated. 2 NetController AeroRP AeroNP AeroNP API AeroTP...visualisation of the running system. For example using the Google Maps API , the main logging web page can show all the running nodes in the system. By...communication between AeroNP and AeroRP and runs on the operating system as daemon. Furthermore, it creates an API interface to mange the communication between
Visualizing Moon Data and Imagery with Google Earth
NASA Astrophysics Data System (ADS)
Weiss-Malik, M.; Scharff, T.; Nefian, A.; Moratto, Z.; Kolb, E.; Lundy, M.; Hancher, M.; Gorelick, N.; Broxton, M.; Beyer, R. A.
2009-12-01
There is a vast store of planetary geospatial data that has been collected by NASA but is difficult to access and visualize. Virtual globes have revolutionized the way we visualize and understand the Earth, but other planetary bodies including Mars and the Moon can be visualized in similar ways. Extraterrestrial virtual globes are poised to revolutionize planetary science, bring an exciting new dimension to science education, and allow ordinary users to explore imagery being sent back to Earth by planetary science satellites. The original Google Moon Web site was a limited series of maps and Apollo content. The new Moon in Google Earth feature provides a similar virtual planet experience for the Moon as we have for the Earth and Mars. We incorporated existing Clementine and Lunar Orbiter imagery for the basemaps and a combination of Kaguya LALT topography and some terrain created from Apollo Metric and Panoramic images. We also have information about the Apollo landings and other robotic landers on the surface, as well as historic maps and charts, and guided tours. Some of the first-released LROC imagery of the Apollo landing sites has been put in place, and we look forward to incorporating more data as it is released from LRO, Chandraayan-1, and Kaguya. These capabilities have obvious public outreach and education benefits, but the potential benefits of allowing planetary scientists to rapidly explore these large and varied data collections — in geological context and within a single user interface — are also becoming evident. Because anyone can produce additional KML content for use in Google Earth, scientists can customize the environment to their needs as well as publish their own processed data and results for others to use. Many scientists and organizations have begun to do this already, resulting in a useful and growing collection of planetary-science-oriented Google Earth layers. Screen shot of Moon in Google Earth, a freely downloadable application for visualizing Moon imagery and data.
Online Public Access Catalog: The Google Maps of the Library World
ERIC Educational Resources Information Center
Bailey, Kieren
2011-01-01
What do Google Maps and a library's Online Public Access Catalog (OPAC) have in common? Google Maps provides users with all the information they need for a trip in one place; users can get directions and find out what attractions, hotels, and restaurants are close by. Librarians must find the ultimate OPAC that will provide, in one place, all the…
USGS Coastal and Marine Geology Survey Data in Google Earth
NASA Astrophysics Data System (ADS)
Reiss, C.; Steele, C.; Ma, A.; Chin, J.
2006-12-01
The U.S. Geological Survey (USGS) Coastal and Marine Geology (CMG) program has a rich data catalog of geologic field activities and metadata called InfoBank, which has been a standard tool for researchers within and outside of the agency. Along with traditional web maps, the data are now accessible in Google Earth, which greatly expands the possible user audience. The Google Earth interface provides geographic orientation and panning/zooming capabilities to locate data relative to topography, bathymetry, and coastal areas. Viewing navigation with Google Earth's background imagery allows queries such as, why areas were not surveyed (answer presence of islands, shorelines, cliffs, etc.). Detailed box core subsample photos from selected sampling activities, published geotechnical data, and sample descriptions are now viewable on Google Earth, (for example, M-1-95-MB, P-2-95-MB, and P-1-97- MB box core samples). One example of the use of Google Earth is CMG's surveys of San Francisco's Ocean Beach since 2004. The surveys are conducted with an all-terrain vehicle (ATV) and shallow-water personal watercraft (PWC) equipped with Global Positioning System (GPS), and elevation and echo sounder data collectors. 3D topographic models with centimeter accuracy have been produced from these surveys to monitor beach and nearshore processes, including sand transport, sedimentation patterns, and seasonal trends. Using Google Earth, multiple track line data (examples: OB-1-05-CA and OB-2-05-CA) can be overlaid on beach imagery. The images also help explain the shape of track lines as objects are encountered.
Googling trends in conservation biology.
Proulx, Raphaël; Massicotte, Philippe; Pépino, Marc
2014-02-01
Web-crawling approaches, that is, automated programs data mining the internet to obtain information about a particular process, have recently been proposed for monitoring early signs of ecosystem degradation or for establishing crop calendars. However, lack of a clear conceptual and methodological framework has prevented the development of such approaches within the field of conservation biology. Our objective was to illustrate how Google Trends, a freely accessible web-crawling engine, can be used to track changes in timing of biological processes, spatial distribution of invasive species, and level of public awareness about key conservation issues. Google Trends returns the number of internet searches that were made for a keyword in a given region of the world over a defined period. Using data retrieved online for 13 countries, we exemplify how Google Trends can be used to study the timing of biological processes, such as the seasonal recurrence of pollen release or mosquito outbreaks across a latitudinal gradient. We mapped the spatial extent of results from Google Trends for 5 invasive species in the United States and found geographic patterns in invasions that are consistent with their coarse-grained distribution at state levels. From 2004 through 2012, Google Trends showed that the level of public interest and awareness about conservation issues related to ecosystem services, biodiversity, and climate change increased, decreased, and followed both trends, respectively. Finally, to further the development of research approaches at the interface of conservation biology, collective knowledge, and environmental management, we developed an algorithm that allows the rapid retrieval of Google Trends data. © 2013 Society for Conservation Biology.
NASA Astrophysics Data System (ADS)
Raup, B. H.; Khalsa, S. S.; Armstrong, R.
2007-12-01
The Global Land Ice Measurements from Space (GLIMS) project has built a geospatial and temporal database of glacier data, composed of glacier outlines and various scalar attributes. These data are being derived primarily from satellite imagery, such as from ASTER and Landsat. Each "snapshot" of a glacier is from a specific time, and the database is designed to store multiple snapshots representative of different times. We have implemented two web-based interfaces to the database; one enables exploration of the data via interactive maps (web map server), while the other allows searches based on text-field constraints. The web map server is an Open Geospatial Consortium (OGC) compliant Web Map Server (WMS) and Web Feature Server (WFS). This means that other web sites can display glacier layers from our site over the Internet, or retrieve glacier features in vector format. All components of the system are implemented using Open Source software: Linux, PostgreSQL, PostGIS (geospatial extensions to the database), MapServer (WMS and WFS), and several supporting components such as Proj.4 (a geographic projection library) and PHP. These tools are robust and provide a flexible and powerful framework for web mapping applications. As a service to the GLIMS community, the database contains metadata on all ASTER imagery acquired over glacierized terrain. Reduced-resolution of the images (browse imagery) can be viewed either as a layer in the MapServer application, or overlaid on the virtual globe within Google Earth. The interactive map application allows the user to constrain by time what data appear on the map. For example, ASTER or glacier outlines from 2002 only, or from Autumn in any year, can be displayed. The system allows users to download their selected glacier data in a choice of formats. The results of a query based on spatial selection (using a mouse) or text-field constraints can be downloaded in any of these formats: ESRI shapefiles, KML (Google Earth), MapInfo, GML (Geography Markup Language) and GMT (Generic Mapping Tools). This "clip-and-ship" function allows users to download only the data they are interested in. Our flexible web interfaces to the database, which includes various support layers (e.g. a layer to help collaborators identify satellite imagery over their region of expertise) will facilitate enhanced analysis to be undertaken on glacier systems, their distribution, and their impacts on other Earth systems.
NASA Technical Reports Server (NTRS)
Flores, Sarah L.; Chapman, Bruce D.; Tung, Waye W.; Zheng, Yang
2011-01-01
This new interface will enable Principal Investigators (PIs), as well as UAVSAR (Uninhabited Aerial Vehicle Synthetic Aperture Radar) members to do their own flight planning and time estimation without having to request flight lines through the science coordinator. It uses an all-in-one Google Maps interface, a JPL hosted database, and PI flight requirements to design an airborne flight plan. The application will enable users to see their own flight plan being constructed interactively through a map interface, and then the flight planning software will generate all the files necessary for the flight. Afterward, the UAVSAR team can then complete the flight request, including calendaring and supplying requisite flight request files in the expected format for processing by NASA s airborne science program. Some of the main features of the interface include drawing flight lines on the map, nudging them, adding them to the current flight plan, and reordering them. The user can also search and select takeoff, landing, and intermediate airports. As the flight plan is constructed, all of its components are constantly being saved to the database, and the estimated flight times are updated. Another feature is the ability to import flight lines from previously saved flight plans. One of the main motivations was to make this Web application as simple and intuitive as possible, while also being dynamic and robust. This Web application can easily be extended to support other airborne instruments.
Keemei: cloud-based validation of tabular bioinformatics file formats in Google Sheets.
Rideout, Jai Ram; Chase, John H; Bolyen, Evan; Ackermann, Gail; González, Antonio; Knight, Rob; Caporaso, J Gregory
2016-06-13
Bioinformatics software often requires human-generated tabular text files as input and has specific requirements for how those data are formatted. Users frequently manage these data in spreadsheet programs, which is convenient for researchers who are compiling the requisite information because the spreadsheet programs can easily be used on different platforms including laptops and tablets, and because they provide a familiar interface. It is increasingly common for many different researchers to be involved in compiling these data, including study coordinators, clinicians, lab technicians and bioinformaticians. As a result, many research groups are shifting toward using cloud-based spreadsheet programs, such as Google Sheets, which support the concurrent editing of a single spreadsheet by different users working on different platforms. Most of the researchers who enter data are not familiar with the formatting requirements of the bioinformatics programs that will be used, so validating and correcting file formats is often a bottleneck prior to beginning bioinformatics analysis. We present Keemei, a Google Sheets Add-on, for validating tabular files used in bioinformatics analyses. Keemei is available free of charge from Google's Chrome Web Store. Keemei can be installed and run on any web browser supported by Google Sheets. Keemei currently supports the validation of two widely used tabular bioinformatics formats, the Quantitative Insights into Microbial Ecology (QIIME) sample metadata mapping file format and the Spatially Referenced Genetic Data (SRGD) format, but is designed to easily support the addition of others. Keemei will save researchers time and frustration by providing a convenient interface for tabular bioinformatics file format validation. By allowing everyone involved with data entry for a project to easily validate their data, it will reduce the validation and formatting bottlenecks that are commonly encountered when human-generated data files are first used with a bioinformatics system. Simplifying the validation of essential tabular data files, such as sample metadata, will reduce common errors and thereby improve the quality and reliability of research outcomes.
GIS tool to locate major Sikh temples in USA
NASA Astrophysics Data System (ADS)
Sharma, Saumya
This tool is a GIS based interactive and graphical user interface tool, which locates the major Sikh temples of USA on a map. This tool is using Java programming language along with MOJO (Map Object Java Object) provided by ESRI that is the organization that provides the GIS software. It also includes some of the integration with Google's API's like Google Translator API. This application will tell users about the origin of Sikhism in India and USA, the major Sikh temples in each state of USA, location, name and detail information through their website. The primary purpose of this application is to make people aware about this religion and culture. This tool will also measure the distance between two temple points in a map and display the result in miles and kilometers. Also, there is an added support to convert each temple's website language from English to Punjabi or any other language using a language convertor tool so that people from different nationalities can understand their culture. By clicking on each point on a map, a new window will pop up showing the picture of the temple and a hyperlink that will redirect to the website of that particular temple .It will also contain links to their dance, music, history, and also a help menu to guide the users to use the software efficiently.
2013-01-01
Background Molecular biology knowledge can be formalized and systematically represented in a computer-readable form as a comprehensive map of molecular interactions. There exist an increasing number of maps of molecular interactions containing detailed and step-wise description of various cell mechanisms. It is difficult to explore these large maps, to organize discussion of their content and to maintain them. Several efforts were recently made to combine these capabilities together in one environment, and NaviCell is one of them. Results NaviCell is a web-based environment for exploiting large maps of molecular interactions, created in CellDesigner, allowing their easy exploration, curation and maintenance. It is characterized by a combination of three essential features: (1) efficient map browsing based on Google Maps; (2) semantic zooming for viewing different levels of details or of abstraction of the map and (3) integrated web-based blog for collecting community feedback. NaviCell can be easily used by experts in the field of molecular biology for studying molecular entities of interest in the context of signaling pathways and crosstalk between pathways within a global signaling network. NaviCell allows both exploration of detailed molecular mechanisms represented on the map and a more abstract view of the map up to a top-level modular representation. NaviCell greatly facilitates curation, maintenance and updating the comprehensive maps of molecular interactions in an interactive and user-friendly fashion due to an imbedded blogging system. Conclusions NaviCell provides user-friendly exploration of large-scale maps of molecular interactions, thanks to Google Maps and WordPress interfaces, with which many users are already familiar. Semantic zooming which is used for navigating geographical maps is adopted for molecular maps in NaviCell, making any level of visualization readable. In addition, NaviCell provides a framework for community-based curation of maps. PMID:24099179
Kuperstein, Inna; Cohen, David P A; Pook, Stuart; Viara, Eric; Calzone, Laurence; Barillot, Emmanuel; Zinovyev, Andrei
2013-10-07
Molecular biology knowledge can be formalized and systematically represented in a computer-readable form as a comprehensive map of molecular interactions. There exist an increasing number of maps of molecular interactions containing detailed and step-wise description of various cell mechanisms. It is difficult to explore these large maps, to organize discussion of their content and to maintain them. Several efforts were recently made to combine these capabilities together in one environment, and NaviCell is one of them. NaviCell is a web-based environment for exploiting large maps of molecular interactions, created in CellDesigner, allowing their easy exploration, curation and maintenance. It is characterized by a combination of three essential features: (1) efficient map browsing based on Google Maps; (2) semantic zooming for viewing different levels of details or of abstraction of the map and (3) integrated web-based blog for collecting community feedback. NaviCell can be easily used by experts in the field of molecular biology for studying molecular entities of interest in the context of signaling pathways and crosstalk between pathways within a global signaling network. NaviCell allows both exploration of detailed molecular mechanisms represented on the map and a more abstract view of the map up to a top-level modular representation. NaviCell greatly facilitates curation, maintenance and updating the comprehensive maps of molecular interactions in an interactive and user-friendly fashion due to an imbedded blogging system. NaviCell provides user-friendly exploration of large-scale maps of molecular interactions, thanks to Google Maps and WordPress interfaces, with which many users are already familiar. Semantic zooming which is used for navigating geographical maps is adopted for molecular maps in NaviCell, making any level of visualization readable. In addition, NaviCell provides a framework for community-based curation of maps.
Geolokit: An interactive tool for visualising and exploring geoscientific data in Google Earth
NASA Astrophysics Data System (ADS)
Triantafyllou, Antoine; Watlet, Arnaud; Bastin, Christophe
2017-10-01
Virtual globes have been developed to showcase different types of data combining a digital elevation model and basemaps of high resolution satellite imagery. Hence, they became a standard to share spatial data and information, although they suffer from a lack of toolboxes dedicated to the formatting of large geoscientific dataset. From this perspective, we developed Geolokit: a free and lightweight software that allows geoscientists - and every scientist working with spatial data - to import their data (e.g., sample collections, structural geology, cross-sections, field pictures, georeferenced maps), to handle and to transcribe them to Keyhole Markup Language (KML) files. KML files are then automatically opened in the Google Earth virtual globe and the spatial data accessed and shared. Geolokit comes with a large number of dedicated tools that can process and display: (i) multi-points data, (ii) scattered data interpolations, (iii) structural geology features in 2D and 3D, (iv) rose diagrams, stereonets and dip-plunge polar histograms, (v) cross-sections and oriented rasters, (vi) georeferenced field pictures, (vii) georeferenced maps and projected gridding. Therefore, together with Geolokit, Google Earth becomes not only a powerful georeferenced data viewer but also a stand-alone work platform. The toolbox (available online at http://www.geolokit.org) is written in Python, a high-level, cross-platform programming language and is accessible through a graphical user interface, designed to run in parallel with Google Earth, through a workflow that requires no additional third party software. Geolokit features are demonstrated in this paper using typical datasets gathered from two case studies illustrating its applicability at multiple scales of investigation: a petro-structural investigation of the Ile d'Yeu orthogneissic unit (Western France) and data collection of the Mariana oceanic subduction zone (Western Pacific).
SpaceTime Environmental Image Information for Scene Understanding
2016-04-01
public Internet resources such as Google,65 MapQuest,66 Bing,67 and Yahoo Maps.68 Approved for public release; distribution unlimited. 9 Table 3...azimuth angle 3 Terrain and location: USACE AGC — Satellite/aerial imagery and terrain analysis 4 Terrain and location: Google, MapQuest, Bing, Yahoo ...Maps. [accessed 2015 Dec]. https://www.bing.com/maps/. 68. YAHOO ! Maps. [accessed 2015 Dec]. https://maps.yahoo.com/b/. 69. 557th Weather Wing. US
A web-based screening tool for near-port air quality assessments
Isakov, Vlad; Barzyk, Timothy M.; Smith, Elizabeth R.; Arunachalam, Saravanan; Naess, Brian; Venkatram, Akula
2018-01-01
The Community model for near-PORT applications (C-PORT) is a screening tool with an intended purpose of calculating differences in annual averaged concentration patterns and relative contributions of various source categories over the spatial domain within about 10 km of the port. C-PORT can inform decision-makers and concerned citizens about local air quality due to mobile source emissions related to commercial port activities. It allows users to visualize and evaluate different planning scenarios, helping them identify the best alternatives for making long-term decisions that protect community health and sustainability. The web-based, easy-to-use interface currently includes data from 21 seaports primarily in the Southeastern U.S., and has a map-based interface based on Google Maps. The tool was developed to visualize and assess changes in air quality due to changes in emissions and/or meteorology in order to analyze development scenarios, and is not intended to support or replace any regulatory models or programs. PMID:29681760
Google Maps offers a new way to evaluate claudication.
Khambati, Husain; Boles, Kim; Jetty, Prasad
2017-05-01
Accurate determination of walking capacity is important for the clinical diagnosis and management plan for patients with peripheral arterial disease. The current "gold standard" of measurement is walking distance on a treadmill. However, treadmill testing is not always reflective of the patient's natural walking conditions, and it may not be fully accessible in every vascular clinic. The objective of this study was to determine whether Google Maps, the readily available GPS-based mapping tool, offers an accurate and accessible method of evaluating walking distances in vascular claudication patients. Patients presenting to the outpatient vascular surgery clinic between November 2013 and April 2014 at the Ottawa Hospital with vasculogenic calf, buttock, and thigh claudication symptoms were identified and prospectively enrolled in our study. Onset of claudication symptoms and maximal walking distance (MWD) were evaluated using four tools: history; Walking Impairment Questionnaire (WIQ), a validated claudication survey; Google Maps distance calculator (patients were asked to report their daily walking routes on the Google Maps-based tool runningmap.com, and walking distances were calculated accordingly); and treadmill testing for onset of symptoms and MWD, recorded in a double-blinded fashion. Fifteen patients were recruited for the study. Determination of walking distances using Google Maps proved to be more accurate than by both clinical history and WIQ, correlating highly with the gold standard of treadmill testing for both claudication onset (r = .805; P < .001) and MWD (r = .928; P < .0001). In addition, distances were generally under-reported on history and WIQ. The Google Maps tool was also efficient, with reporting times averaging below 4 minutes. For vascular claudicants with no other walking limitations, Google Maps is a promising new tool that combines the objective strengths of the treadmill test and incorporates real-world walking environments. It offers an accurate, efficient, inexpensive, and readily accessible way to assess walking distances in patients with peripheral vascular disease. Copyright © 2017 Society for Vascular Surgery. Published by Elsevier Inc. All rights reserved.
Google Earth and Geo Applications: A Toolset for Viewing Earth's Geospatial Information
NASA Astrophysics Data System (ADS)
Tuxen-Bettman, K.
2016-12-01
Earth scientists measure and derive fundamental data that can be of broad general interest to the public and policy makers. Yet, one of the challenges that has always faced the Earth science community is how to present their data and findings in an easy-to-use and compelling manner. Google's Geo Tools offer an efficient and dynamic way for scientists, educators, journalists and others to both access data and view or tell stories in a dynamic three-dimensional geospatial context. Google Earth in particular provides a dense canvas of satellite imagery on which can be viewed rich vector and raster datasets using the medium of Keyhole Markup Language (KML). Through KML, Google Earth can combine the analytical capabilities of Earth Engine, collaborative mapping of My Maps, and storytelling of Tour Builder and more to make Google's Geo Applications a coherent suite of tools for exploring our planet.https://earth.google.com/https://earthengine.google.com/https://mymaps.google.com/https://tourbuilder.withgoogle.com/https://www.google.com/streetview/
Web application and database modeling of traffic impact analysis using Google Maps
NASA Astrophysics Data System (ADS)
Yulianto, Budi; Setiono
2017-06-01
Traffic impact analysis (TIA) is a traffic study that aims at identifying the impact of traffic generated by development or change in land use. In addition to identifying the traffic impact, TIA is also equipped with mitigation measurement to minimize the arising traffic impact. TIA has been increasingly important since it was defined in the act as one of the requirements in the proposal of Building Permit. The act encourages a number of TIA studies in various cities in Indonesia, including Surakarta. For that reason, it is necessary to study the development of TIA by adopting the concept Transportation Impact Control (TIC) in the implementation of the TIA standard document and multimodal modeling. It includes TIA's standardization for technical guidelines, database and inspection by providing TIA checklists, monitoring and evaluation. The research was undertaken by collecting the historical data of junctions, modeling of the data in the form of relational database, building a user interface for CRUD (Create, Read, Update and Delete) the TIA data in the form of web programming with Google Maps libraries. The result research is a system that provides information that helps the improvement and repairment of TIA documents that exist today which is more transparent, reliable and credible.
2013-08-09
CAPE CANAVERAL, Fla. – Google used an assortment of vehicles to precisely map NASA's Kennedy Space Center in Florida to be featured on the company's map page. The work allows Internet users to see inside buildings at Kennedy as they were used during the space shuttle era. Google used a car, tricycle and pushcart to maneuver around the center and through some of its facilities. Photo credit: Google/Wendy Wang
Recent Advances in Geospatial Visualization with the New Google Earth
NASA Astrophysics Data System (ADS)
Anderson, J. C.; Poyart, E.; Yan, S.; Sargent, R.
2017-12-01
Google Earth's detailed, world-wide imagery and terrain data provide a rich backdrop for geospatial visualization at multiple scales, from global to local. The Keyhole Markup Language (KML) is an open standard that has been the primary way for users to author and share data visualizations in Google Earth. Despite its ease of use and flexibility for relatively small amounts of data, users can quickly run into difficulties and limitations working with large-scale or time-varying datasets using KML in Google Earth. Recognizing these challenges, we present our recent work toward extending Google Earth to be a more powerful data visualization platform. We describe a new KML extension to simplify the display of multi-resolution map tile pyramids - which can be created by analysis platforms like Google Earth Engine, or by a variety of other map tile production pipelines. We also describe how this implementation can pave the way to creating novel data visualizations by leveraging custom graphics shaders. Finally, we present our investigations into native support in Google Earth for data storage and transport formats that are well-suited for big raster and vector data visualization. Taken together, these capabilities make it easier to create and share new scientific data visualization experiences using Google Earth, and simplify the integration of Google Earth with existing map data products, services, and analysis pipelines.
Price, Richard; Marsh, Abbie J; Fisher, Marisa H
2018-03-01
Facilitating the use of public transportation enhances opportunities for independent living and competitive, community-based employment for individuals with intellectual and developmental disabilities (IDD). Four young adults with IDD were taught through total-task chaining to use the Google Maps application, a self-prompting, visual navigation system, to take the bus to locations around a college campus and the community. Three of four participants learned to use Google Maps to independently navigate public transportation. Google Maps may be helpful in supporting independent travel, highlighting the importance of future research in teaching navigation skills. Learning to independently use public transportation increases access to autonomous activities, such as opportunities to work and to attend postsecondary education programs on large college campuses.Individuals with IDD can be taught through chaining procedures to use the Google Maps application to navigate public transportation.Mobile map applications are an effective and functional modern tool that can be used to teach community navigation.
ERIC Educational Resources Information Center
Hsu, Hsiao-Ping; Tsai, Bor-Wen; Chen, Che-Ming
2018-01-01
Teaching high-school geomorphological concepts and topographic map reading entails many challenges. This research reports the applicability and effectiveness of Google Earth in teaching topographic map skills and geomorphological concepts, by a single teacher, in a one-computer classroom. Compared to learning via a conventional instructional…
NASA Astrophysics Data System (ADS)
Ivankovic, D.; Dadic, V.
2009-04-01
Some of oceanographic parameters have to be manually inserted into database; some (for example data from CTD probe) are inserted from various files. All this parameters requires visualization, validation and manipulation from research vessel or scientific institution, and also public presentation. For these purposes is developed web based system, containing dynamic sql procedures and java applets. Technology background is Oracle 10g relational database, and Oracle application server. Web interfaces are developed using PL/SQL stored database procedures (mod PL/SQL). Additional parts for data visualization include use of Java applets and JavaScript. Mapping tool is Google maps API (javascript) and as alternative java applet. Graph is realized as dynamically generated web page containing java applet. Mapping tool and graph are georeferenced. That means that click on some part of graph, automatically initiate zoom or marker onto location where parameter was measured. This feature is very useful for data validation. Code for data manipulation and visualization are partially realized with dynamic SQL and that allow as to separate data definition and code for data manipulation. Adding new parameter in system requires only data definition and description without programming interface for this kind of data.
Regional early flood warning system: design and implementation
NASA Astrophysics Data System (ADS)
Chang, L. C.; Yang, S. N.; Kuo, C. L.; Wang, Y. F.
2017-12-01
This study proposes a prototype of the regional early flood inundation warning system in Tainan City, Taiwan. The AI technology is used to forecast multi-step-ahead regional flood inundation maps during storm events. The computing time is only few seconds that leads to real-time regional flood inundation forecasting. A database is built to organize data and information for building real-time forecasting models, maintaining the relations of forecasted points, and displaying forecasted results, while real-time data acquisition is another key task where the model requires immediately accessing rain gauge information to provide forecast services. All programs related database are constructed in Microsoft SQL Server by using Visual C# to extracting real-time hydrological data, managing data, storing the forecasted data and providing the information to the visual map-based display. The regional early flood inundation warning system use the up-to-date Web technologies driven by the database and real-time data acquisition to display the on-line forecasting flood inundation depths in the study area. The friendly interface includes on-line sequentially showing inundation area by Google Map, maximum inundation depth and its location, and providing KMZ file download of the results which can be watched on Google Earth. The developed system can provide all the relevant information and on-line forecast results that helps city authorities to make decisions during typhoon events and make actions to mitigate the losses.
Towards Large-area Field-scale Operational Evapotranspiration for Water Use Mapping
NASA Astrophysics Data System (ADS)
Senay, G. B.; Friedrichs, M.; Morton, C.; Huntington, J. L.; Verdin, J.
2017-12-01
Field-scale evapotranspiration (ET) estimates are needed for improving surface and groundwater use and water budget studies. Ideally, field-scale ET estimates would be at regional to national levels and cover long time periods. As a result of large data storage and computational requirements associated with processing field-scale satellite imagery such as Landsat, numerous challenges remain to develop operational ET estimates over large areas for detailed water use and availability studies. However, the combination of new science, data availability, and cloud computing technology is enabling unprecedented capabilities for ET mapping. To demonstrate this capability, we used Google's Earth Engine cloud computing platform to create nationwide annual ET estimates with 30-meter resolution Landsat ( 16,000 images) and gridded weather data using the Operational Simplified Surface Energy Balance (SSEBop) model in support of the National Water Census, a USGS research program designed to build decision support capacity for water management agencies and other natural resource managers. By leveraging Google's Earth Engine Application Programming Interface (API) and developing software in a collaborative, open-platform environment, we rapidly advance from research towards applications for large-area field-scale ET mapping. Cloud computing of the Landsat image archive combined with other satellite, climate, and weather data, is creating never imagined opportunities for assessing ET model behavior and uncertainty, and ultimately providing the ability for more robust operational monitoring and assessment of water use at field-scales.
GIS Application Management for Disabled People
NASA Astrophysics Data System (ADS)
Tongkaw, Sasalak
2017-08-01
This research aimed to develop and design Geographical Information Systems (GIS) for facilitating disabled people by presenting some useful disabled information on the Google Map. The map could provide information about disabled types of people such as blind, deaf and physical movement. This research employed the Multiview 2 theory and method to plan and find out the problems in real world situation. This research used many designing data structure methods such as Data Flow Diagram, and ER-Diagram. The research focused into two parts: server site and client site which included the interface for Web-based application. The clear information of disable people on the map was useful for facilitating disabled people to find some useful information. In addition, it provided specialized data for company and government officers for managing and planning local facilities for disabled people in the cities. The disable could access the system through the Internet access at any time by using mobile or portable devices.
Learning GIS and exploring geolocated data with the all-in-one Geolokit toolbox for Google Earth
NASA Astrophysics Data System (ADS)
Watlet, A.; Triantafyllou, A.; Bastin, C.
2016-12-01
GIS software are today's essential tools to gather and visualize geological data, to apply spatial and temporal analysis and finally, to create and share interactive maps for further investigations in geosciences. Such skills are especially essential to learn for students who go through fieldtrips, samples collections or field experiments. However, time is generally missing to teach in detail all the aspects of visualizing geolocated geoscientific data. For these purposes, we developed Geolokit: a lightweight freeware dedicated to geodata visualization and written in Python, a high-level, cross-platform programming language. Geolokit software is accessible through a graphical user interface, designed to run in parallel with Google Earth, benefitting from the numerous interactive capabilities. It is designed as a very user-friendly toolbox that allows `geo-users' to import their raw data (e.g. GPS, sample locations, structural data, field pictures, maps), to use fast data analysis tools and to visualize these into the Google Earth environment using KML code; with no require of third party software, except Google Earth itself. Geolokit comes with a large number of geosciences labels, symbols, colours and placemarks and is applicable to display several types of geolocated data, including: Multi-points datasets Automatically computed contours of multi-points datasets via several interpolation methods Discrete planar and linear structural geology data in 2D or 3D supporting large range of structures input format Clustered stereonets and rose diagrams 2D cross-sections as vertical sections Georeferenced maps and grids with user defined coordinates Field pictures using either geo-tracking metadata from a camera built-in GPS module, or the same-day track of an external GPS In the end, Geolokit is helpful for quickly visualizing and exploring data without losing too much time in the numerous capabilities of GIS software suites. We are looking for students and teachers to discover all the functionalities of Geolokit. As this project is under development and planned to be open source, we are definitely looking to discussions regarding particular needs or ideas, and to contributions in the Geolokit project.
Supporting our scientists with Google Earth-based UIs.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Scott, Janine
2010-10-01
Google Earth and Google Maps are incredibly useful for researchers looking for easily-digestible displays of data. This presentation will provide a step-by-step tutorial on how to begin using Google Earth to create tools that further the mission of the DOE national lab complex.
a Map Mash-Up Application: Investigation the Temporal Effects of Climate Change on Salt Lake Basin
NASA Astrophysics Data System (ADS)
Kirtiloglu, O. S.; Orhan, O.; Ekercin, S.
2016-06-01
The main purpose of this paper is to investigate climate change effects that have been occurred at the beginning of the twenty-first century at the Konya Closed Basin (KCB) located in the semi-arid central Anatolian region of Turkey and particularly in Salt Lake region where many major wetlands located in and situated in KCB and to share the analysis results online in a Web Geographical Information System (GIS) environment. 71 Landsat 5-TM, 7-ETM+ and 8-OLI images and meteorological data obtained from 10 meteorological stations have been used at the scope of this work. 56 of Landsat images have been used for extraction of Salt Lake surface area through multi-temporal Landsat imagery collected from 2000 to 2014 in Salt lake basin. 15 of Landsat images have been used to make thematic maps of Normalised Difference Vegetation Index (NDVI) in KCB, and 10 meteorological stations data has been used to generate the Standardized Precipitation Index (SPI), which was used in drought studies. For the purpose of visualizing and sharing the results, a Web GIS-like environment has been established by using Google Maps and its useful data storage and manipulating product Fusion Tables which are all Google's free of charge Web service elements. The infrastructure of web application includes HTML5, CSS3, JavaScript, Google Maps API V3 and Google Fusion Tables API technologies. These technologies make it possible to make effective "Map Mash-Ups" involving an embedded Google Map in a Web page, storing the spatial or tabular data in Fusion Tables and add this data as a map layer on embedded map. The analysing process and map mash-up application have been discussed in detail as the main sections of this paper.
PhyloGeoViz: a web-based program that visualizes genetic data on maps.
Tsai, Yi-Hsin E
2011-05-01
The first step of many population genetic studies is the simple visualization of allele frequencies on a landscape. This basic data exploration can be challenging without proprietary software, and the manual plotting of data is cumbersome and unfeasible at large sample sizes. I present an open source, web-based program that plots any kind of frequency or count data as pie charts in Google Maps (Google Inc., Mountain View, CA). Pie polygons are then exportable to Google Earth (Google Inc.), a free Geographic Information Systems platform. Import of genetic data into Google Earth allows phylogeographers access to a wealth of spatial information layers integral to forming hypotheses and understanding patterns in the data. © 2010 Blackwell Publishing Ltd.
NASA Astrophysics Data System (ADS)
Morton, J. J.; Ferrini, V. L.
2015-12-01
The Marine Geoscience Data System (MGDS, www.marine-geo.org) operates an interactive digital data repository and metadata catalog that provides access to a variety of marine geology and geophysical data from throughout the global oceans. Its Marine-Geo Digital Library includes common marine geophysical data types and supporting data and metadata, as well as complementary long-tail data. The Digital Library also includes community data collections and custom data portals for the GeoPRISMS, MARGINS and Ridge2000 programs, for active source reflection data (Academic Seismic Portal), and for marine data acquired by the US Antarctic Program (Antarctic and Southern Ocean Data Portal). Ensuring that these data are discoverable not only through our own interfaces but also through standards-compliant web services is critical for enabling investigators to find data of interest.Over the past two years, MGDS has developed several new RESTful web services that enable programmatic access to metadata and data holdings. These web services are compliant with the EarthCube GeoWS Building Blocks specifications and are currently used to drive our own user interfaces. New web applications have also been deployed to provide a more intuitive user experience for searching, accessing and browsing metadata and data. Our new map-based search interface combines components of the Google Maps API with our web services for dynamic searching and exploration of geospatially constrained data sets. Direct introspection of nearly all data formats for hundreds of thousands of data files curated in the Marine-Geo Digital Library has allowed for precise geographic bounds, which allow geographic searches to an extent not previously possible. All MGDS map interfaces utilize the web services of the Global Multi-Resolution Topography (GMRT) synthesis for displaying global basemap imagery and for dynamically provide depth values at the cursor location.
Feature Positioning on Google Street View Panoramas
NASA Astrophysics Data System (ADS)
Tsai, V. J. D.; Chang, C.-T.
2012-07-01
Location-based services (LBS) on web-based maps and images have come into real-time since Google launched its Street View imaging services in 2007. This research employs Google Maps API and Web Service, GAE for JAVA, AJAX, Proj4js, CSS and HTML in developing an internet platform for accessing the orientation parameters of Google Street View (GSV) panoramas in order to determine the three dimensional position of interest features that appear on two overlapping panoramas by geometric intersection. A pair of GSV panoramas was examined using known points located on the Library Building of National Chung Hsing University (NCHU) with the root-mean-squared errors of ±0.522m, ±1.230m, and ±5.779m for intersection and ±0.142m, ±1.558m, and ±5.733m for resection in X, Y, and h (elevation), respectively. Potential error sources in GSV positioning were analyzed and illustrated that the errors in Google provided GSV positional parameters dominate the errors in geometric intersection. The developed system is suitable for data collection in establishing LBS applications integrated with Google Maps and Google Earth in traffic sign and infrastructure inventory by adding automatic extraction and matching techniques for points of interest (POI) from GSV panoramas.
ERIC Educational Resources Information Center
Lin, Yu-Tzu; Chang, Chia-Hu; Hou, Huei-Tse; Wu, Ke-Chou
2016-01-01
This study investigated the effectiveness of using Google Docs in collaborative concept mapping (CCM) by comparing it with a paper-and-pencil approach. A quasi-experimental study was conducted in a physics course. The control group drew concept maps using the paper-and-pencil method and face-to-face discussion, whereas the experimental group…
Usability analysis of indoor map application in a shopping centre
NASA Astrophysics Data System (ADS)
Dewi, R. S.; Hadi, R. K.
2018-04-01
Although indoor navigation is still new in Indonesia, its future development is very promising. Similar to the outdoor one, the indoor navigation technology provides several important functions to support route and landmark findings. Furthermore, there is also a need that indoor navigation can support the public safety especially during disaster evacuation process in a building. It is a common that the indoor navigation technologies are built as applications where users can access this technology using their smartphones, tablets, or personal computers. Therefore, a usability analysis is important to ensure the indoor navigation applications can be operated by users with highest functionality. Among several indoor map applications which were available in the market, this study chose to analyse indoor Google Maps due to its availability and popularity in Indonesia. The experiments to test indoor Google Maps was conducted in one of the biggest shopping centre building in Surabaya, Indonesia. The usability was measured by employing System Usability Scale (SUS) questionnaire. The result showed that the SUS score of indoor Google Maps was below the average score of other cellular applications to indicate the users still had high difficulty in operating and learning the features of indoor Google Maps.
Kuperstein, I; Bonnet, E; Nguyen, H-A; Cohen, D; Viara, E; Grieco, L; Fourquet, S; Calzone, L; Russo, C; Kondratova, M; Dutreix, M; Barillot, E; Zinovyev, A
2015-01-01
Cancerogenesis is driven by mutations leading to aberrant functioning of a complex network of molecular interactions and simultaneously affecting multiple cellular functions. Therefore, the successful application of bioinformatics and systems biology methods for analysis of high-throughput data in cancer research heavily depends on availability of global and detailed reconstructions of signalling networks amenable for computational analysis. We present here the Atlas of Cancer Signalling Network (ACSN), an interactive and comprehensive map of molecular mechanisms implicated in cancer. The resource includes tools for map navigation, visualization and analysis of molecular data in the context of signalling network maps. Constructing and updating ACSN involves careful manual curation of molecular biology literature and participation of experts in the corresponding fields. The cancer-oriented content of ACSN is completely original and covers major mechanisms involved in cancer progression, including DNA repair, cell survival, apoptosis, cell cycle, EMT and cell motility. Cell signalling mechanisms are depicted in detail, together creating a seamless ‘geographic-like' map of molecular interactions frequently deregulated in cancer. The map is browsable using NaviCell web interface using the Google Maps engine and semantic zooming principle. The associated web-blog provides a forum for commenting and curating the ACSN content. ACSN allows uploading heterogeneous omics data from users on top of the maps for visualization and performing functional analyses. We suggest several scenarios for ACSN application in cancer research, particularly for visualizing high-throughput data, starting from small interfering RNA-based screening results or mutation frequencies to innovative ways of exploring transcriptomes and phosphoproteomes. Integration and analysis of these data in the context of ACSN may help interpret their biological significance and formulate mechanistic hypotheses. ACSN may also support patient stratification, prediction of treatment response and resistance to cancer drugs, as well as design of novel treatment strategies. PMID:26192618
Beckers, Niek; Schreiner, Sam; Bertrand, Pierre; Mehler, Bruce; Reimer, Bryan
2017-01-01
The relative impact of using a Google Glass based voice interface to enter a destination address compared to voice and touch-entry methods using a handheld Samsung Galaxy S4 smartphone was assessed in a driving simulator. Voice entry (Google Glass and Samsung) had lower subjective workload ratings, lower standard deviation of lateral lane position, shorter task durations, faster remote Detection Response Task (DRT) reaction times, lower DRT miss rates, and resulted in less time glancing off-road than the primary visual-manual interaction with the Samsung Touch interface. Comparing voice entry methods, using Google Glass took less time, while glance metrics and reaction time to DRT events responded to were similar. In contrast, DRT miss rate was higher for Google Glass, suggesting that drivers may be under increased distraction levels but for a shorter period of time; whether one or the other equates to an overall safer driving experience is an open question. Copyright © 2016 Elsevier Ltd. All rights reserved.
Federal Register 2010, 2011, 2012, 2013, 2014
2012-07-10
... explained in the legislative history of the Omnibus Trade and Competitiveness Act of 1988, the Department... Google Maps: https://maps.google.com . The rates were in effect prior to the POR, so we adjusted them to...
Reisinger, Florian; Krishna, Ritesh; Ghali, Fawaz; Ríos, Daniel; Hermjakob, Henning; Vizcaíno, Juan Antonio; Jones, Andrew R
2012-03-01
We present a Java application programming interface (API), jmzIdentML, for the Human Proteome Organisation (HUPO) Proteomics Standards Initiative (PSI) mzIdentML standard for peptide and protein identification data. The API combines the power of Java Architecture of XML Binding (JAXB) and an XPath-based random-access indexer to allow a fast and efficient mapping of extensible markup language (XML) elements to Java objects. The internal references in the mzIdentML files are resolved in an on-demand manner, where the whole file is accessed as a random-access swap file, and only the relevant piece of XMLis selected for mapping to its corresponding Java object. The APIis highly efficient in its memory usage and can handle files of arbitrary sizes. The APIfollows the official release of the mzIdentML (version 1.1) specifications and is available in the public domain under a permissive licence at http://www.code.google.com/p/jmzidentml/. © 2012 WILEY-VCH Verlag GmbH & Co. KGaA, Weinheim.
The New USGS Volcano Hazards Program Web Site
NASA Astrophysics Data System (ADS)
Venezky, D. Y.; Graham, S. E.; Parker, T. J.; Snedigar, S. F.
2008-12-01
The U.S. Geological Survey's (USGS) Volcano Hazard Program (VHP) has launched a revised web site that uses a map-based interface to display hazards information for U.S. volcanoes. The web site is focused on better communication of hazards and background volcano information to our varied user groups by reorganizing content based on user needs and improving data display. The Home Page provides a synoptic view of the activity level of all volcanoes for which updates are written using a custom Google® Map. Updates are accessible by clicking on one of the map icons or clicking on the volcano of interest in the adjacent color-coded list of updates. The new navigation provides rapid access to volcanic activity information, background volcano information, images and publications, volcanic hazards, information about VHP, and the USGS volcano observatories. The Volcanic Activity section was tailored for emergency managers but provides information for all our user groups. It includes a Google® Map of the volcanoes we monitor, an Elevated Activity Page, a general status page, information about our Volcano Alert Levels and Aviation Color Codes, monitoring information, and links to monitoring data from VHP's volcano observatories: Alaska Volcano Observatory (AVO), Cascades Volcano Observatory (CVO), Long Valley Observatory (LVO), Hawaiian Volcano Observatory (HVO), and Yellowstone Volcano Observatory (YVO). The YVO web site was the first to move to the new navigation system and we are working on integrating the Long Valley Observatory web site next. We are excited to continue to implement new geospatial technologies to better display our hazards and supporting volcano information.
Positional Accuracy Assessment of Googleearth in Riyadh
NASA Astrophysics Data System (ADS)
Farah, Ashraf; Algarni, Dafer
2014-06-01
Google Earth is a virtual globe, map and geographical information program that is controlled by Google corporation. It maps the Earth by the superimposition of images obtained from satellite imagery, aerial photography and GIS 3D globe. With millions of users all around the globe, GoogleEarth® has become the ultimate source of spatial data and information for private and public decision-support systems besides many types and forms of social interactions. Many users mostly in developing countries are also using it for surveying applications, the matter that raises questions about the positional accuracy of the Google Earth program. This research presents a small-scale assessment study of the positional accuracy of GoogleEarth® Imagery in Riyadh; capital of Kingdom of Saudi Arabia (KSA). The results show that the RMSE of the GoogleEarth imagery is 2.18 m and 1.51 m for the horizontal and height coordinates respectively.
Google earth mapping of damage from the Nigata-Ken-Chuetsu M6.6 earthquake of 16 July 2007
Kayen, Robert E.; Steele, WM. Clint; Collins, Brian; Walker, Kevin
2008-01-01
We describe the use of Google Earth during and after a large damaging earthquake thatstruck the central Japan coast on 16 July 2007 to collect and organize damage information and guide the reconnaissance activities. This software enabled greater real-time collaboration among scientists and engineers. After the field investigation, the Google Earth map is used as a final reporting product that was directly linked to the more traditional research report document. Finally, we analyze the use of the software within the context of a post-disaster reconnaissance investigation, and link it to student use of GoogleEarth in field situations
2013-08-09
CAPE CANAVERAL, Fla. – As seen on Google Maps, the massive F-1 engines of the Saturn V's first stage on display inside the Apollo/Saturn V Center at the Kennedy Space Center Visitor Complex. Each engine stands 19 feet tall with a diameter of more than 12 feet. The five engines on the first stage produced 7.5 million pounds of thrust at liftoff. The Saturn V was used to launch NASA's Apollo missions to the moon which saw 12 astronauts land and work on the lunar surface. Google precisely mapped Kennedy Space Center and some of its historical facilities for the company's map page. Photo credit: Google/Wendy Wang
Using Mobile App Development Tools to Build a GIS Application
NASA Astrophysics Data System (ADS)
Mital, A.; Catchen, M.; Mital, K.
2014-12-01
Our group designed and built working web, android, and IOS applications using different mapping libraries as bases on which to overlay fire data from NASA. The group originally planned to make app versions for Google Maps, Leaflet, and OpenLayers. However, because the Leaflet library did not properly load on Android, the group focused efforts on the other two mapping libraries. For Google Maps, the group first designed a UI for the web app and made a working version of the app. After updating the source of fire data to one which also provided historical fire data, the design had to be modified to include the extra data. After completing a working version of the web app, the group used webview in android, a built in resource which allowed porting the web app to android without rewriting the code for android. Upon completing this, the group found Apple IOS devices had a similar capability, and so decided to add an IOS app to the project using a function similar to webview. Alongside this effort, the group began implementing an OpenLayers fire map using a simpler UI. This web app was completed fairly quickly relative to Google Maps; however, it did not include functionality such as satellite imagery or searchable locations. The group finished the project with a working android version of the Google Maps based app supporting API levels 14-19 and an OpenLayers based app supporting API levels 8-19, as well as a Google Maps based IOS app supporting both old and new screen formats. This project was implemented by high school and college students under an SGT Inc. STEM internship program
Visualizing Cross-sectional Data in a Real-World Context
NASA Astrophysics Data System (ADS)
Van Noten, K.; Lecocq, T.
2016-12-01
If you could fly around your research results in three dimensions, wouldn't you like to do it? Visualizing research results properly during scientific presentations already does half the job of informing the public on the geographic framework of your research. Many scientists use the Google Earth™ mapping service (V7.1.2.2041) because it's a great interactive mapping tool for assigning geographic coordinates to individual data points, localizing a research area, and draping maps of results over Earth's surface for 3D visualization. However, visualizations of research results in vertical cross-sections are often not shown simultaneously with the maps in Google Earth. A few tutorials and programs to display cross-sectional data in Google Earth do exist, and the workflow is rather simple. By importing a cross-sectional figure into in the open software SketchUp Make [Trimble Navigation Limited, 2016], any spatial model can be exported to a vertical figure in Google Earth. In this presentation a clear workflow/tutorial is presented how to image cross-sections manually in Google Earth. No software skills, nor any programming codes are required. It is very easy to use, offers great possibilities for teaching and allows fast figure manipulation in Google Earth. The full workflow can be found in "Van Noten, K. 2016. Visualizing Cross-Sectional Data in a Real-World Context. EOS, Transactions AGU, 97, 16-19".The video tutorial can be found here: https://www.youtube.com/watch?v=Tr8LwFJ4RYU&Figure: Cross-sectional Research Examples Illustrated in Google Earth
HCLS 2.0/3.0: health care and life sciences data mashup using Web 2.0/3.0.
Cheung, Kei-Hoi; Yip, Kevin Y; Townsend, Jeffrey P; Scotch, Matthew
2008-10-01
We describe the potential of current Web 2.0 technologies to achieve data mashup in the health care and life sciences (HCLS) domains, and compare that potential to the nascent trend of performing semantic mashup. After providing an overview of Web 2.0, we demonstrate two scenarios of data mashup, facilitated by the following Web 2.0 tools and sites: Yahoo! Pipes, Dapper, Google Maps and GeoCommons. In the first scenario, we exploited Dapper and Yahoo! Pipes to implement a challenging data integration task in the context of DNA microarray research. In the second scenario, we exploited Yahoo! Pipes, Google Maps, and GeoCommons to create a geographic information system (GIS) interface that allows visualization and integration of diverse categories of public health data, including cancer incidence and pollution prevalence data. Based on these two scenarios, we discuss the strengths and weaknesses of these Web 2.0 mashup technologies. We then describe Semantic Web, the mainstream Web 3.0 technology that enables more powerful data integration over the Web. We discuss the areas of intersection of Web 2.0 and Semantic Web, and describe the potential benefits that can be brought to HCLS research by combining these two sets of technologies.
HCLS 2.0/3.0: Health Care and Life Sciences Data Mashup Using Web 2.0/3.0
Cheung, Kei-Hoi; Yip, Kevin Y.; Townsend, Jeffrey P.; Scotch, Matthew
2010-01-01
We describe the potential of current Web 2.0 technologies to achieve data mashup in the health care and life sciences (HCLS) domains, and compare that potential to the nascent trend of performing semantic mashup. After providing an overview of Web 2.0, we demonstrate two scenarios of data mashup, facilitated by the following Web 2.0 tools and sites: Yahoo! Pipes, Dapper, Google Maps and GeoCommons. In the first scenario, we exploited Dapper and Yahoo! Pipes to implement a challenging data integration task in the context of DNA microarray research. In the second scenario, we exploited Yahoo! Pipes, Google Maps, and GeoCommons to create a geographic information system (GIS) interface that allows visualization and integration of diverse categories of public health data, including cancer incidence and pollution prevalence data. Based on these two scenarios, we discuss the strengths and weaknesses of these Web 2.0 mashup technologies. We then describe Semantic Web, the mainstream Web 3.0 technology that enables more powerful data integration over the Web. We discuss the areas of intersection of Web 2.0 and Semantic Web, and describe the potential benefits that can be brought to HCLS research by combining these two sets of technologies. PMID:18487092
The Adversarial Route Analysis Tool: A Web Application
DOE Office of Scientific and Technical Information (OSTI.GOV)
Casson, William H. Jr.
2012-08-02
The Adversarial Route Analysis Tool is a type of Google maps for adversaries. It's a web-based Geospatial application similar to Google Maps. It helps the U.S. government plan operations that predict where an adversary might be. It's easily accessible and maintainble and it's simple to use without much training.
Ameisen, David; Deroulers, Christophe; Perrier, Valérie; Bouhidel, Fatiha; Battistella, Maxime; Legrès, Luc; Janin, Anne; Bertheau, Philippe; Yunès, Jean-Baptiste
2014-01-01
Since microscopic slides can now be automatically digitized and integrated in the clinical workflow, quality assessment of Whole Slide Images (WSI) has become a crucial issue. We present a no-reference quality assessment method that has been thoroughly tested since 2010 and is under implementation in multiple sites, both public university-hospitals and private entities. It is part of the FlexMIm R&D project which aims to improve the global workflow of digital pathology. For these uses, we have developed two programming libraries, in Java and Python, which can be integrated in various types of WSI acquisition systems, viewers and image analysis tools. Development and testing have been carried out on a MacBook Pro i7 and on a bi-Xeon 2.7GHz server. Libraries implementing the blur assessment method have been developed in Java, Python, PHP5 and MySQL5. For web applications, JavaScript, Ajax, JSON and Sockets were also used, as well as the Google Maps API. Aperio SVS files were converted into the Google Maps format using VIPS and Openslide libraries. We designed the Java library as a Service Provider Interface (SPI), extendable by third parties. Analysis is computed in real-time (3 billion pixels per minute). Tests were made on 5000 single images, 200 NDPI WSI, 100 Aperio SVS WSI converted to the Google Maps format. Applications based on our method and libraries can be used upstream, as calibration and quality control tool for the WSI acquisition systems, or as tools to reacquire tiles while the WSI is being scanned. They can also be used downstream to reacquire the complete slides that are below the quality threshold for surgical pathology analysis. WSI may also be displayed in a smarter way by sending and displaying the regions of highest quality before other regions. Such quality assessment scores could be integrated as WSI's metadata shared in clinical, research or teaching contexts, for a more efficient medical informatics workflow.
Return of the Google Game: More Fun Ideas to Transform Students into Skilled Researchers
ERIC Educational Resources Information Center
Watkins, Katrine
2008-01-01
Teens are impatient and unsophisticated online researchers who are often limited by their poor reading skills. Because they are attracted to clean and simple Web interfaces, they often turn to Google--and now Wikipedia--to help meet their research needs. The Google Game, co-authored by this author, teaches kids that there is a well-thought-out…
National Geothermal Data System: Open Access to Geoscience Data, Maps, and Documents
NASA Astrophysics Data System (ADS)
Caudill, C. M.; Richard, S. M.; Musil, L.; Sonnenschein, A.; Good, J.
2014-12-01
The U.S. National Geothermal Data System (NGDS) provides free open access to millions of geoscience data records, publications, maps, and reports via distributed web services to propel geothermal research, development, and production. NGDS is built on the US Geoscience Information Network (USGIN) data integration framework, which is a joint undertaking of the USGS and the Association of American State Geologists (AASG), and is compliant with international standards and protocols. NGDS currently serves geoscience information from 60+ data providers in all 50 states. Free and open source software is used in this federated system where data owners maintain control of their data. This interactive online system makes geoscience data easily discoverable, accessible, and interoperable at no cost to users. The dynamic project site http://geothermaldata.org serves as the information source and gateway to the system, allowing data and applications discovery and availability of the system's data feed. It also provides access to NGDS specifications and the free and open source code base (on GitHub), a map-centric and library style search interface, other software applications utilizing NGDS services, NGDS tutorials (via YouTube and USGIN site), and user-created tools and scripts. The user-friendly map-centric web-based application has been created to support finding, visualizing, mapping, and acquisition of data based on topic, location, time, provider, or key words. Geographic datasets visualized through the map interface also allow users to inspect the details of individual GIS data points (e.g. wells, geologic units, etc.). In addition, the interface provides the information necessary for users to access the GIS data from third party software applications such as GoogleEarth, UDig, and ArcGIS. A redistributable, free and open source software package called GINstack (USGIN software stack) was also created to give data providers a simple way to release data using interoperable and shareable standards, upload data and documents, and expose those data as a node in the NGDS or any larger data system through a CSW endpoint. The easy-to-use interface is supported by back-end software including Postgres, GeoServer, and custom CKAN extensions among others.
Data Access and Web Services at the EarthScope Plate Boundary Observatory
NASA Astrophysics Data System (ADS)
Matykiewicz, J.; Anderson, G.; Henderson, D.; Hodgkinson, K.; Hoyt, B.; Lee, E.; Persson, E.; Torrez, D.; Smith, J.; Wright, J.; Jackson, M.
2007-12-01
The EarthScope Plate Boundary Observatory (PBO) at UNAVCO, Inc., part of the NSF-funded EarthScope project, is designed to study the three-dimensional strain field resulting from deformation across the active boundary zone between the Pacific and North American plates in the western United States. To meet these goals, PBO will install 880 continuous GPS stations, 103 borehole strainmeter stations, and five laser strainmeters, as well as manage data for 209 previously existing continuous GPS stations and one previously existing laser strainmeter. UNAVCO provides access to data products from these stations, as well as general information about the PBO project, via the PBO web site (http://pboweb.unavco.org). GPS and strainmeter data products can be found using a variety of access methods, incuding map searches, text searches, and station specific data retrieval. In addition, the PBO construction status is available via multiple mapping interfaces, including custom web based map widgets and Google Earth. Additional construction details can be accessed from PBO operational pages and station specific home pages. The current state of health for the PBO network is available with the statistical snap-shot, full map interfaces, tabular web based reports, and automatic data mining and alerts. UNAVCO is currently working to enhance the community access to this information by developing a web service framework for the discovery of data products, interfacing with operational engineers, and exposing data services to third party participants. In addition, UNAVCO, through the PBO project, provides advanced data management and monitoring systems for use by the community in operating geodetic networks in the United States and beyond. We will demonstrate these systems during the AGU meeting, and we welcome inquiries from the community at any time.
The Plate Boundary Observatory: Community Focused Web Services
NASA Astrophysics Data System (ADS)
Matykiewicz, J.; Anderson, G.; Lee, E.; Hoyt, B.; Hodgkinson, K.; Persson, E.; Wright, J.; Torrez, D.; Jackson, M.
2006-12-01
The Plate Boundary Observatory (PBO), part of the NSF-funded EarthScope project, is designed to study the three-dimensional strain field resulting from deformation across the active boundary zone between the Pacific and North American plates in the western United States. To meet these goals, PBO will install 852 continuous GPS stations, 103 borehole strainmeter stations, 28 tiltmeters, and five laser strainmeters, as well as manage data for 209 previously existing continuous GPS stations. UNAVCO provides access to data products from these stations, as well as general information about the PBO project, via the PBO web site (http://pboweb.unavco.org). GPS and strainmeter data products can be found using a variety of channels, including map searches, text searches, and station specific data retrieval. In addition, the PBO construction status is available via multiple mapping interfaces, including custom web based map widgets and Google Earth. Additional construction details can be accessed from PBO operational pages and station specific home pages. The current state of health for the PBO network is available with the statistical snap-shot, full map interfaces, tabular web based reports, and automatic data mining and alerts. UNAVCO is currently working to enhance the community access to this information by developing a web service framework for the discovery of data products, interfacing with operational engineers, and exposing data services to third party participants. In addition, UNAVCO, through the PBO project, provides advanced data management and monitoring systems for use by the community in operating geodetic networks in the United States and beyond. We will demonstrate these systems during the AGU meeting, and we welcome inquiries from the community at any time.
Creating Web Area Segments with Google Analytics
Segments allow you to quickly access data for a predefined set of Sessions or Users, such as government or education users, or sessions in a particular state. You can then apply this segment to any report within the Google Analytics (GA) interface.
Building a Smart Portal for Astronomy
NASA Astrophysics Data System (ADS)
Derriere, S.; Boch, T.
2011-07-01
The development of a portal for accessing astronomical resources is not an easy task. The ever-increasing complexity of the data products can result in very complex user interfaces, requiring a lot of effort and learning from the user in order to perform searches. This is often a design choice, where the user must explicitly set many constraints, while the portal search logic remains simple. We investigated a different approach, where the query interface is kept as simple as possible (ideally, a simple text field, like for Google search), and the search logic is made much more complex to interpret the query in a relevant manner. We will present the implications of this approach in terms of interpretation and categorization of the query parameters (related to astronomical vocabularies), translation (mapping) of these concepts into the portal components metadata, identification of query schemes and use cases matching the input parameters, and delivery of query results to the user.
From Google Maps to Google Models (Invited)
NASA Astrophysics Data System (ADS)
Moore, R. V.
2010-12-01
Why hasn’t integrated modelling taken off? To its advocates, it is self-evidently the best and arguably the only tool available for understanding and predicting the likely response of the environment to events and policies. Legislation requires managers to ensure that their plans are sustainable. How, other than by modelling the interacting processes involved, can the option with the greatest benefits be identified? Integrated modelling (IM) is seen to have huge potential. In science, IM is used to extend and encapsulate our understanding of the whole earth system. Such models are beginning to be incorporated in operational decision support systems and used to seek sustainable solutions to society’s problems, but only on a limited scale. Commercial take up is negligible yet the opportunities would appear limitless. The need is there; the potential is there, so what is inhibiting IM’s take up? What must be done to reap the rewards of the R & D to date? To answer the question, it useful to look back at the developments which have seen paper maps evolve into Google Maps and the systems that now surround it; facilities available not just to experts and governments but to anyone with a an iphone and an internet connection. The initial objective was to automate the process of drawing lines on paper, though it was quickly realised that digitising maps was the key to unlocking the information they held. However, it took thousands of PhD and MSc projects before a computer could generate a map comparable to that produced by a cartographer and many more before it was possible to extract reliable useful information from maps. It also required advances in IT and a change of mindset from one focused on paper map production to one focused on information delivery. To move from digital maps to Google Maps required the availability of data on a world scale, the resources to bring them together, the development of remote sensing, satellite navigation and communications technology and the creation of a commercial climate and conditions that allowed businesses anywhere to exploit the new information. This talk will draw lessons from the experience and imagine how Google Maps could become Google Models. The first lesson is time scale, it took far longer for digital mapping to move out of the development phase than most expected. Its first real customers were the public utilities. They are large organisations, risk averse and take time to change their ways of working; integrated modellers should not be surprised by the slow take up. Few of the early commercial entrants made any significant profits. It was only when the data reached critical mass and became accessible, when the systems became easy to use, affordable and accessible via the web, when convincing demonstrations became available and the necessary standards emerged that Google Maps could emerge. IM has yet to reach this point. It has far bigger technical, scientific and institutional challenges to overcome. The resources required will be large. It is possible though that they could be marshalled by creating an open source community of practice. However, that community will need a facilitating core group and standards to succeed. Having seen what Google Maps made possible, the innovative ideas it released, it is not difficult to imagine where a community of practice might take IM.
KML-based teaching lessons developed by Google in partnership with the University of Alaska.
NASA Astrophysics Data System (ADS)
Kolb, E. J.; Bailey, J.; Bishop, A.; Cain, J.; Goddard, M.; Hurowitz, K.; Kennedy, K.; Ornduff, T.; Sfraga, M.; Wernecke, J.
2008-12-01
The focus of Google's Geo Education outreach efforts (http://www.google.com/educators/geo.html) is on helping primary, secondary, and post-secondary educators incorporate Google Earth and Sky, Google Maps, and SketchUp into their classroom lessons. In this poster and demonstration, we will show our KML-based science lessons that were developed in partnership with the University of Alaska and used in classroom teachings by our team to Alaskan high-school students.
Rousselet, Jérôme; Imbert, Charles-Edouard; Dekri, Anissa; Garcia, Jacques; Goussard, Francis; Vincent, Bruno; Denux, Olivier; Robinet, Christelle; Dorkeld, Franck; Roques, Alain; Rossi, Jean-Pierre
2013-01-01
Mapping species spatial distribution using spatial inference and prediction requires a lot of data. Occurrence data are generally not easily available from the literature and are very time-consuming to collect in the field. For that reason, we designed a survey to explore to which extent large-scale databases such as Google maps and Google Street View could be used to derive valid occurrence data. We worked with the Pine Processionary Moth (PPM) Thaumetopoea pityocampa because the larvae of that moth build silk nests that are easily visible. The presence of the species at one location can therefore be inferred from visual records derived from the panoramic views available from Google Street View. We designed a standardized procedure allowing evaluating the presence of the PPM on a sampling grid covering the landscape under study. The outputs were compared to field data. We investigated two landscapes using grids of different extent and mesh size. Data derived from Google Street View were highly similar to field data in the large-scale analysis based on a square grid with a mesh of 16 km (96% of matching records). Using a 2 km mesh size led to a strong divergence between field and Google-derived data (46% of matching records). We conclude that Google database might provide useful occurrence data for mapping the distribution of species which presence can be visually evaluated such as the PPM. However, the accuracy of the output strongly depends on the spatial scales considered and on the sampling grid used. Other factors such as the coverage of Google Street View network with regards to sampling grid size and the spatial distribution of host trees with regards to road network may also be determinant.
Dekri, Anissa; Garcia, Jacques; Goussard, Francis; Vincent, Bruno; Denux, Olivier; Robinet, Christelle; Dorkeld, Franck; Roques, Alain; Rossi, Jean-Pierre
2013-01-01
Mapping species spatial distribution using spatial inference and prediction requires a lot of data. Occurrence data are generally not easily available from the literature and are very time-consuming to collect in the field. For that reason, we designed a survey to explore to which extent large-scale databases such as Google maps and Google street view could be used to derive valid occurrence data. We worked with the Pine Processionary Moth (PPM) Thaumetopoea pityocampa because the larvae of that moth build silk nests that are easily visible. The presence of the species at one location can therefore be inferred from visual records derived from the panoramic views available from Google street view. We designed a standardized procedure allowing evaluating the presence of the PPM on a sampling grid covering the landscape under study. The outputs were compared to field data. We investigated two landscapes using grids of different extent and mesh size. Data derived from Google street view were highly similar to field data in the large-scale analysis based on a square grid with a mesh of 16 km (96% of matching records). Using a 2 km mesh size led to a strong divergence between field and Google-derived data (46% of matching records). We conclude that Google database might provide useful occurrence data for mapping the distribution of species which presence can be visually evaluated such as the PPM. However, the accuracy of the output strongly depends on the spatial scales considered and on the sampling grid used. Other factors such as the coverage of Google street view network with regards to sampling grid size and the spatial distribution of host trees with regards to road network may also be determinant. PMID:24130675
Community Near-Port Modeling System (C-PORT): Briefing for ...
What C-PORT is: Screening level tool for assessing port activities and exploring the range of potential impacts that changes to port operations might have on local air quality; Analysis of decision alternatives through mapping of the likely pattern of potential pollutant dispersion and an estimated change in pollutant concentrations for user-designated scenarios; Designed primarily to evaluate the local air quality impacts of proposed port expansion or modernization, as well as to identify options for mitigating any impacts; Currently includes data from 21 US seaports and features a map-based interface similar to the widely used Google Earth; Still under development, C-PORT is designed as an easy-to-use computer modeling tool for users, such as state air quality managers and planners. This is part of our product outreach prior to model public release and to solicit for additional beta testers.
2013-08-09
CAPE CANAVERAL, Fla. – As seen on Google Maps, a Space Shuttle Main Engine, or SSME, stands inside the Engine Shop at Orbiter Processing Facility 3 at NASA's Kennedy Space Center. Each orbiter used three of the engines during launch and ascent into orbit. The engines burn super-cold liquid hydrogen and liquid oxygen and each one produces 155,000 pounds of thrust. The engines, known in the industry as RS-25s, could be reused on multiple shuttle missions. They will be used again later this decade for NASA's Space Launch System rocket. Google precisely mapped the space center and some of its historical facilities for the company's map page. The work allows Internet users to see inside buildings at Kennedy as they were used during the space shuttle era. Photo credit: Google/Wendy Wang
NASA Astrophysics Data System (ADS)
Welton, B.; Chouinard, K.; Sultan, M.; Becker, D.; Milewski, A.; Becker, R.
2010-12-01
Rising populations in the arid and semi arid parts of the World are increasing the demand for fresh water supplies worldwide. Many data sets needed for assessment of hydrologic applications across vast regions of the world are expensive, unpublished, difficult to obtain, or at varying scales which complicates their use. Fortunately, this situation is changing with the development of global remote sensing datasets and web-based platforms such as GIS Server. GIS provides a cost effective vehicle for comparing, analyzing, and querying a variety of spatial datasets as geographically referenced layers. We have recently constructed a web-based GIS, that incorporates all relevant geological, geochemical, geophysical, and remote sensing data sets that were readily used to identify reservoir types and potential well locations on local and regional scales in various tectonic settings including: (1) extensional environment (Red Sea rift), (2) transcurrent fault system (Najd Fault in the Arabian-Nubian Shield), and (3) compressional environments (Himalayas). The web-based GIS could also be used to detect spatial and temporal trends in precipitation, recharge, and runoff in large watersheds on local, regional, and continental scales. These applications were enabled through the construction of a web-based ArcGIS Server with Google Map’s interface and the development of customized geoprocessing tools. ArcGIS Server provides out-of-the-box setups that are generic in nature. This platform includes all of the standard web based GIS tools (e.g. pan, zoom, identify, search, data querying, and measurement). In addition to the standard suite of tools provided by ArcGIS Server an additional set of advanced data manipulation and display tools was also developed to allow for a more complete and customizable view of the area of interest. The most notable addition to the standard GIS Server tools is the custom on-demand geoprocessing tools (e.g., graph, statistical functions, custom raster creation, profile, TRMM). The generation of a wide range of derivative maps (e.g., buffer zone, contour map, graphs, temporal rainfall distribution maps) from various map layers (e.g., geologic maps, geophysics, satellite images) allows for more user flexibility. The use of these tools along with Google Map’s API which enables the website user to utilize high quality GeoEye 2 images provide by Google in conjunction with our data, creates a more complete image of the area being observed and allows for custom derivative maps to be created in the field and viewed immediately on the web, processes that were restricted to offline databases.
Kaewpitoon, Soraya J; Rujirakul, Ratana; Joosiri, Apinya; Jantakate, Sirinun; Sangkudloa, Amnat; Kaewthani, Sarochinee; Chimplee, Kanokporn; Khemplila, Kritsakorn; Kaewpitoon, Natthawut
2016-01-01
Cholangiocarcinoma (CCA) is a serious problem in Thailand, particularly in the northeastern and northern regions. Database of population at risk are need required for monitoring, surveillance, home health care, and home visit. Therefore, this study aimed to develop a geographic information system (GIS) database and Google map of the population at risk of CCA in Mueang Yang district, Nakhon Ratchasima province, northeastern Thailand during June to October 2015. Populations at risk were screened using the Korat CCA verbal screening test (KCVST). Software included Microsoft Excel, ArcGIS, and Google Maps. The secondary data included the point of villages, sub-district boundaries, district boundaries, point of hospital in Mueang Yang district, used for created the spatial databese. The populations at risk for CCA and opisthorchiasis were used to create an arttribute database. Data were tranfered to WGS84 UTM ZONE 48. After the conversion, all of the data were imported into Google Earth using online web pages www.earthpoint.us. Some 222 from a 4,800 population at risk for CCA constituted a high risk group. Geo-visual display available at following www.google.com/maps/d/u/0/ edit?mid=zPxtcHv_iDLo.kvPpxl5mAs90 and hl=th. Geo-visual display 5 layers including: layer 1, village location and number of the population at risk for CCA; layer 2, sub-district health promotion hospital in Mueang Yang district and number of opisthorchiasis; layer 3, sub-district district and the number of population at risk for CCA; layer 4, district hospital and the number of population at risk for CCA and number of opisthorchiasis; and layer 5, district and the number of population at risk for CCA and number of opisthorchiasis. This GIS database and Google map production process is suitable for further monitoring, surveillance, and home health care for CCA sufferers.
Generating and Visualizing Climate Indices using Google Earth Engine
NASA Astrophysics Data System (ADS)
Erickson, T. A.; Guentchev, G.; Rood, R. B.
2017-12-01
Climate change is expected to have largest impacts on regional and local scales. Relevant and credible climate information is needed to support the planning and adaptation efforts in our communities. The volume of climate projections of temperature and precipitation is steadily increasing, as datasets are being generated on finer spatial and temporal grids with an increasing number of ensembles to characterize uncertainty. Despite advancements in tools for querying and retrieving subsets of these large, multi-dimensional datasets, ease of access remains a barrier for many existing and potential users who want to derive useful information from these data, particularly for those outside of the climate modelling research community. Climate indices, that can be derived from daily temperature and precipitation data, such as annual number of frost days or growing season length, can provide useful information to practitioners and stakeholders. For this work the NASA Earth Exchange Global Daily Downscaled Projections (NEX-GDDP) dataset was loaded into Google Earth Engine, a cloud-based geospatial processing platform. Algorithms that use the Earth Engine API to generate several climate indices were written. The indices were chosen from the set developed by the joint CCl/CLIVAR/JCOMM Expert Team on Climate Change Detection and Indices (ETCCDI). Simple user interfaces were created that allow users to query, produce maps and graphs of the indices, as well as download results for additional analyses. These browser-based interfaces could allow users in low-bandwidth environments to access climate information. This research shows that calculating climate indices from global downscaled climate projection datasets and sharing them widely using cloud computing technologies is feasible. Further development will focus on exposing the climate indices to existing applications via the Earth Engine API, and building custom user interfaces for presenting climate indices to a diverse set of user groups.
Facilitating Semantic Interoperability Among Ocean Data Systems: ODIP-R2R Student Outcomes
NASA Astrophysics Data System (ADS)
Stocks, K. I.; Chen, Y.; Shepherd, A.; Chandler, C. L.; Dockery, N.; Elya, J. L.; Smith, S. R.; Ferreira, R.; Fu, L.; Arko, R. A.
2014-12-01
With informatics providing an increasingly important set of tools for geoscientists, it is critical to train the next generation of scientists in information and data techniques. The NSF-supported Rolling Deck to Repository (R2R) Program works with the academic fleet community to routinely document, assess, and preserve the underway sensor data from U.S. research vessels. The Ocean Data Interoperability Platform (ODIP) is an EU-US-Australian collaboration fostering interoperability among regional e-infrastructures through workshops and joint prototype development. The need to align terminology between systems is a common challenge across all of the ODIP prototypes. Five R2R students were supported to address aspects of semantic interoperability within ODIP. Developing a vocabulary matching service that links terms from different vocabularies with similar concept. The service implements Google Refine reconciliation service interface such that users can leverage Google Refine application as a friendly user interface while linking different vocabulary terms. Developing Resource Description Framework (RDF) resources that map Shipboard Automated Meteorological Oceanographic System (SAMOS) vocabularies to internationally served vocabularies. Each SAMOS vocabulary term (data parameter and quality control flag) will be described as an RDF resource page. These RDF resources allow for enhanced discoverability and retrieval of SAMOS data by enabling data searches based on parameter. Improving data retrieval and interoperability by exposing data and mapped vocabularies using Semantic Web technologies. We have collaborated with ODIP participating organizations in order to build a generalized data model that will be used to populate a SPARQL endpoint in order to provide expressive querying over our data files. Mapping local and regional vocabularies used by R2R to those used by ODIP partners. This work is described more fully in a companion poster. Making published Linked Data Web developer-friendly with a RESTful service. This goal was achieved by defining a proxy layer on top of the existing SPARQL endpoint that 1) translates HTTP requests into SPARQL queries, and 2) renders the returned results as required by the request sender using content negotiation, suffixes and parameters.
Pulse oximeter for cyclists in smartphone
NASA Astrophysics Data System (ADS)
Martins, L.; Gaidos, O.; dos Santos, I.
2015-01-01
The monitoring of cyclists during physical activity is an important factor to improve their performance. We discuss a new approaches based on smartphone for monitoring physiological signal wirelessly for cyclists, using a pulse oximeter sensor attached to the rider's forehead. This paper presents a wireless pulse Oximeter that was developed with a Nellcor's module, which uses the Standard Host Interface Protocol (SHIP) for communication with the Bluetooth module and sends data for a Smartphone with Android O.S. Then these data are shown in the screen: the heartbeat and saturation percentage. The application was created with App Inventor and the data are sent to Google Maps via Twitter. The results demonstrate the possibility of developing a successful prototype.
NASA Technical Reports Server (NTRS)
Perez Guerrero, Geraldo A.; Armstrong, Duane; Underwood, Lauren
2015-01-01
This project is creating a cloud-enabled, HTML 5 web application to help oyster fishermen and state agencies apply Earth science to improve the management of this important natural and economic resource. The Oyster Fisheries app gathers and analyzes environmental and water quality information, and alerts fishermen and resources managers about problems in oyster fishing waters. An intuitive interface based on Google Maps displays the geospatial information and provides familiar interactive controls to the users. Alerts can be tailored to notify users when conditions in specific leases or public fishing areas require attention. The app is hosted on the Amazon Web Services cloud. It is being developed and tested using some of the latest web development tools such as web components and Polymer.
ERIC Educational Resources Information Center
Rochkind, Jonathan
2007-01-01
The ability to search and receive results in more than one database through a single interface--or metasearch--is something many users want. Google Scholar--the search engine of specifically scholarly content--and library metasearch products like Ex Libris's MetaLib, Serials Solution's Central Search, WebFeat, and products based on MuseGlobal used…
Google Voice: Connecting Your Telephone to the 21st Century
ERIC Educational Resources Information Center
Johnson, Benjamin E.
2010-01-01
The foundation of the mighty Google Empire rests upon an algorithm that connects people to information--things such as websites, maps, and restaurant reviews. Lately it seems that people are less interested in connecting with information than they are with connecting to one another, which begs the question, "Is Facebook the new Google?" Given this…
Cultural Adventures for the Google[TM] Generation
ERIC Educational Resources Information Center
Dann, Tammy
2010-01-01
Google Earth is a computer program that allows users to view the Earth through satellite imagery and maps, to see cities from above and through street views, and to search for addresses and browse locations. Many famous buildings and structures from around the world have detailed 3D views accessible on Google Earth. It is possible to explore the…
Visualizing Geographic Data in Google Earth for Education and Outreach
NASA Astrophysics Data System (ADS)
Martin, D. J.; Treves, R.
2008-12-01
Google Earth is an excellent tool to help students and the public visualize scientific data as with low technical skill scientific content can be shown in three dimensions against a background of remotely sensed imagery. It therefore has a variety of uses in university education and as a tool for public outreach. However, in both situations it is of limited value if it is only used to attract attention with flashy three dimensional animations. In this poster we shall illustrate several applications that represent what we believe is good educational practice. The first example shows how the combination of a floor map and a projection of Google Earth on a screen can be used to produce active learning. Students are asked to imagine where they would build a house on Big Island Hawaii in order to avoid volcanic hazards. In the second example Google Earth is used to illustrate evidence over a range of scales in a description of Lake Agassiz flood events which would be more difficult to comprehend in a traditional paper based format. In the final example a simple text manipulation application "TMapper" is used to change the color palette of a thematic map generated by the students in Google Earth to teach them about the use of color in map design.
Visualize Your Data with Google Fusion Tables
NASA Astrophysics Data System (ADS)
Brisbin, K. E.
2011-12-01
Google Fusion Tables is a modern data management platform that makes it easy to host, manage, collaborate on, visualize, and publish tabular data online. Fusion Tables allows users to upload their own data to the Google cloud, which they can then use to create compelling and interactive visualizations with the data. Users can view data on a Google Map, plot data in a line chart, or display data along a timeline. Users can share these visualizations with others to explore and discover interesting trends about various types of data, including scientific data such as invasive species or global trends in disease. Fusion Tables has been used by many organizations to visualize a variety of scientific data. One example is the California Redistricting Map created by the LA Times: http://goo.gl/gwZt5 The Pacific Institute and Circle of Blue have used Fusion Tables to map the quality of water around the world: http://goo.gl/T4SX8 The World Resources Institute mapped the threat level of coral reefs using Fusion Tables: http://goo.gl/cdqe8 What attendees will learn in this session: This session will cover all the steps necessary to use Fusion Tables to create a variety of interactive visualizations. Attendees will begin by learning about the various options for uploading data into Fusion Tables, including Shapefile, KML file, and CSV file import. Attendees will then learn how to use Fusion Tables to manage their data by merging it with other data and controlling the permissions of the data. Finally, the session will cover how to create a customized visualization from the data, and share that visualization with others using both Fusion Tables and the Google Maps API.
[Who Hits the Mark? A Comparative Study of the Free Geocoding Services of Google and OpenStreetMap].
Lemke, D; Mattauch, V; Heidinger, O; Hense, H W
2015-09-01
Geocoding, the process of converting textual information (addresses) into geographic coordinates is increasingly used in public health/epidemiological research and practice. To date, little attention has been paid to geocoding quality and its impact on different types of spatially-related health studies. The primary aim of this study was to compare 2 freely available geocoding services (Google and OpenStreetMap) with regard to matching rate (percentage of address records capable of being geocoded) and positional accuracy (distance between geocodes and the ground truth locations). Residential addresses were geocoded by the NRW state office for information and technology and were considered as reference data (gold standard). The gold standard included the coordinates, the quality of the addresses (4 categories), and a binary urbanity indicator based on the CORINE land cover data. 2 500 addresses were randomly sampled after stratification for address quality and urbanity indicator (approximately 20 000 addresses). These address samples were geocoded using the geocoding services from Google and OSM. In general, both geocoding services showed a decrease in the matching rate with decreasing address quality and urbanity. Google showed consistently a higher completeness than OSM (>93 vs. >82%). Also, the cartographic confounding between urban and rural regions was less distinct with Google's geocoding API. Regarding the positional accuracy of the geo-coordinates, Google also showed the smallest deviations from the reference coordinates, with a median of <9 vs. <175.8 m. The cumulative density function derived from the positional accuracy showed for Google that nearly 95% and for OSM 50% of the addresses were geocoded within <50 m of their reference coordinates. The geocoding API from Google is superior to OSM regarding completeness and positional accuracy of the geocoded addresses. On the other hand, Google has several restrictions, such as the limitation of the requests to 2 500 addresses per 24 h and the presentation of the results exclusively on Google Maps, which may complicate the use for scientific purposes. © Georg Thieme Verlag KG Stuttgart · New York.
2013-08-09
CAPE CANAVERAL, Fla. – As seen on Google Maps, the Rotating Service Structure at Launch Complex 39A at NASA's Kennedy Space Center housed space shuttle payloads temporarily so they could be loaded inside the 60-foot-long cargo bay of a shuttle before launch. The RSS, as the structure was known, was hinged to the Fixed Service Structure on one side and rolled on a rail on the other. As its name suggests, the enclosed facility would rotate into place around the shuttle as it stood at the launch pad. Once in place, the RSS protected the shuttle and its cargo. Google precisely mapped the space center and some of its historical facilities for the company's map page. The work allows Internet users to see inside buildings at Kennedy as they were used during the space shuttle era. Photo credit: Google/Wendy Wang
Active Fire Mapping Program Current Large Incidents (Home) New Large Incidents Fire Detection Maps MODIS Satellite Imagery VIIRS Satellite Imagery Fire Detection GIS Data Fire Data in Google Earth ...
DOE Office of Scientific and Technical Information (OSTI.GOV)
Susandi, Armi, E-mail: armi@meteo.itb.ac.id; Tamamadin, Mamad, E-mail: mamadtama@meteo.itb.ac.id; Djamal, Erizal, E-mail: erizal-jamal@yahoo.com
This paper describes information system of rice planting calendar to help farmers in determining the time for rice planting. The information includes rainfall prediction in ten days (dasarian) scale overlaid to map of rice field to produce map of rice planting in village level. The rainfall prediction was produced by stochastic modeling using Fast Fourier Transform (FFT) and Non-Linier Least Squares methods to fit the curve of function to the rainfall data. In this research, the Fourier series has been modified become non-linear function to follow the recent characteristics of rainfall that is non stationary. The results have been alsomore » validated in 4 steps, including R-Square, RMSE, R-Skill, and comparison with field data. The development of information system (cyber extension) provides information such as rainfall prediction, prediction of the planting time, and interactive space for farmers to respond to the information submitted. Interfaces for interactive response will be critical to the improvement of prediction accuracy of information, both rainfall and planting time. The method used to get this information system includes mapping on rice planting prediction, converting the format file, developing database system, developing website, and posting website. Because of this map was overlaid with the Google map, the map files must be converted to the .kml file format.« less
ERIC Educational Resources Information Center
Fluke, Christopher J.
2009-01-01
I report on a pilot study on the use of Google Maps to provide virtual field trips as a component of a wholly online graduate course on the history of astronomy. The Astronomical Tourist Web site (http://astronomy.swin.edu.au/sao/tourist), themed around the role that specific locations on Earth have contributed to the development of astronomical…
Spatio-temporal Change Patterns of Tropical Forests from 2000 to 2014 Using MOD09A1 Dataset
NASA Astrophysics Data System (ADS)
Qin, Y.; Xiao, X.; Dong, J.
2016-12-01
Large-scale deforestation and forest degradation in the tropical region have resulted in extensive carbon emissions and biodiversity loss. However, restricted by the availability of good-quality observations, large uncertainty exists in mapping the spatial distribution of forests and their spatio-temporal changes. In this study, we proposed a pixel- and phenology-based algorithm to identify and map annual tropical forests from 2000 to 2014, using the 8-day, 500-m MOD09A1 (v005) product, under the support of Google cloud computing (Google Earth Engine). A temporal filter was applied to reduce the random noises and to identify the spatio-temporal changes of forests. We then built up a confusion matrix and assessed the accuracy of the annual forest maps based on the ground reference interpreted from high spatial resolution images in Google Earth. The resultant forest maps showed the consistent forest/non-forest, forest loss, and forest gain in the pan-tropical zone during 2000 - 2014. The proposed algorithm showed the potential for tropical forest mapping and the resultant forest maps are important for the estimation of carbon emission and biodiversity loss.
2014-01-01
Background Since microscopic slides can now be automatically digitized and integrated in the clinical workflow, quality assessment of Whole Slide Images (WSI) has become a crucial issue. We present a no-reference quality assessment method that has been thoroughly tested since 2010 and is under implementation in multiple sites, both public university-hospitals and private entities. It is part of the FlexMIm R&D project which aims to improve the global workflow of digital pathology. For these uses, we have developed two programming libraries, in Java and Python, which can be integrated in various types of WSI acquisition systems, viewers and image analysis tools. Methods Development and testing have been carried out on a MacBook Pro i7 and on a bi-Xeon 2.7GHz server. Libraries implementing the blur assessment method have been developed in Java, Python, PHP5 and MySQL5. For web applications, JavaScript, Ajax, JSON and Sockets were also used, as well as the Google Maps API. Aperio SVS files were converted into the Google Maps format using VIPS and Openslide libraries. Results We designed the Java library as a Service Provider Interface (SPI), extendable by third parties. Analysis is computed in real-time (3 billion pixels per minute). Tests were made on 5000 single images, 200 NDPI WSI, 100 Aperio SVS WSI converted to the Google Maps format. Conclusions Applications based on our method and libraries can be used upstream, as calibration and quality control tool for the WSI acquisition systems, or as tools to reacquire tiles while the WSI is being scanned. They can also be used downstream to reacquire the complete slides that are below the quality threshold for surgical pathology analysis. WSI may also be displayed in a smarter way by sending and displaying the regions of highest quality before other regions. Such quality assessment scores could be integrated as WSI's metadata shared in clinical, research or teaching contexts, for a more efficient medical informatics workflow. PMID:25565494
Interactive Computing and Processing of NASA Land Surface Observations Using Google Earth Engine
NASA Technical Reports Server (NTRS)
Molthan, Andrew; Burks, Jason; Bell, Jordan
2016-01-01
Google's Earth Engine offers a "big data" approach to processing large volumes of NASA and other remote sensing products. h\\ps://earthengine.google.com/ Interfaces include a Javascript or Python-based API, useful for accessing and processing over large periods of record for Landsat and MODIS observations. Other data sets are frequently added, including weather and climate model data sets, etc. Demonstrations here focus on exploratory efforts to perform land surface change detection related to severe weather, and other disaster events.
GIS Technologies For The New Planetary Science Archive (PSA)
NASA Astrophysics Data System (ADS)
Docasal, R.; Barbarisi, I.; Rios, C.; Macfarlane, A. J.; Gonzalez, J.; Arviset, C.; De Marchi, G.; Martinez, S.; Grotheer, E.; Lim, T.; Besse, S.; Heather, D.; Fraga, D.; Barthelemy, M.
2015-12-01
Geographical information system (GIS) is becoming increasingly used for planetary science. GIS are computerised systems for the storage, retrieval, manipulation, analysis, and display of geographically referenced data. Some data stored in the Planetary Science Archive (PSA), for instance, a set of Mars Express/Venus Express data, have spatial metadata associated to them. To facilitate users in handling and visualising spatial data in GIS applications, the new PSA should support interoperability with interfaces implementing the standards approved by the Open Geospatial Consortium (OGC). These standards are followed in order to develop open interfaces and encodings that allow data to be exchanged with GIS Client Applications, well-known examples of which are Google Earth and NASA World Wind as well as open source tools such as Openlayers. The technology already exists within PostgreSQL databases to store searchable geometrical data in the form of the PostGIS extension. An existing open source maps server is GeoServer, an instance of which has been deployed for the new PSA, uses the OGC standards to allow, among others, the sharing, processing and editing of data and spatial data through the Web Feature Service (WFS) standard as well as serving georeferenced map images through the Web Map Service (WMS). The final goal of the new PSA, being developed by the European Space Astronomy Centre (ESAC) Science Data Centre (ESDC), is to create an archive which enables science exploitation of ESA's planetary missions datasets. This can be facilitated through the GIS framework, offering interfaces (both web GUI and scriptable APIs) that can be used more easily and scientifically by the community, and that will also enable the community to build added value services on top of the PSA.
Masthi, N R Ramesh; Madhusudan, M; Puthussery, Yannick P
2015-11-01
The global positioning system (GPS) technology along with Google Earth is used to measure (spatial map) the accurate distribution of morbidity, mortality and planning of interventions in the community. We used this technology to find out its role in the investigation of a cholera outbreak, and also to identify the cause of the outbreak. This study was conducted in a village near Bengaluru, Karnataka in June 2013 during a cholera outbreak. House-to-house survey was done to identify acute watery diarrhoea cases. A hand held GPS receiver was used to record north and east coordinates of the households of cases and these values were subsequently plotted on Google Earth map. Water samples were collected from suspected sources for microbiological analysis. A total of 27 cases of acute watery diarrhoea were reported. Fifty per cent of cases were in the age group of 14-44 yr and one death was reported. GPS technology and Google Earth described the accurate location of household of cases and spot map generated showed clustering of cases around the suspected water sources. The attack rate was 6.92 per cent and case fatality rate was 3.7 per cent. Water samples collected from suspected sources showed the presence of Vibrio cholera O1 Ogawa. GPS technology and Google Earth were easy to use, helpful to accurately pinpoint the location of household of cases, construction of spot map and follow up of cases. Outbreak was found to be due to contamination of drinking water sources.
NASA Astrophysics Data System (ADS)
Bashev, A.
2012-04-01
Currently there is an enormous amount of various geoscience databases. Unfortunately the only users of the majority of the databases are their elaborators. There are several reasons for that: incompaitability, specificity of tasks and objects and so on. However the main obstacles for wide usage of geoscience databases are complexity for elaborators and complication for users. The complexity of architecture leads to high costs that block the public access. The complication prevents users from understanding when and how to use the database. Only databases, associated with GoogleMaps don't have these drawbacks, but they could be hardly named "geoscience" Nevertheless, open and simple geoscience database is necessary at least for educational purposes (see our abstract for ESSI20/EOS12). We developed a database and web interface to work with them and now it is accessible at maps.sch192.ru. In this database a result is a value of a parameter (no matter which) in a station with a certain position, associated with metadata: the date when the result was obtained; the type of a station (lake, soil etc); the contributor that sent the result. Each contributor has its own profile, that allows to estimate the reliability of the data. The results can be represented on GoogleMaps space image as a point in a certain position, coloured according to the value of the parameter. There are default colour scales and each registered user can create the own scale. The results can be also extracted in *.csv file. For both types of representation one could select the data by date, object type, parameter type, area and contributor. The data are uploaded in *.csv format: Name of the station; Lattitude(dd.dddddd); Longitude(ddd.dddddd); Station type; Parameter type; Parameter value; Date(yyyy-mm-dd). The contributor is recognised while entering. This is the minimal set of features that is required to connect a value of a parameter with a position and see the results. All the complicated data treatment could be conducted in other programs after extraction the filtered data into *.csv file. It makes the database understandable for non-experts. The database employs open data format (*.csv) and wide spread tools: PHP as the program language, MySQL as database management system, JavaScript for interaction with GoogleMaps and JQueryUI for create user interface. The database is multilingual: there are association tables, which connect with elements of the database. In total the development required about 150 hours. The database still has several problems. The main problem is the reliability of the data. Actually it needs an expert system for estimation the reliability, but the elaboration of such a system would take more resources than the database itself. The second problem is the problem of stream selection - how to select the stations that are connected with each other (for example, belong to one water stream) and indicate their sequence. Currently the interface is English and Russian. However it can be easily translated to your language. But some problems we decided. For example problem "the problem of the same station" (sometimes the distance between stations is smaller, than the error of position): when you adding new station to the database our application automatically find station near this place. Also we decided problem of object and parameter type (how to regard "EC" and "electrical conductivity" as the same parameter). This problem has been solved using "associative tables". If you would like to see the interface on your language, just contact us. We should send you the list of terms and phrases for translation on your language. The main advantage of the database is that it is totally open: everybody can see, extract the data from the database and use them for non-commercial purposes with no charge. Registered users can contribute to the database without getting paid. We hope, that it will be widely used first of all for education purposes, but professional scientists could use it also.
The implementation of a modernized Dynamic Digital Map on Gale Crater, Mars
NASA Astrophysics Data System (ADS)
McBeck, J.; Condit, C. D.
2012-12-01
Currently, geology instructors present information to students via PowerPoint, Word, Excel and other programs that are not designed to parse or present geologic data. More tech-savvy, and perhaps better-funded, instructors use Google Earth or ArcGIS to display geologic maps and other visual information. However, Google Earth lacks the ability to present large portions of text, and ArcGIS restricts such functionality to labels and annotations. The original Dynamic Digital Map, which we have renamed Dynamic Digital Map Classic (DDMC), allows instructors to represent both visual and large portions of textual information to students. This summer we generalized the underlying architecture of DDMC, redesigned the user interface, modernized the analytical functionality, renamed the older version and labeled this new creature Dynamic Digital Map Extended (DDME). With the new DDME instructors can showcase maps, images, articles and movies, and create digital field trips. They can set the scale, coordinate system and caption of maps and images, add symbol links to maps and images that can transport the user to any specified destination—either internally (to data contained within the DDME) or externally (to a website address). Instructors and students can also calculate non-linear distances and irregular areas of maps and images, and create digital field trips with any number of stops—complete with notes and driving directions. DDMEs are perhaps best described as a sort of computerized, self-authored, interactive textbook. To display the vast capabilities of DDME, we created a DDME of Gale Crater (DDME-GC), which is the landing site of the most sophisticated NASA Mars Rover—Curiosity. DDME-GC hosts six thematic maps: a detailed geologic map provided by Brad Thompson of the Boston University Center for Remote Sensing (Thompson, et al., 2010), and five maps maintained in ASU's JMARS system, including global mosaics from Mars Global Surveyor's Mars Orbiter Laser Altimeter (MOLA), Mars Odyssey's Thermal Emission Imaging System (THEMIS), and the Mars Digital Image Model. DDME-GC offers a diverse suite of images, with over 40 images captured in the High Resolution Imaging Science Experiment (HiRISE), as well as several global mosaics created from Viking Orbiter, Hubble Telescope, THEMIS, MOLA and HiRISE data. DDME-GC also provides more than 25 articles that span subjects from the possible origins of the mound located in Gale Crater to the goals of NASA's Mars Exploration Program. The movies hosted by DDME-GC describe the difficulties of selecting a landing site for Curiosity, landing Curiosity on Mars and several other dynamic topics. The most significant advantage of the modernized DDME is its easily augmented functionality. In the future, DDME will be able to communicate with databases, import Keyhole Markup Language (KML) files from Google Earth, and be available on iOS and Android operating system. (Imagine: a field trip without the burden of notebooks, pens or pencils, paper or clipboards, with this information maintained on a mobile device.) The most recent DDME is a mere skeleton of its full capabilities—a robust architecture upon which myriad functionality can be supplemented.
NASA Astrophysics Data System (ADS)
Yudono, Adipandang
2017-06-01
Recently, crowd-sourced information is used to produce and improve collective knowledge and community capacity building. Triggered by broadening and expanding access to the Internet and cellular telephones, the utilisation of crowd-sourcing for policy advocacy, e-government and e-participation has increased globally [1]. Crowd-sourced information can conceivably support government’s or general social initiatives to inform, counsel, and cooperate, by engaging subjects and empowering decentralisation and democratization [2]. Crowd-sourcing has turned into a major technique for interactive mapping initiatives by urban or rural community because of its capability to incorporate a wide range of data. Continuously accumulated spatial data can be sorted, layered, and envisioned in ways that even beginners can comprehend with ease. Interactive spatial visualization has the possibility to be a useful democratic planning tool to empower citizens participating in spatial data provision and sharing in government programmes. Since the global emergence of World Wide Web (WWW) technology, the interaction between information providers and users has increased. Local communities are able to produce and share spatial data to produce web interfaces with territorial information in mapping application programming interfaces (APIs) public, such as Google maps, OSM and Wikimapia [3][4][5]. In terms of the democratic spatial planning action, Volunteered Geographic Information (VGI) is considered an effective voluntary method of helping people feel comfortable with the technology and other co-participants in order to shape coalitions of local knowledge. This paper has aim to investigate ‘How is spatial data created by citizens used in Indonesia?’ by discussing the characteristics of spatial data usage by citizens to support spatial policy formulation, starting with the history of participatory mapping to current VGI development in Indonesia.
2013-08-09
CAPE CANAVERAL, Fla. – As seen on Google Maps, the view from the top of the Fixed Service Structure at Launch Complex 39A at NASA's Kennedy Space Center. The FSS, as the structure is known, is 285 feet high and overlooks the Rotating Service Structure that was rolled into place when a space shuttle was at the pad. The path taken by NASA's massive crawler-transporters that carried the shuttle stack 3 miles from Vehicle Assembly Building are also visible leading up to the launch pad. In the distance are seen the launch pads and support structures at Cape Canaveral Air Force Station for the Atlas V, Delta IV and Falcon 9 rockets. Google precisely mapped the space center and some of its historical facilities for the company's map page. The work allows Internet users to see inside buildings at Kennedy as they were used during the space shuttle era. Photo credit: Google/Wendy Wang
2013-08-09
CAPE CANAVERAL, Fla. – As seen on Google Maps, space shuttle Endeavour goes through transition and retirement processing in high bay 4 of the Vehicle Assembly Building at NASA's Kennedy Space Center. The spacecraft completed 25 missions beginning with its first flight, STS-49, in May 1992, and ending with STS-134 in May 2011. It helped construct the International Space Station in orbit and travelled more than 122 million miles in orbit during its career. The reaction control system pods in the shuttle's nose and aft section were removed for processing before Endeavour was put on public display at the California Science Center in Los Angeles. Google precisely mapped the space center and some of its historical facilities for the company's map page. The work allows Internet users to see inside buildings at Kennedy as they were used during the space shuttle era. Photo credit: Google/Wendy Wang
Southern California Earthquake Center Geologic Vertical Motion Database
NASA Astrophysics Data System (ADS)
Niemi, Nathan A.; Oskin, Michael; Rockwell, Thomas K.
2008-07-01
The Southern California Earthquake Center Geologic Vertical Motion Database (VMDB) integrates disparate sources of geologic uplift and subsidence data at 104- to 106-year time scales into a single resource for investigations of crustal deformation in southern California. Over 1800 vertical deformation rate data points in southern California and northern Baja California populate the database. Four mature data sets are now represented: marine terraces, incised river terraces, thermochronologic ages, and stratigraphic surfaces. An innovative architecture and interface of the VMDB exposes distinct data sets and reference frames, permitting user exploration of this complex data set and allowing user control over the assumptions applied to convert geologic and geochronologic information into absolute uplift rates. Online exploration and download tools are available through all common web browsers, allowing the distribution of vertical motion results as HTML tables, tab-delimited GIS-compatible text files, or via a map interface through the Google Maps™ web service. The VMDB represents a mature product for research of fault activity and elastic deformation of southern California.
Assessing species habitat using Google Street View: a case study of cliff-nesting vultures.
Olea, Pedro P; Mateo-Tomás, Patricia
2013-01-01
The assessment of a species' habitat is a crucial issue in ecology and conservation. While the collection of habitat data has been boosted by the availability of remote sensing technologies, certain habitat types have yet to be collected through costly, on-ground surveys, limiting study over large areas. Cliffs are ecosystems that provide habitat for a rich biodiversity, especially raptors. Because of their principally vertical structure, however, cliffs are not easy to study by remote sensing technologies, posing a challenge for many researches and managers working with cliff-related biodiversity. We explore the feasibility of Google Street View, a freely available on-line tool, to remotely identify and assess the nesting habitat of two cliff-nesting vultures (the griffon vulture and the globally endangered Egyptian vulture) in northwestern Spain. Two main usefulness of Google Street View to ecologists and conservation biologists were evaluated: i) remotely identifying a species' potential habitat and ii) extracting fine-scale habitat information. Google Street View imagery covered 49% (1,907 km) of the roads of our study area (7,000 km²). The potential visibility covered by on-ground surveys was significantly greater (mean: 97.4%) than that of Google Street View (48.1%). However, incorporating Google Street View to the vulture's habitat survey would save, on average, 36% in time and 49.5% in funds with respect to the on-ground survey only. The ability of Google Street View to identify cliffs (overall accuracy = 100%) outperformed the classification maps derived from digital elevation models (DEMs) (62-95%). Nonetheless, high-performance DEM maps may be useful to compensate Google Street View coverage limitations. Through Google Street View we could examine 66% of the vultures' nesting-cliffs existing in the study area (n = 148): 64% from griffon vultures and 65% from Egyptian vultures. It also allowed us the extraction of fine-scale features of cliffs. This World Wide Web-based methodology may be a useful, complementary tool to remotely map and assess the potential habitat of cliff-dependent biodiversity over large geographic areas, saving survey-related costs.
Assessing Species Habitat Using Google Street View: A Case Study of Cliff-Nesting Vultures
Olea, Pedro P.; Mateo-Tomás, Patricia
2013-01-01
The assessment of a species’ habitat is a crucial issue in ecology and conservation. While the collection of habitat data has been boosted by the availability of remote sensing technologies, certain habitat types have yet to be collected through costly, on-ground surveys, limiting study over large areas. Cliffs are ecosystems that provide habitat for a rich biodiversity, especially raptors. Because of their principally vertical structure, however, cliffs are not easy to study by remote sensing technologies, posing a challenge for many researches and managers working with cliff-related biodiversity. We explore the feasibility of Google Street View, a freely available on-line tool, to remotely identify and assess the nesting habitat of two cliff-nesting vultures (the griffon vulture and the globally endangered Egyptian vulture) in northwestern Spain. Two main usefulness of Google Street View to ecologists and conservation biologists were evaluated: i) remotely identifying a species’ potential habitat and ii) extracting fine-scale habitat information. Google Street View imagery covered 49% (1,907 km) of the roads of our study area (7,000 km2). The potential visibility covered by on-ground surveys was significantly greater (mean: 97.4%) than that of Google Street View (48.1%). However, incorporating Google Street View to the vulture’s habitat survey would save, on average, 36% in time and 49.5% in funds with respect to the on-ground survey only. The ability of Google Street View to identify cliffs (overall accuracy = 100%) outperformed the classification maps derived from digital elevation models (DEMs) (62–95%). Nonetheless, high-performance DEM maps may be useful to compensate Google Street View coverage limitations. Through Google Street View we could examine 66% of the vultures’ nesting-cliffs existing in the study area (n = 148): 64% from griffon vultures and 65% from Egyptian vultures. It also allowed us the extraction of fine-scale features of cliffs. This World Wide Web-based methodology may be a useful, complementary tool to remotely map and assess the potential habitat of cliff-dependent biodiversity over large geographic areas, saving survey-related costs. PMID:23355880
A method for vreating a three dimensional model from published geologic maps and cross sections
Walsh, Gregory J.
2009-01-01
This brief report presents a relatively inexpensive and rapid method for creating a 3D model of geology from published quadrangle-scale maps and cross sections using Google Earth and Google SketchUp software. An example from the Green Mountains of Vermont, USA, is used to illustrate the step by step methods used to create such a model. A second example is provided from the Jebel Saghro region of the Anti-Atlas Mountains of Morocco. The report was published to help enhance the public?s ability to use and visualize geologic map data.
The impact of geo-tagging on the photo industry and creating revenue streams
NASA Astrophysics Data System (ADS)
Richter, Rolf; Böge, Henning; Weckmann, Christoph; Schloen, Malte
2010-02-01
Internet geo and mapping services like Google Maps, Google Earth and Microsoft Bing Maps have reinvented the use of geographical information and have reached an enormous popularity. Besides that, location technologies like GPS have become affordable and are now being integrated in many camera phones. GPS is also available for standalone cameras as add on products or integrated in cameras. These developments are the enabler for new products for the photo industry or they enhance existing products. New commercial opportunities have been identified in the areas of photo hardware, internet/software and photo finishing.
Route Sanitizer: Connected Vehicle Trajectory De-Identification Tool
DOE Office of Scientific and Technical Information (OSTI.GOV)
Carter, Jason M; Ferber, Aaron E
Route Sanitizer is ORNL's connected vehicle moving object database de-identification tool and a graphical user interface to ORNL's connected vehicle de-identification algorithm. It uses the Google Chrome (soon to be Electron) platform so it will run on different computing platforms. The basic de-identification strategy is record redaction: portions of a vehicle trajectory (e.g. sequences of precise temporal spatial records) are removed. It does not alter retained records. The algorithm uses custom techniques to find areas within trajectories that may be considered private, then it suppresses those in addition to enough of the trajectory surrounding those locations to protect against "inferencemore » attacks" in a mathematically sound way. Map data is integrated into the process to make this possible.« less
Caltrans - California Department of Transportation
Caltrans QuickMap QuickMap Mobile QuickMap Android App Check Current Highway Conditions: Enter Highway the App Store. Google Play Apple Store Quickmap Mobile Version Quickmap Full Version CA Safety
Li, Ya-pin; Fang, Li-qun; Gao, Su-qing; Wang, Zhen; Gao, Hong-wei; Liu, Peng; Wang, Ze-Rui; Li, Yan-Li; Zhu, Xu-Guang; Li, Xin-Lou; Xu, Bo; Li, Yin-Jun; Yang, Hong; de Vlas, Sake J; Shi, Tao-Xing; Cao, Wu-Chun
2013-01-01
For years, emerging infectious diseases have appeared worldwide and threatened the health of people. The emergence and spread of an infectious-disease outbreak are usually unforeseen, and have the features of suddenness and uncertainty. Timely understanding of basic information in the field, and the collection and analysis of epidemiological information, is helpful in making rapid decisions and responding to an infectious-disease emergency. Therefore, it is necessary to have an unobstructed channel and convenient tool for the collection and analysis of epidemiologic information in the field. Baseline information for each county in mainland China was collected and a database was established by geo-coding information on a digital map of county boundaries throughout the country. Google Maps was used to display geographic information and to conduct calculations related to maps, and the 3G wireless network was used to transmit information collected in the field to the server. This study established a decision support system for the response to infectious-disease emergencies based on WebGIS and mobile services (DSSRIDE). The DSSRIDE provides functions including data collection, communication and analyses in real time, epidemiological detection, the provision of customized epidemiological questionnaires and guides for handling infectious disease emergencies, and the querying of professional knowledge in the field. These functions of the DSSRIDE could be helpful for epidemiological investigations in the field and the handling of infectious-disease emergencies. The DSSRIDE provides a geographic information platform based on the Google Maps application programming interface to display information of infectious disease emergencies, and transfers information between workers in the field and decision makers through wireless transmission based on personal computers, mobile phones and personal digital assistants. After a 2-year practice and application in infectious disease emergencies, the DSSRIDE is becoming a useful platform and is a useful tool for investigations in the field carried out by response sections and individuals. The system is suitable for use in developing countries and low-income districts.
Muellner, Ulrich J; Vial, Flavie; Wohlfender, Franziska; Hadorn, Daniela; Reist, Martin; Muellner, Petra
2015-01-01
The reporting of outputs from health surveillance systems should be done in a near real-time and interactive manner in order to provide decision makers with powerful means to identify, assess, and manage health hazards as early and efficiently as possible. While this is currently rarely the case in veterinary public health surveillance, reporting tools do exist for the visual exploration and interactive interrogation of health data. In this work, we used tools freely available from the Google Maps and Charts library to develop a web application reporting health-related data derived from slaughterhouse surveillance and from a newly established web-based equine surveillance system in Switzerland. Both sets of tools allowed entry-level usage without or with minimal programing skills while being flexible enough to cater for more complex scenarios for users with greater programing skills. In particular, interfaces linking statistical softwares and Google tools provide additional analytical functionality (such as algorithms for the detection of unusually high case occurrences) for inclusion in the reporting process. We show that such powerful approaches could improve timely dissemination and communication of technical information to decision makers and other stakeholders and could foster the early-warning capacity of animal health surveillance systems.
A Web-Based Information System for Field Data Management
NASA Astrophysics Data System (ADS)
Weng, Y. H.; Sun, F. S.
2014-12-01
A web-based field data management system has been designed and developed to allow field geologists to store, organize, manage, and share field data online. System requirements were analyzed and clearly defined first regarding what data are to be stored, who the potential users are, and what system functions are needed in order to deliver the right data in the right way to the right user. A 3-tiered architecture was adopted to create this secure, scalable system that consists of a web browser at the front end while a database at the back end and a functional logic server in the middle. Specifically, HTML, CSS, and JavaScript were used to implement the user interface in the front-end tier, the Apache web server runs PHP scripts, and MySQL to server is used for the back-end database. The system accepts various types of field information, including image, audio, video, numeric, and text. It allows users to select data and populate them on either Google Earth or Google Maps for the examination of the spatial relations. It also makes the sharing of field data easy by converting them into XML format that is both human-readable and machine-readable, and thus ready for reuse.
IRIS Earthquake Browser with Integration to the GEON IDV for 3-D Visualization of Hypocenters.
NASA Astrophysics Data System (ADS)
Weertman, B. R.
2007-12-01
We present a new generation of web based earthquake query tool - the IRIS Earthquake Browser (IEB). The IEB combines the DMC's large set of earthquake catalogs (provided by USGS/NEIC, ISC and the ANF) with the popular Google Maps web interface. With the IEB you can quickly and easily find earthquakes in any region of the globe. Using Google's detailed satellite images, earthquakes can be easily co-located with natural geographic features such as volcanoes as well as man made features such as commercial mines. A set of controls allow earthquakes to be filtered by time, magnitude, and depth range as well as catalog name, contributor name and magnitude type. Displayed events can be easily exported in NetCDF format into the GEON Integrated Data Viewer (IDV) where hypocenters may be visualized in three dimensions. Looking "under the hood", the IEB is based on AJAX technology and utilizes REST style web services hosted at the IRIS DMC. The IEB is part of a broader effort at the DMC aimed at making our data holdings available via web services. The IEB is useful both educationally and as a research tool.
Moving Forward: The Next-Gen Catalog and the New Discovery Tools
ERIC Educational Resources Information Center
Weare, William H., Jr.; Toms, Sue; Breeding, Marshall
2011-01-01
Do students prefer to use Google instead of the library catalog? Ever wondered why? Google is easier to use and delivers plenty of "good enough" resources to meet their needs. The current generation of online catalogs has two main problems. First, the look and feel of the interface doesn't reflect the conventions adhered to elsewhere on the web,…
Development of a Google-based search engine for data mining radiology reports.
Erinjeri, Joseph P; Picus, Daniel; Prior, Fred W; Rubin, David A; Koppel, Paul
2009-08-01
The aim of this study is to develop a secure, Google-based data-mining tool for radiology reports using free and open source technologies and to explore its use within an academic radiology department. A Health Insurance Portability and Accountability Act (HIPAA)-compliant data repository, search engine and user interface were created to facilitate treatment, operations, and reviews preparatory to research. The Institutional Review Board waived review of the project, and informed consent was not required. Comprising 7.9 GB of disk space, 2.9 million text reports were downloaded from our radiology information system to a fileserver. Extensible markup language (XML) representations of the reports were indexed using Google Desktop Enterprise search engine software. A hypertext markup language (HTML) form allowed users to submit queries to Google Desktop, and Google's XML response was interpreted by a practical extraction and report language (PERL) script, presenting ranked results in a web browser window. The query, reason for search, results, and documents visited were logged to maintain HIPAA compliance. Indexing averaged approximately 25,000 reports per hour. Keyword search of a common term like "pneumothorax" yielded the first ten most relevant results of 705,550 total results in 1.36 s. Keyword search of a rare term like "hemangioendothelioma" yielded the first ten most relevant results of 167 total results in 0.23 s; retrieval of all 167 results took 0.26 s. Data mining tools for radiology reports will improve the productivity of academic radiologists in clinical, educational, research, and administrative tasks. By leveraging existing knowledge of Google's interface, radiologists can quickly perform useful searches.
Sally Ride EarthKAM - Automated Image Geo-Referencing Using Google Earth Web Plug-In
NASA Technical Reports Server (NTRS)
Andres, Paul M.; Lazar, Dennis K.; Thames, Robert Q.
2013-01-01
Sally Ride EarthKAM is an educational program funded by NASA that aims to provide the public the ability to picture Earth from the perspective of the International Space Station (ISS). A computer-controlled camera is mounted on the ISS in a nadir-pointing window; however, timing limitations in the system cause inaccurate positional metadata. Manually correcting images within an orbit allows the positional metadata to be improved using mathematical regressions. The manual correction process is time-consuming and thus, unfeasible for a large number of images. The standard Google Earth program allows for the importing of KML (keyhole markup language) files that previously were created. These KML file-based overlays could then be manually manipulated as image overlays, saved, and then uploaded to the project server where they are parsed and the metadata in the database is updated. The new interface eliminates the need to save, download, open, re-save, and upload the KML files. Everything is processed on the Web, and all manipulations go directly into the database. Administrators also have the control to discard any single correction that was made and validate a correction. This program streamlines a process that previously required several critical steps and was probably too complex for the average user to complete successfully. The new process is theoretically simple enough for members of the public to make use of and contribute to the success of the Sally Ride EarthKAM project. Using the Google Earth Web plug-in, EarthKAM images, and associated metadata, this software allows users to interactively manipulate an EarthKAM image overlay, and update and improve the associated metadata. The Web interface uses the Google Earth JavaScript API along with PHP-PostgreSQL to present the user the same interface capabilities without leaving the Web. The simpler graphical user interface will allow the public to participate directly and meaningfully with EarthKAM. The use of similar techniques is being investigated to place ground-based observations in a Google Mars environment, allowing the MSL (Mars Science Laboratory) Science Team a means to visualize the rover and its environment.
Fazeli Dehkordy, Soudabeh; Carlos, Ruth C; Hall, Kelli S; Dalton, Vanessa K
2014-09-01
Millions of people use online search engines everyday to find health-related information and voluntarily share their personal health status and behaviors in various Web sites. Thus, data from tracking of online information seeker's behavior offer potential opportunities for use in public health surveillance and research. Google Trends is a feature of Google which allows Internet users to graph the frequency of searches for a single term or phrase over time or by geographic region. We used Google Trends to describe patterns of information-seeking behavior in the subject of dense breasts and to examine their correlation with the passage or introduction of dense breast notification legislation. To capture the temporal variations of information seeking about dense breasts, the Web search query "dense breast" was entered in the Google Trends tool. We then mapped the dates of legislative actions regarding dense breasts that received widespread coverage in the lay media to information-seeking trends about dense breasts over time. Newsworthy events and legislative actions appear to correlate well with peaks in search volume of "dense breast". Geographic regions with the highest search volumes have passed, denied, or are currently considering the dense breast legislation. Our study demonstrated that any legislative action and respective news coverage correlate with increase in information seeking for "dense breast" on Google, suggesting that Google Trends has the potential to serve as a data source for policy-relevant research. Copyright © 2014 AUR. Published by Elsevier Inc. All rights reserved.
Google's Geo Education Outreach: Results and Discussion of Outreach Trip to Alaskan High Schools.
NASA Astrophysics Data System (ADS)
Kolb, E. J.; Bailey, J.; Bishop, A.; Cain, J.; Goddard, M.; Hurowitz, K.; Kennedy, K.; Ornduff, T.; Sfraga, M.; Wernecke, J.
2008-12-01
The focus of Google's Geo Education outreach efforts (http://www.google.com/educators/geo.html) is on helping primary, secondary, and post-secondary educators incorporate Google Earth and Sky, Google Maps, and SketchUp into their classroom lessons. In partnership with the University of Alaska, our Geo Education team members visited several remote Alaskan high schools during a one-week period in September. At each school, we led several 40-minute hands-on learning sessions in which Google products were used by the students to investigate local geologic and environmental processes. For the teachers, we provided several resources including follow-on lesson plans, example KML-based lessons, useful URL's, and website resources that multiple users can contribute to. This talk will highlight results of the trip and discuss how educators can access and use Google's Geo Education resources.
Geospatial Data Science Applications and Visualizations | Geospatial Data
. Since before the time of Google Maps, NREL has used the internet to allow stakeholders to view and world, these maps drive understanding. See our collection of key maps for examples. Featured Analysis
2013-08-09
CAPE CANAVERAL, Fla. – As seen on Google Maps, Firing Room 3 inside the Launch Control Center at NASA's Kennedy Space Center was one of the four control rooms used by NASA and contractor launch teams to oversee a space shuttle countdown. This firing room is furnished in the classic style with the same metal computer cabinets and some of the same monitors in place when the first shuttle mission launched April 12, 1981. Specialized operators worked at consoles tailored to keep track of the status of shuttle systems while the spacecraft was processed in the Orbiter Processing Facility, being stacked inside the Vehicle Assembly Building and standing at the launch pad before liftoff. The firing rooms, including 3, were also used during NASA's Apollo Program. Google precisely mapped the space center and some of its historical facilities for the company's map page. The work allows Internet users to see inside buildings at Kennedy as they were used during the space shuttle era. Photo credit: Google/Wendy Wang
An interactive GIS based tool on Chinese history and its topography
NASA Astrophysics Data System (ADS)
Konda, Ashish Reddy
The aim of the thesis is to demonstrate how China was attacked by the foreign powers, the rise and fall of the empires, the border conflicts with India, Russia, Vietnam and territorial disputes in South China Sea. This thesis is focused on creating a GIS tool showcasing the modern Chinese history, which includes the major wars fought during that period. This tool is developed using the features of Google Maps that shows the location of the wars. The topography of China is also represented on the interactive Google Map by creating layers for rivers, mountain ranges and deserts. The provinces with highest population are also represented on the Google Map with circles. The application also shows the historical events in chronological order using a timeline feature. This has been implemented using JQuery, JavaScript, HTML5 and CSS. Chinese culture and biographies of important leaders are also included in this thesis, which is embedded with pictures and videos.
Interactive web-based mapping: bridging technology and data for health.
Highfield, Linda; Arthasarnprasit, Jutas; Ottenweller, Cecelia A; Dasprez, Arnaud
2011-12-23
The Community Health Information System (CHIS) online mapping system was first launched in 1998. Its overarching goal was to provide researchers, residents and organizations access to health related data reflecting the overall health and well-being of their communities within the Greater Houston area. In September 2009, initial planning and development began for the next generation of CHIS. The overarching goal for the new version remained to make health data easily accessible for a wide variety of research audiences. However, in the new version we specifically sought to make the CHIS truly interactive and give the user more control over data selection and reporting. In July 2011, a beta version of the next-generation of the application was launched. This next-generation is also a web based interactive mapping tool comprised of two distinct portals: the Breast Health Portal and Project Safety Net. Both are accessed via a Google mapping interface. Geographic coverage for the portals is currently an 8 county region centered on Harris County, Texas. Data accessed by the application include Census 2000, Census 2010 (underway), cancer incidence from the Texas Cancer Registry (TX Dept. of State Health Services), death data from Texas Vital Statistics, clinic locations for free and low-cost health services, along with service lists, hours of operation, payment options and languages spoken, uninsured and poverty data. The system features query on the fly technology, which means the data is not generated until the query is provided to the system. This allows users to interact in real-time with the databases and generate customized reports and maps. To the author's knowledge, the Breast Health Portal and Project Safety Net are the first local-scale interactive online mapping interfaces for public health data which allow users to control the data generated. For example, users may generate breast cancer incidence rates by Census tract, in real time, for women aged 40-64. Conversely, they could then generate the same rates for women aged 35-55. The queries are user controlled.
Texting while driving using Google Glass™: Promising but not distraction-free.
He, Jibo; Choi, William; McCarley, Jason S; Chaparro, Barbara S; Wang, Chun
2015-08-01
Texting while driving is risky but common. This study evaluated how texting using a Head-Mounted Display, Google Glass, impacts driving performance. Experienced drivers performed a classic car-following task while using three different interfaces to text: fully manual interaction with a head-down smartphone, vocal interaction with a smartphone, and vocal interaction with Google Glass. Fully manual interaction produced worse driving performance than either of the other interaction methods, leading to more lane excursions and variable vehicle control, and higher workload. Compared to texting vocally with a smartphone, texting using Google Glass produced fewer lane excursions, more braking responses, and lower workload. All forms of texting impaired driving performance compared to undistracted driving. These results imply that the use of Google Glass for texting impairs driving, but its Head-Mounted Display configuration and speech recognition technology may be safer than texting using a smartphone. Copyright © 2015 Elsevier Ltd. All rights reserved.
Google glass-based remote control of a mobile robot
NASA Astrophysics Data System (ADS)
Yu, Song; Wen, Xi; Li, Wei; Chen, Genshe
2016-05-01
In this paper, we present an approach to remote control of a mobile robot via a Google Glass with the multi-function and compact size. This wearable device provides a new human-machine interface (HMI) to control a robot without need for a regular computer monitor because the Google Glass micro projector is able to display live videos around robot environments. In doing it, we first develop a protocol to establish WI-FI connection between Google Glass and a robot and then implement five types of robot behaviors: Moving Forward, Turning Left, Turning Right, Taking Pause, and Moving Backward, which are controlled by sliding and clicking the touchpad located on the right side of the temple. In order to demonstrate the effectiveness of the proposed Google Glass-based remote control system, we navigate a virtual Surveyor robot to pass a maze. Experimental results demonstrate that the proposed control system achieves the desired performance.
NASA Astrophysics Data System (ADS)
Farda, N. M.
2017-12-01
Coastal wetlands provide ecosystem services essential to people and the environment. Changes in coastal wetlands, especially on land use, are important to monitor by utilizing multi-temporal imagery. The Google Earth Engine (GEE) provides many machine learning algorithms (10 algorithms) that are very useful for extracting land use from imagery. The research objective is to explore machine learning in Google Earth Engine and its accuracy for multi-temporal land use mapping of coastal wetland area. Landsat 3 MSS (1978), Landsat 5 TM (1991), Landsat 7 ETM+ (2001), and Landsat 8 OLI (2014) images located in Segara Anakan lagoon are selected to represent multi temporal images. The input for machine learning are visible and near infrared bands, PCA band, invers PCA bands, bare soil index, vegetation index, wetness index, elevation from ASTER GDEM, and GLCM (Harralick) texture, and also polygon samples in 140 locations. There are 10 machine learning algorithms applied to extract coastal wetlands land use from Landsat imagery. The algorithms are Fast Naive Bayes, CART (Classification and Regression Tree), Random Forests, GMO Max Entropy, Perceptron (Multi Class Perceptron), Winnow, Voting SVM, Margin SVM, Pegasos (Primal Estimated sub-GrAdient SOlver for Svm), IKPamir (Intersection Kernel Passive Aggressive Method for Information Retrieval, SVM). Machine learning in Google Earth Engine are very helpful in multi-temporal land use mapping, the highest accuracy for land use mapping of coastal wetland is CART with 96.98 % Overall Accuracy using K-Fold Cross Validation (K = 10). GEE is particularly useful for multi-temporal land use mapping with ready used image and classification algorithms, and also very challenging for other applications.
3D Viewer Platform of Cloud Clustering Management System: Google Map 3D
NASA Astrophysics Data System (ADS)
Choi, Sung-Ja; Lee, Gang-Soo
The new management system of framework for cloud envrionemnt is needed by the platfrom of convergence according to computing environments of changes. A ISV and small business model is hard to adapt management system of platform which is offered from super business. This article suggest the clustering management system of cloud computing envirionments for ISV and a man of enterprise in small business model. It applies the 3D viewer adapt from map3D & earth of google. It is called 3DV_CCMS as expand the CCMS[1].
Application of neogeographic tools for geochemistry
NASA Astrophysics Data System (ADS)
Zhilin, Denis
2010-05-01
Neogeography is a usage of geographical tools for utilization by a non-expert group of users. It have been rapidly developing last ten years and is founded on (a) availability of Global Positioning System (GPS) receivers, that allows to obtain very precise geographical position (b) services, that allows linking geographical position with satellite images, GoogleEarth for example and (c) programs as GPS Track Maker or OziExplorer, that allows linking geographical coordinates with other raster images (for example, maps). However, the possibilities of neogeographic approach are much wider. It allows linking different parameters with geographical coordinates on the one hand and space image or map - on the other. If it is easy to measure a parameter, a great database could be collected for a very small time. The results can be presented in very different ways. One can plot a parameter versus the distance from a particular point (for example, a source of a substance), make two-dimension distribution of parameter of put the results onto a map or space image. In the case of chemical parameters it can help finding the source of pollution, trace the influence of pollution, reveal geochemical processes and patterns. The main advantage of neogeograpic approach is the employment of non-experts in collecting data. Now non-experts can easily measure electrical conductivity and pH of natural waters, concentration of different gases in the atmosphere, solar irradiation, radioactivity and so on. If the results are obtained (for example, by students of secondary schools) and shared, experts can proceed them and make significant conclusions. An interface of sharing the results (http://maps.sch192.ru/) was elaborated by V. Ilyin. Within the interface a user can load *.csv file with coordinates, type of parameter and the value of parameter in a particular point. The points are marked on the GoogleEarth map with the color corresponding the value of the parameter. The color scale can be edited manually. We would like to show some results of practical and scientific importance, obtained by non-experts. At 2006 our secondary school students investigated the distribution of snow salinity around Kosygina Street in Moscow. One can conclude that the distribution of salinity is reproducible and that the street influences the snow up to 150 meters. Another example obtained by our students is the distribution of electrical conductivity of swamp water showing extreme irregularity of this parameter within the small area (about 0.5x0.5 km) the electrical conductivity varied from 22 to 77 uS with no regularity. It points out the key role of local processes in swamp water chemistry. The third example (maps of electrical conductivity and pH of water on a large area) one can see at http://fenevo.narod.ru/maps/ec-maps.htm and http://fenevo.narod.ru/maps/ph-maps.htm. Basing on the map one can conclude mechanisms of formation of water mineralization in the area. Availability of GPS receivers and systems for easy measuring of chemical parameters can lead to neogeochemical revolution as GPS receivers have led to neogeographical. A great number of non-experts can share their geochemical results, forming huge amount of available geochemical data. It will help to falsify and visualize concepts of geochemistry and environmental chemistry and, maybe, develop new ones. Geophysical and biological data could be shared as well with the same advantages for corresponding sciences.
Rattanasing, Wannaporn; Kaewpitoon, Soraya J; Loyd, Ryan A; Rujirakul, Ratana; Yodkaw, Eakachai; Kaewpitoon, Natthawut
2015-01-01
Cholangiocarcinoma (CCA) is a serious public health problem in the Northeast of Thailand. CCA is considered to be an incurable and rapidly lethal disease. Knowledge of the distribution of CCA patients is necessary for management strategies. This study aimed to utilize the Geographic Information System and Google EarthTM for distribution mapping of cholangiocarcinoma in Satuek District, Buriram, Thailand, during a 5-year period (2008-2012). In this retrospective study data were collected and reviewed from the OPD cards, definitive cases of CCA were patients who were treated in Satuek hospital and were diagnosed with CCA or ICD-10 code C22.1. CCA cases were used to analyze and calculate with ArcGIS 9.2, all of data were imported into Google Earth using the online web page www.earthpoint.us. Data were displayed at village points. A total of 53 cases were diagnosed and identified as CCA. The incidence was 53.57 per 100,000 population (65.5 for males and 30.8 for females) and the majority of CCA cases were in stages IV and IIA. The average age was 67 years old. The highest attack rate was observed in Thung Wang sub-district (161.4 per 100,000 population). The map display at village points for CCA patients based on Google Earth gave a clear visual deistribution. CCA is still a major problem in Satuek district, Buriram province of Thailand. The Google Earth production process is very simple and easy to learn. It is suitable for the use in further development of CCA management strategies.
Dehkordy, Soudabeh Fazeli; Carlos, Ruth C.; Hall, Kelli S.; Dalton, Vanessa K.
2015-01-01
Rationale and Objectives Millions of people use online search engines every day to find health-related information and voluntarily share their personal health status and behaviors in various Web sites. Thus, data from tracking of online information seeker’s behavior offer potential opportunities for use in public health surveillance and research. Google Trends is a feature of Google which allows internet users to graph the frequency of searches for a single term or phrase over time or by geographic region. We used Google Trends to describe patterns of information seeking behavior in the subject of dense breasts and to examine their correlation with the passage or introduction of dense breast notification legislation. Materials and Methods In order to capture the temporal variations of information seeking about dense breasts, the web search query “dense breast” was entered in the Google Trends tool. We then mapped the dates of legislative actions regarding dense breasts that received widespread coverage in the lay media to information seeking trends about dense breasts over time. Results Newsworthy events and legislative actions appear to correlate well with peaks in search volume of “dense breast”. Geographic regions with the highest search volumes have either passed, denied, or are currently considering the dense breast legislation. Conclusions Our study demonstrated that any legislative action and respective news coverage correlate with increase in information seeking for “dense breast” on Google, suggesting that Google Trends has the potential to serve as a data source for policy-relevant research. PMID:24998689
Google glass based immunochromatographic diagnostic test analysis
NASA Astrophysics Data System (ADS)
Feng, Steve; Caire, Romain; Cortazar, Bingen; Turan, Mehmet; Wong, Andrew; Ozcan, Aydogan
2015-03-01
Integration of optical imagers and sensors into recently emerging wearable computational devices allows for simpler and more intuitive methods of integrating biomedical imaging and medical diagnostics tasks into existing infrastructures. Here we demonstrate the ability of one such device, the Google Glass, to perform qualitative and quantitative analysis of immunochromatographic rapid diagnostic tests (RDTs) using a voice-commandable hands-free software-only interface, as an alternative to larger and more bulky desktop or handheld units. Using the built-in camera of Glass to image one or more RDTs (labeled with Quick Response (QR) codes), our Glass software application uploads the captured image and related information (e.g., user name, GPS, etc.) to our servers for remote analysis and storage. After digital analysis of the RDT images, the results are transmitted back to the originating Glass device, and made available through a website in geospatial and tabular representations. We tested this system on qualitative human immunodeficiency virus (HIV) and quantitative prostate-specific antigen (PSA) RDTs. For qualitative HIV tests, we demonstrate successful detection and labeling (i.e., yes/no decisions) for up to 6-fold dilution of HIV samples. For quantitative measurements, we activated and imaged PSA concentrations ranging from 0 to 200 ng/mL and generated calibration curves relating the RDT line intensity values to PSA concentration. By providing automated digitization of both qualitative and quantitative test results, this wearable colorimetric diagnostic test reader platform on Google Glass can reduce operator errors caused by poor training, provide real-time spatiotemporal mapping of test results, and assist with remote monitoring of various biomedical conditions.
NASA Astrophysics Data System (ADS)
Lilly, M. R.; Feditova, A.; Levine, K.; Giardino, J. R.
2017-12-01
The Harris County Flood Control District has an impressive amount of information available for the public related to flood management and response. During Hurricane Harvey, this information was used by the authors to help address daily questions from family and friends living in the Houston area. Common near-real-time reporting data included precipitation and water levels. Maps included locations of data stations, stream or bayou conditions (in bank, out of bank) and watershed or drainage boundaries. In general, the data station reporting and online information was updating well throughout the hurricane and post-flooding period. Only a few of the data reporting stations had problems with water level sensor measurements. The overall information was helpful to hydrologists and floodplain managers. The online information could not easily answer all common questions residents may have during a flood event. Some of the more common questions were how to use the water-level information to know the potential extent of flooding and relative location of flooding to the location of residents. To help address the questions raised during the flooding on how to use the available water level data, we used Google Earth to get lot and intersection locations to help show the relative differences between nearby water-level stations and residences of interest. The reported resolution of the Google Earth elevation data is 1-foot. To help confirm the use of this data, we compared Google Earth approximate elevations with reported Harris County Floodplain Reference Mark individual reports. This method helped verify we could use the Google Earth information for approximate comparisons. We also faced questions on what routes to take if evacuation was needed, and where to go to get to higher ground elevations. Google Earth again provided a helpful and easy to use interface to look at road and intersection elevations and develop suggested routes for family and friends to take to avoid low areas that may be subject to flooding. These and other recommendations that helped answer common questions by residents reacting to the hurricane and subsequent flooding conditions are summarized with examples.
Marek, Lukáš; Tuček, Pavel; Pászto, Vít
2015-01-28
Visual analytics aims to connect the processing power of information technologies and the user's ability of logical thinking and reasoning through the complex visual interaction. Moreover, the most of the data contain the spatial component. Therefore, the need for geovisual tools and methods arises. Either one can develop own system but the dissemination of findings and its usability might be problematic or the widespread and well-known platform can be utilized. The aim of this paper is to prove the applicability of Google Earth™ software as a tool for geovisual analytics that helps to understand the spatio-temporal patterns of the disease distribution. We combined the complex joint spatio-temporal analysis with comprehensive visualisation. We analysed the spatio-temporal distribution of the campylobacteriosis in the Czech Republic between 2008 and 2012. We applied three main approaches in the study: (1) the geovisual analytics of the surveillance data that were visualised in the form of bubble chart; (2) the geovisual analytics of the disease's weekly incidence surfaces computed by spatio-temporal kriging and (3) the spatio-temporal scan statistics that was employed in order to identify high or low rates clusters of affected municipalities. The final data are stored in Keyhole Markup Language files and visualised in Google Earth™ in order to apply geovisual analytics. Using geovisual analytics we were able to display and retrieve information from complex dataset efficiently. Instead of searching for patterns in a series of static maps or using numerical statistics, we created the set of interactive visualisations in order to explore and communicate results of analyses to the wider audience. The results of the geovisual analytics identified periodical patterns in the behaviour of the disease as well as fourteen spatio-temporal clusters of increased relative risk. We prove that Google Earth™ software is a usable tool for the geovisual analysis of the disease distribution. Google Earth™ has many indisputable advantages (widespread, freely available, intuitive interface, space-time visualisation capabilities and animations, communication of results), nevertheless it is still needed to combine it with pre-processing tools that prepare the data into a form suitable for the geovisual analytics itself.
The World in Spatial Terms: Mapmaking and Map Reading
ERIC Educational Resources Information Center
Ekiss, Gale Olp; Trapido-Lurie, Barbara; Phillips, Judy; Hinde, Elizabeth
2007-01-01
Maps and mapping activities are essential in the primary grades. Maps are truly ubiquitous today, as evidenced by the popularity of websites such as Google Earth and Mapquest, and by devices such as Global Positioning System (GPS) units in cars, planes, and boats. Maps can give visual settings to travel stories and historical narratives and can…
Building a Dashboard of the Planet with Google Earth and Earth Engine
NASA Astrophysics Data System (ADS)
Moore, R. T.; Hancher, M.
2016-12-01
In 2005 Google Earth, a popular 3-D virtual globe, was first released. Scientists immediately recognized how it could be used to tell stories about the Earth. From 2006 to 2009, the "Virtual Globes" sessions of AGU included innovative examples of scientists and educators using Google Earth, and since that time it has become a commonplace tool for communicating scientific results. In 2009 Google Earth Engine, a cloud-based platform for planetary-scale geospatial analysis, was first announced. Earth Engine was initially used to extract information about the world's forests from raw Landsat data. Since then, the platform has proven highly effective for general analysis of georeferenced data, and users have expanded the list of use cases to include high-impact societal issues such as conservation, drought, disease, food security, water management, climate change and environmental monitoring. To support these use cases, the platform has continuously evolved with new datasets, analysis functions, and user interface tools. This talk will give an overview of the latest Google Earth and Earth Engine functionality that allow partners to understand, monitor and tell stories about of our living, breathing Earth. https://earth.google.com https://earthengine.google.com
Wampler, Peter J; Rediske, Richard R; Molla, Azizur R
2013-01-18
A remote sensing technique was developed which combines a Geographic Information System (GIS); Google Earth, and Microsoft Excel to identify home locations for a random sample of households in rural Haiti. The method was used to select homes for ethnographic and water quality research in a region of rural Haiti located within 9 km of a local hospital and source of health education in Deschapelles, Haiti. The technique does not require access to governmental records or ground based surveys to collect household location data and can be performed in a rapid, cost-effective manner. The random selection of households and the location of these households during field surveys were accomplished using GIS, Google Earth, Microsoft Excel, and handheld Garmin GPSmap 76CSx GPS units. Homes were identified and mapped in Google Earth, exported to ArcMap 10.0, and a random list of homes was generated using Microsoft Excel which was then loaded onto handheld GPS units for field location. The development and use of a remote sensing method was essential to the selection and location of random households. A total of 537 homes initially were mapped and a randomized subset of 96 was identified as potential survey locations. Over 96% of the homes mapped using Google Earth imagery were correctly identified as occupied dwellings. Only 3.6% of the occupants of mapped homes visited declined to be interviewed. 16.4% of the homes visited were not occupied at the time of the visit due to work away from the home or market days. A total of 55 households were located using this method during the 10 days of fieldwork in May and June of 2012. The method used to generate and field locate random homes for surveys and water sampling was an effective means of selecting random households in a rural environment lacking geolocation infrastructure. The success rate for locating households using a handheld GPS was excellent and only rarely was local knowledge required to identify and locate households. This method provides an important technique that can be applied to other developing countries where a randomized study design is needed but infrastructure is lacking to implement more traditional participant selection methods.
ERIC Educational Resources Information Center
Lucking, Robert A.; Christmann, Edwin P.; Whiting, Mervyn J.
2008-01-01
"Mashup" is a new technology term used to describe a web application that combines data or technology from several different sources. You can apply this concept in your classroom by having students create their own mashup maps. Google Maps provides you with the simple tools, map databases, and online help you'll need to quickly master this…
2013-08-09
CAPE CANAVERAL, Fla. – As seen on Google Maps, Firing Room 4 inside the Launch Control Center at NASA's Kennedy Space Center was one of the four control rooms used by NASA and contractor launch teams to oversee a space shuttle countdown. This firing room was the most advanced of the control rooms used for shuttle missions and was the primary firing room for the shuttle's final series of launches before retirement. It is furnished in a more contemporary style with wood cabinets and other features, although it retains many of the computer systems the shuttle counted on to operate safely. Specialized operators worked at consoles tailored to keep track of the status of shuttle systems while the spacecraft was processed in the Orbiter Processing Facility, being stacked inside the Vehicle Assembly Building and standing at the launch pad before liftoff. The firing rooms, including 3, were also used during NASA's Apollo Program. Google precisely mapped the space center and some of its historical facilities for the company's map page. The work allows Internet users to see inside buildings at Kennedy as they were used during the space shuttle era. Photo credit: Google/Wendy Wang
NASA Astrophysics Data System (ADS)
Askay, S.
2009-12-01
Published on Memorial Day 2009, Map the Fallen is a Google Earth visualization of the 5500+ US and international soldiers that have died in Iraq and Afghanistan since 2001. In addition to providing photos, stories and links for each solider, the time-animated map visually connects hometowns to places of death. This novel way of representing casualty data brings the geographic reach and magnitude of the issue into focus together with the very personal nature of individual stories. Innovative visualizations techniques were used that illustrate the spatio-temporal nature of this information and to show the global reach and interconnectivity of this issue. Several of advanced KML techniques employed to create this engaging and performance-conscious map will be discussed during this session. These include: 1) the use of HTML iframes and javascript to minimize the KML size, and extensive cross-linking throughout content; 2) the creation of a time-animated, on-screen casualty counter; 3) the use of parabolic arcs to connect each hometown to place of death; 4) the use of concentric spirals to represent chronological data; and 5) numerous performance optimizations to ensure the 23K placemarks, 2500 screen overlays and nearly 250k line vertices performed well in Google Earth. This session will include a demonstration of the map, conceptual discussions of the techniques used, and some in-depth technical explanation of the KML code.
Development of a Web-Based Visualization Platform for Climate Research Using Google Earth
NASA Technical Reports Server (NTRS)
Sun, Xiaojuan; Shen, Suhung; Leptoukh, Gregory G.; Wang, Panxing; Di, Liping; Lu, Mingyue
2011-01-01
Recently, it has become easier to access climate data from satellites, ground measurements, and models from various data centers, However, searching. accessing, and prc(essing heterogeneous data from different sources are very tim -consuming tasks. There is lack of a comprehensive visual platform to acquire distributed and heterogeneous scientific data and to render processed images from a single accessing point for climate studies. This paper. documents the design and implementation of a Web-based visual, interoperable, and scalable platform that is able to access climatological fields from models, satellites, and ground stations from a number of data sources using Google Earth (GE) as a common graphical interface. The development is based on the TCP/IP protocol and various data sharing open sources, such as OPeNDAP, GDS, Web Processing Service (WPS), and Web Mapping Service (WMS). The visualization capability of integrating various measurements into cE extends dramatically the awareness and visibility of scientific results. Using embedded geographic information in the GE, the designed system improves our understanding of the relationships of different elements in a four dimensional domain. The system enables easy and convenient synergistic research on a virtual platform for professionals and the general public, gr$tly advancing global data sharing and scientific research collaboration.
Monitoring Global Precipitation through UCI CHRS's RainMapper App on Mobile Devices
NASA Astrophysics Data System (ADS)
Nguyen, P.; Huynh, P.; Braithwaite, D.; Hsu, K. L.; Sorooshian, S.
2014-12-01
The Water and Development Information for Arid Lands-a Global Network (G-WADI) Precipitation Estimation from Remotely Sensed Information using Artificial Neural Networks—Cloud Classification System (PERSIANN-CCS) GeoServer has been developed through a collaboration between the Center for Hydrometeorology and Remote Sensing (CHRS) at the University of California, Irvine (UCI) and the UNESCO's International Hydrological Program (IHP). G-WADI PERSIANN-CCS GeoServer provides near real-time high resolution (0.04o, approx 4km) global (60oN - 60oS) satellite precipitation estimated by the PERSIANN-CCS algorithm developed by the scientists at CHRS. The G-WADI PERSIANN-CCS GeoServer utilizes the open-source MapServer software from the University of Minnesota to provide a user-friendly web-based mapping and visualization of satellite precipitation data. Recent efforts have been made by the scientists at CHRS to provide free on-the-go access to the PERSIANN-CCS precipitation data through an application named RainMapper for mobile devices. RainMapper provides visualization of global satellite precipitation of the most recent 3, 6, 12, 24, 48 and 72-hour periods overlaid with various basemaps. RainMapper uses the Google maps application programing interface (API) and embedded global positioning system (GPS) access to better monitor the global precipitation data on mobile devices. Functionalities include using geographical searching with voice recognition technologies make it easy for the user to explore near real-time precipitation in a certain location. RainMapper also allows for conveniently sharing the precipitation information and visualizations with the public through social networks such as Facebook and Twitter. RainMapper is available for iOS and Android devices and can be downloaded (free) from the App Store and Google Play. The usefulness of RainMapper was demonstrated through an application in tracking the evolution of the recent Rammasun Typhoon over the Philippines in mid July 2014.
NASA Astrophysics Data System (ADS)
García-Flores, Agustín.; Paz-Gallardo, Abel; Plaza, Antonio; Li, Jun
2016-10-01
This paper describes a new web platform dedicated to the classification of satellite images called Hypergim. The current implementation of this platform enables users to perform classification of satellite images from any part of the world thanks to the worldwide maps provided by Google Maps. To perform this classification, Hypergim uses unsupervised algorithms like Isodata and K-means. Here, we present an extension of the original platform in which we adapt Hypergim in order to use supervised algorithms to improve the classification results. This involves a significant modification of the user interface, providing the user with a way to obtain samples of classes present in the images to use in the training phase of the classification process. Another main goal of this development is to improve the runtime of the image classification process. To achieve this goal, we use a parallel implementation of the Random Forest classification algorithm. This implementation is a modification of the well-known CURFIL software package. The use of this type of algorithms to perform image classification is widespread today thanks to its precision and ease of training. The actual implementation of Random Forest was developed using CUDA platform, which enables us to exploit the potential of several models of NVIDIA graphics processing units using them to execute general purpose computing tasks as image classification algorithms. As well as CUDA, we use other parallel libraries as Intel Boost, taking advantage of the multithreading capabilities of modern CPUs. To ensure the best possible results, the platform is deployed in a cluster of commodity graphics processing units (GPUs), so that multiple users can use the tool in a concurrent way. The experimental results indicate that this new algorithm widely outperform the previous unsupervised algorithms implemented in Hypergim, both in runtime as well as precision of the actual classification of the images.
NASA Astrophysics Data System (ADS)
Cao, Y. B.; Hua, Y. X.; Zhao, J. X.; Guo, S. M.
2013-11-01
With China's rapid economic development and comprehensive national strength growing, Border work has become a long-term and important task in China's diplomatic work. How to implement rapid plotting, real-time sharing and mapping surrounding affairs has taken great significance for government policy makers and diplomatic staff. However, at present the already exists Boundary information system are mainly have problems of Geospatial data update is heavily workload, plotting tools are in a state of serious lack of, Geographic events are difficult to share, this phenomenon has seriously hampered the smooth development of the border task. The development and progress of Geographic information system technology especially the development of Web GIS offers the possibility to solve the above problems, this paper adopts four layers of B/S architecture, with the support of Google maps service, uses the free API which is offered by Google maps and its features of openness, ease of use, sharing characteristics, highresolution images to design and implement the surrounding transaction plotting and management system based on the web development technology of ASP.NET, C#, Ajax. The system can provide decision support for government policy makers as well as diplomatic staff's real-time plotting and sharing of surrounding information. The practice has proved that the system has good usability and strong real-time.
Creating a Geo-Referenced Bibliography with Google Earth and Geocommons: The Coos Bay Bibliography
ERIC Educational Resources Information Center
Schmitt, Jenni; Butler, Barb
2012-01-01
We compiled a geo-referenced bibliography of research including theses, peer-reviewed articles, agency literature, and books having sample collection sites in and around Coos Bay, Oregon. Using Google Earth and GeoCommons we created a map that allows users such as visiting researchers, faculty, students, and local agencies to identify previous…
Landsat Based Woody Vegetation Loss Detection in Queensland, Australia Using the Google Earth Engine
NASA Astrophysics Data System (ADS)
Johansen, K.; Phinn, S. R.; Taylor, M.
2014-12-01
Land clearing detection and woody Foliage Projective Cover (FPC) monitoring at the state and national level in Australia has mainly been undertaken by state governments and the Terrestrial Ecosystem Research Network (TERN) because of the considerable expense, expertise, sustained duration of activities and staffing levels needed. Only recently have services become available, providing low budget, generalized access to change detection tools suited to this task. The objective of this research was to examine if a globally available service, Google Earth Engine Beta, could be used to predict woody vegetation loss with accuracies approaching the methods used by TERN and the government of the state of Queensland, Australia. Two change detection approaches were investigated using Landsat Thematic Mapper time series and the Google Earth Engine Application Programming Interface: (1) CART and Random Forest classifiers; and (2) a normalized time series of Foliage Projective Cover (FPC) and NDVI combined with a spectral index. The CART and Random Forest classifiers produced high user's and producer's mapping accuracies of clearing (77-92% and 54-77%, respectively) when detecting change within epochs for which training data were available, but extrapolation to epochs without training data reduced the mapping accuracies. The use of FPC and NDVI time series provided a more robust approach for calculation of a clearing probability, as it did not rely on training data but instead on the difference of the normalized FPC / NDVI mean and standard deviation of a single year at the change point in relation to the remaining time series. However, the FPC and NDVI time series approach represented a trade-off between user's and producer's accuracies. Both change detection approaches explored in this research were sensitive to ephemeral greening and drying of the landscape. However, the developed normalized FPC and NDVI time series approach can be tuned to provide automated alerts for large woody vegetation clearing events by selecting suitable thresholds to identify very likely clearing. This research provides a comprehensive foundation to build further capacity to use globally accessible, free, online image datasets and processing tools to accurately detect woody vegetation clearing in an automated and rapid manner.
Global Analysis of River Planform Change using the Google Earth Engine
NASA Astrophysics Data System (ADS)
Bryk, A.; Dietrich, W. E.; Gorelick, N.; Sargent, R.; Braudrick, C. A.
2014-12-01
Geomorphologists have historically tracked river dynamics using a combination of maps, aerial photographs, and the stratigraphic record. Although stratigraphic records can extend into deep time, maps and aerial photographs often confine our record of change to sparse measurements over the last ~80 years and in some cases much less time. For the first time Google's Earth Engine (GEE) cloud based platform allows researchers the means to analyze quantitatively the pattern and pace of river channel change over the last 30 years with high temporal resolution across the entire planet. The GEE provides an application programing interface (API) that enables quantitative analysis of various data sets including the entire Landsat L1T archive. This allows change detection for channels wider than about 150 m over 30 years of successive, georeferenced imagery. Qualitatively, it becomes immediately evident that the pace of channel morphodynamics for similar planforms varies by orders of magnitude across the planet and downstream along individual rivers. To quantify these rates of change and to explore their controls we have developed methods for differentiating channels from floodplain along large alluvial rivers. We introduce a new metric of morphodynamics: the ratio of eroded area to channel area per unit time, referred to as "M". We also keep track of depositional areas resulting from channel shifting. To date our quantitative analysis has focused on rivers in the Andean foreland. Our analysis shows channel bank erosion rates, M, varies by orders of magnitude for these rivers, from 0 to ~0.25 yr-1, yet these rivers have essentially identical curvature and sinuosity and are visually indistinguishable. By tracking both bank paths in time, we find that, for some meandering rivers, a significant fraction of new floodplain is produced through outer-bank accretion rather than point bar deposition. This process is perhaps more important in generating floodplain stratigraphy than previously recognized. These initial findings indicate a new set of quantitative observations will emerge to further test and advance morphodynamic theory. The Google Earth Engine offers the opportunity to explore river morphodynamics on an unprecedented scale and provides a powerful tool for addressing fundamental questions in river morphodynamics.
Gao, Su-qing; Wang, Zhen; Gao, Hong-wei; Liu, Peng; Wang, Ze-rui; Li, Yan-li; Zhu, Xu-guang; Li, Xin-lou; Xu, Bo; Li, Yin-jun; Yang, Hong; de Vlas, Sake J.; Shi, Tao-xing; Cao, Wu-chun
2013-01-01
Background For years, emerging infectious diseases have appeared worldwide and threatened the health of people. The emergence and spread of an infectious-disease outbreak are usually unforeseen, and have the features of suddenness and uncertainty. Timely understanding of basic information in the field, and the collection and analysis of epidemiological information, is helpful in making rapid decisions and responding to an infectious-disease emergency. Therefore, it is necessary to have an unobstructed channel and convenient tool for the collection and analysis of epidemiologic information in the field. Methodology/Principal Findings Baseline information for each county in mainland China was collected and a database was established by geo-coding information on a digital map of county boundaries throughout the country. Google Maps was used to display geographic information and to conduct calculations related to maps, and the 3G wireless network was used to transmit information collected in the field to the server. This study established a decision support system for the response to infectious-disease emergencies based on WebGIS and mobile services (DSSRIDE). The DSSRIDE provides functions including data collection, communication and analyses in real time, epidemiological detection, the provision of customized epidemiological questionnaires and guides for handling infectious disease emergencies, and the querying of professional knowledge in the field. These functions of the DSSRIDE could be helpful for epidemiological investigations in the field and the handling of infectious-disease emergencies. Conclusions/Significance The DSSRIDE provides a geographic information platform based on the Google Maps application programming interface to display information of infectious disease emergencies, and transfers information between workers in the field and decision makers through wireless transmission based on personal computers, mobile phones and personal digital assistants. After a 2-year practice and application in infectious disease emergencies, the DSSRIDE is becoming a useful platform and is a useful tool for investigations in the field carried out by response sections and individuals. The system is suitable for use in developing countries and low-income districts. PMID:23372780
Phan, Thanh G; Beare, Richard; Chen, Jian; Clissold, Benjamin; Ly, John; Singhal, Shaloo; Ma, Henry; Srikanth, Velandai
2017-05-01
There is great interest in how endovascular clot retrieval hubs provide services to a population. We applied a computational method to objectively generate service boundaries for such endovascular clot retrieval hubs, defined by traveling time to hub. Stroke incidence data merged with population census to estimate numbers of stroke in metropolitan Melbourne, Australia. Traveling time from randomly generated addresses to 4 endovascular clot retrieval-capable hubs (Royal Melbourne Hospital [RMH], Monash Medical Center [MMC], Alfred Hospital [ALF], and Austin Hospital [AUS]) estimated using Google Map application program interface. Boundary maps generated based on traveling time at various times of day for combinations of hubs. In a 2-hub model, catchment was best distributed when RMH was paired with MMC (model 1a, RMH 1765 km 2 and MMC 1164 km 2 ) or with AUS (model 1c, RMH 1244 km 2 and AUS 1685 km 2 ), with no statistical difference between models ( P =0.20). Catchment was poorly distributed when RMH was paired with ALF (model 1b, RMH 2252 km 2 and ALF 676 km 2 ), significantly different from both models 1a and 1c (both P <0.05). Model 1a had the greatest proportion of patients arriving within ideal time of 30 minutes followed by model 1c ( P <0.001). In a 3-hub model, the combination of RMH, MMC, and AUS was superior to that of RMH, MMC, and ALF in catchment distribution and travel time. The method was also successfully applied to the city of Adelaide demonstrating wider applicability. We provide proof of concept for a novel computational method to objectively designate service boundaries for endovascular clot retrieval hubs. © 2017 American Heart Association, Inc.
NASA Astrophysics Data System (ADS)
Díaz, Elkin; Arguello, Henry
2016-05-01
Urban ecosystem studies require monitoring, controlling and planning to analyze building density, urban density, urban planning, atmospheric modeling and land use. In urban planning, there are many methods for building height estimation using optical remote sensing images. These methods however, highly depend on sun illumination and cloud-free weather. In contrast, high resolution synthetic aperture radar provides images independent from daytime and weather conditions, although, these images rely on special hardware and expensive acquisition. Most of the biggest cities around the world have been photographed by Google street view under different conditions. Thus, thousands of images from the principal streets of a city can be accessed online. The availability of this and similar rich city imagery such as StreetSide from Microsoft, represents huge opportunities in computer vision because these images can be used as input in many applications such as 3D modeling, segmentation, recognition and stereo correspondence. This paper proposes a novel algorithm to estimate building heights using public Google Street-View imagery. The objective of this work is to obtain thousands of geo-referenced images from Google Street-View using a representational state transfer system, and estimate their average height using single view metrology. Furthermore, the resulting measurements and image metadata are used to derive a layer of heights in a Google map available online. The experimental results show that the proposed algorithm can estimate an accurate average building height map of thousands of images using Google Street-View Imagery of any city.
The EPA Recovery Mapper is an Internet interactive mapping application that allows users to discover information about every American Recovery and Reinvestment Act (ARRA) award that EPA has funded for six programs. By integrating data reported by the recipients of Recovery Act funding and data created by EPA, this application delivers a level of transparency and public accessibility to users interested in EPA's use of Recovery Act monies. The application is relatively easy to use and builds on the same mapping model as Google, Bing, MapQuest and other commonly used mapping interfaces. EPA Recovery Mapper tracks each award made by each program and gives basic Quick Facts information for each award including award name, location, award date, dollar amounts and more. Data Summaries for each EPA program or for each state are provided displaying dollars for Total Awarded, Total Received (Paid), and Total Jobs This Quarter by Recovery for the latest quarter of data released by Recovery.gov. The data are reported to the government and EPA four times a year by the award recipients. The latest quarterly report will always be displayed in the EPA Recovery Mapper. In addition, the application provides many details about each award. Users will learn more about how to access and interpret these data later in this document. Data shown in the EPA Recovery Mapper are derived from information reported back to FederalReporting.gov from the recipients of Recovery Act funding. EPA
Change and Anomaly Detection in Real-Time GPS Data
NASA Astrophysics Data System (ADS)
Granat, R.; Pierce, M.; Gao, X.; Bock, Y.
2008-12-01
The California Real-Time Network (CRTN) is currently generating real-time GPS position data at a rate of 1-2Hz at over 80 locations. The CRTN data presents the possibility of studying dynamical solid earth processes in a way that complements existing seismic networks. To realize this possibility we have developed a prototype system for detecting changes and anomalies in the real-time data. Through this system, we can can correlate changes in multiple stations in order to detect signals with geographical extent. Our approach involves developing a statistical model for each GPS station in the network, and then using those models to segment the time series into a number of discrete states described by the model. We use a hidden Markov model (HMM) to describe the behavior of each station; fitting the model to the data requires neither labeled training examples nor a priori information about the system. As such, HMMs are well suited to this problem domain, in which the data remains largely uncharacterized. There are two main components to our approach. The first is the model fitting algorithm, regularized deterministic annealing expectation- maximization (RDAEM), which provides robust, high-quality results. The second is a web service infrastructure that connects the data to the statistical modeling analysis and allows us to easily present the results of that analysis through a web portal interface. This web service approach facilitates the automatic updating of station models to keep pace with dynamical changes in the data. Our web portal interface is critical to the process of interpreting the data. A Google Maps interface allows users to visually interpret state changes not only on individual stations but across the entire network. Users can drill down from the map interface to inspect detailed results for individual stations, download the time series data, and inspect fitted models. Alternatively, users can use the web portal look at the evolution of changes on the network by moving backwards and forwards in time.
ERIC Educational Resources Information Center
Giorgis, Scott
2015-01-01
Three-dimensional thinking skills are extremely useful for geoscientists, and at the undergraduate level, these skills are often emphasized in structural geology courses. Google Earth is a powerful tool for visualizing the three-dimensional nature of data collected on the surface of Earth. The results of a 5 y pre- and posttest study of the…
Assessing natural hazard risk using images and data
NASA Astrophysics Data System (ADS)
Mccullough, H. L.; Dunbar, P. K.; Varner, J. D.; Mungov, G.
2012-12-01
Photographs and other visual media provide valuable pre- and post-event data for natural hazard assessment. Scientific research, mitigation, and forecasting rely on visual data for risk analysis, inundation mapping and historic records. Instrumental data only reveal a portion of the whole story; photographs explicitly illustrate the physical and societal impacts from the event. Visual data is rapidly increasing as the availability of portable high resolution cameras and video recorders becomes more attainable. Incorporating these data into archives ensures a more complete historical account of events. Integrating natural hazards data, such as tsunami, earthquake and volcanic eruption events, socio-economic information, and tsunami deposits and runups along with images and photographs enhances event comprehension. Global historic databases at NOAA's National Geophysical Data Center (NGDC) consolidate these data, providing the user with easy access to a network of information. NGDC's Natural Hazards Image Database (ngdc.noaa.gov/hazardimages) was recently improved to provide a more efficient and dynamic user interface. It uses the Google Maps API and Keyhole Markup Language (KML) to provide geographic context to the images and events. Descriptive tags, or keywords, have been applied to each image, enabling easier navigation and discovery. In addition, the Natural Hazards Map Viewer (maps.ngdc.noaa.gov/viewers/hazards) provides the ability to search and browse data layers on a Mercator-projection globe with a variety of map backgrounds. This combination of features creates a simple and effective way to enhance our understanding of hazard events and risks using imagery.
Drawing the Line with Google Earth: The Place of Digital Mapping outside of Geography
ERIC Educational Resources Information Center
Mercier, O. Ripeka; Rata, Arama
2017-01-01
The "Te Kawa a Maui Atlas" project explores how mapping activities support undergraduate student engagement and learning in Maori studies. This article describes two specific assignments, which used online mapping allowing students to engage with the work of their peers. By analysing student evaluations of these activities, we identify…
How to Use This Website | USDA Plant Hardiness Zone Map
, regional or national plant hardiness zone maps in three different resolutions using the following steps. To default printing menu option or button. Viewing the Map - Open Full Map Button c. Save Full Map Button can copy the e-mail address and paste it into a different e-mail client (e.g., Google Gmail, Yahoo
Interactive web-based mapping: bridging technology and data for health
2011-01-01
Background The Community Health Information System (CHIS) online mapping system was first launched in 1998. Its overarching goal was to provide researchers, residents and organizations access to health related data reflecting the overall health and well-being of their communities within the Greater Houston area. In September 2009, initial planning and development began for the next generation of CHIS. The overarching goal for the new version remained to make health data easily accessible for a wide variety of research audiences. However, in the new version we specifically sought to make the CHIS truly interactive and give the user more control over data selection and reporting. Results In July 2011, a beta version of the next-generation of the application was launched. This next-generation is also a web based interactive mapping tool comprised of two distinct portals: the Breast Health Portal and Project Safety Net. Both are accessed via a Google mapping interface. Geographic coverage for the portals is currently an 8 county region centered on Harris County, Texas. Data accessed by the application include Census 2000, Census 2010 (underway), cancer incidence from the Texas Cancer Registry (TX Dept. of State Health Services), death data from Texas Vital Statistics, clinic locations for free and low-cost health services, along with service lists, hours of operation, payment options and languages spoken, uninsured and poverty data. Conclusions The system features query on the fly technology, which means the data is not generated until the query is provided to the system. This allows users to interact in real-time with the databases and generate customized reports and maps. To the author's knowledge, the Breast Health Portal and Project Safety Net are the first local-scale interactive online mapping interfaces for public health data which allow users to control the data generated. For example, users may generate breast cancer incidence rates by Census tract, in real time, for women aged 40-64. Conversely, they could then generate the same rates for women aged 35-55. The queries are user controlled. PMID:22195603
Traffic Sign Inventory from Google Street View Images
NASA Astrophysics Data System (ADS)
Tsai, Victor J. D.; Chen, Jyun-Han; Huang, Hsun-Sheng
2016-06-01
Traffic sign detection and recognition (TSDR) has drawn considerable attention on developing intelligent transportation systems (ITS) and autonomous vehicle driving systems (AVDS) since 1980's. Unlikely to the general TSDR systems that deal with real-time images captured by the in-vehicle cameras, this research aims on developing techniques for detecting, extracting, and positioning of traffic signs from Google Street View (GSV) images along user-selected routes for low-cost, volumetric and quick establishment of the traffic sign infrastructural database that may be associated with Google Maps. The framework and techniques employed in the proposed system are described.
Using Google Streetview Panoramic Imagery for Geoscience Education
NASA Astrophysics Data System (ADS)
De Paor, D. G.; Dordevic, M. M.
2014-12-01
Google Streetview is a feature of Google Maps and Google Earth that allows viewers to switch from map or satellite view to 360° panoramic imagery recorded close to the ground. Most panoramas are recorded by Google engineers using special cameras mounted on the roofs of cars. Bicycles, snowmobiles, and boats have also been used and sometimes the camera has been mounted on a backpack for off-road use by hikers and skiers or attached to scuba-diving gear for "Underwater Streetview (sic)." Streetview panoramas are linked together so that the viewer can change viewpoint by clicking forward and reverse buttons. They therefore create a 4-D touring effect. As part of the GEODE project ("Google Earth for Onsite and Distance Education"), we are experimenting with the use of Streetview imagery for geoscience education. Our web-based test application allows instructors to select locations for students to study. Students are presented with a set of questions or tasks that they must address by studying the panoramic imagery. Questions include identification of rock types, structures such as faults, and general geological setting. The student view is locked into Streetview mode until they submit their answers, whereupon the map and satellite views become available, allowing students to zoom out and verify their location on Earth. Student learning is scaffolded by automatic computerized feedback. There are lots of existing Streetview panoramas with rich geological content. Additionally, instructors and members of the general public can create panoramas, including 360° Photo Spheres, by stitching images taken with their mobiles devices and submitting them to Google for evaluation and hosting. A multi-thousand-dollar, multi-directional camera and mount can be purchased from DIY-streetview.com. This allows power users to generate their own high-resolution panoramas. A cheaper, 360° video camera is soon to be released according to geonaute.com. Thus there are opportunities for geoscience educators both to use existing Streetview imagery and to generate new imagery for specific locations of geological interest. The GEODE team includes the authors and: H. Almquist, C. Bentley, S. Burgin, C. Cervato, G. Cooper, P. Karabinos, T. Pavlis, J. Piatek, B. Richards, J. Ryan, R. Schott, K. St. John, B. Tewksbury, and S. Whitmeyer.
Cartographic analyses of geographic information available on Google Earth Images
NASA Astrophysics Data System (ADS)
Oliveira, J. C.; Ramos, J. R.; Epiphanio, J. C.
2011-12-01
The propose was to evaluate planimetric accuracy of satellite images available on database of Google Earth. These images are referents to the vicinities of the Federal Univertisity of Viçosa, Minas Gerais - Brazil. The methodology developed evaluated the geographical information of three groups of images which were in accordance to the level of detail presented in the screen images (zoom). These groups of images were labeled to Zoom 1000 (a single image for the entire study area), Zoom 100 (formed by a mosaic of 73 images) and Zoom 100 with geometric correction (this mosaic is like before, however, it was applied a geometric correction through control points). In each group of image was measured the Cartographic Accuracy based on statistical analyses and brazilian's law parameters about planimetric mapping. For this evaluation were identified 22 points in each group of image, where the coordinates of each point were compared to the coordinates of the field obtained by GPS (Global Positioning System). The Table 1 show results related to accuracy (based on a threshold equal to 0.5 mm * mapping scale) and tendency (abscissa and ordinate) between the coordinates of the image and the coordinates of field. Table 1 The geometric correction applied to the Group Zoom 100 reduced the trends identified earlier, and the statistical tests pointed a usefulness of the data for a mapping at a scale of 1/5000 with error minor than 0.5 mm * scale. The analyses proved the quality of cartographic data provided by Google, as well as the possibility of reduce the divergences of positioning present on the data. It can be concluded that it is possible to obtain geographic information database available on Google Earth, however, the level of detail (zoom) used at the time of viewing and capturing information on the screen influences the quality cartographic of the mapping. Although cartographic and thematic potential present in the database, it is important to note that both the software as data distributed by Google Earth has policies for use and distribution.
Table 1 - PLANIMETRIC ANALYSIS
Interfaces Visualize Data for Airline Safety, Efficiency
NASA Technical Reports Server (NTRS)
2014-01-01
As the A-Train Constellation orbits Earth to gather data, NASA scientists and partners visualize, analyze, and communicate the information. To this end, Langley Research Center awarded SBIR funding to Fairfax, Virginia-based WxAnalyst Ltd. to refine the company's existing user interface for Google Earth to visualize data. Hawaiian Airlines is now using the technology to help manage its flights.
Sean A. Parks; Lisa M. Holsinger; Morgan A. Voss; Rachel A. Loehman; Nathaniel P. Robinson
2018-01-01
Landsat-based fire severity datasets are an invaluable resource for monitoring and research purposes. These gridded fire severity datasets are generally produced with pre- and post-fire imagery to estimate the degree of fire-induced ecological change. Here, we introduce methods to produce three Landsat-based fire severity metrics using the Google Earth Engine (GEE)...
NASA Astrophysics Data System (ADS)
Griffith, P. C.; Wilcox, L. E.; Morrell, A.
2009-12-01
The central objective of the North American Carbon Program (NACP), a core element of the US Global Change Research Program, is to quantify the sources and sinks of carbon dioxide, carbon monoxide, and methane in North America and adjacent ocean regions. The NACP consists of a wide range of investigators at universities and federal research centers. Although many of these investigators have worked together in the past, many have had few prior interactions and may not know of similar work within knowledge domains, much less across the diversity of environments and scientific approaches in the Program. Coordinating interactions and sharing data are major challenges in conducting NACP. The Google Earth and Google Map Collections on the NACP website (www.nacarbon.org) provide a geographical view of the research products contributed by each core and affiliated NACP project. Other relevant data sources (e.g. AERONET, LVIS) can also be browsed in spatial context with NACP contributions. Each contribution links to project-oriented metadata, or “project profiles”, that provide a greater understanding of the scientific and social context of each dataset and are an important means of communicating within the NACP and to the larger carbon cycle science community. Project profiles store information such as a project's title, leaders, participants, an abstract, keywords, funding agencies, associated intensive campaigns, expected data products, data needs, publications, and URLs to associated data centers, datasets, and metadata. Data products are research contributions that include biometric inventories, flux tower estimates, remote sensing land cover products, tools, services, and model inputs / outputs. Project leaders have been asked to identify these contributions to the site level whenever possible, either through simple latitude/longitude pair, or by uploading a KML, KMZ, or shape file. Project leaders may select custom icons to graphically categorize their contributions; for example, a ship for oceanographic samples, a tower for tower measurements. After post-processing, research contributions are added to the NACP Google Earth and Google Map Collection to facilitate discovery and use in synthesis activities of the Program.
NASA Astrophysics Data System (ADS)
Gorelick, Noel
2013-04-01
The Google Earth Engine platform is a system designed to enable petabyte-scale, scientific analysis and visualization of geospatial datasets. Earth Engine provides a consolidated environment including a massive data catalog co-located with thousands of computers for analysis. The user-friendly front-end provides a workbench environment to allow interactive data and algorithm development and exploration and provides a convenient mechanism for scientists to share data, visualizations and analytic algorithms via URLs. The Earth Engine data catalog contains a wide variety of popular, curated datasets, including the world's largest online collection of Landsat scenes (> 2.0M), numerous MODIS collections, and many vector-based data sets. The platform provides a uniform access mechanism to a variety of data types, independent of their bands, projection, bit-depth, resolution, etc..., facilitating easy multi-sensor analysis. Additionally, a user is able to add and curate their own data and collections. Using a just-in-time, distributed computation model, Earth Engine can rapidly process enormous quantities of geo-spatial data. All computation is performed lazily; nothing is computed until it's required either for output or as input to another step. This model allows real-time feedback and preview during algorithm development, supporting a rapid algorithm development, test, and improvement cycle that scales seamlessly to large-scale production data processing. Through integration with a variety of other services, Earth Engine is able to bring to bear considerable analytic and technical firepower in a transparent fashion, including: AI-based classification via integration with Google's machine learning infrastructure, publishing and distribution at Google scale through integration with the Google Maps API, Maps Engine and Google Earth, and support for in-the-field activities such as validation, ground-truthing, crowd-sourcing and citizen science though the Android Open Data Kit.
NASA Astrophysics Data System (ADS)
Gorelick, N.
2012-12-01
The Google Earth Engine platform is a system designed to enable petabyte-scale, scientific analysis and visualization of geospatial datasets. Earth Engine provides a consolidated environment including a massive data catalog co-located with thousands of computers for analysis. The user-friendly front-end provides a workbench environment to allow interactive data and algorithm development and exploration and provides a convenient mechanism for scientists to share data, visualizations and analytic algorithms via URLs. The Earth Engine data catalog contains a wide variety of popular, curated datasets, including the world's largest online collection of Landsat scenes (> 2.0M), numerous MODIS collections, and many vector-based data sets. The platform provides a uniform access mechanism to a variety of data types, independent of their bands, projection, bit-depth, resolution, etc..., facilitating easy multi-sensor analysis. Additionally, a user is able to add and curate their own data and collections. Using a just-in-time, distributed computation model, Earth Engine can rapidly process enormous quantities of geo-spatial data. All computation is performed lazily; nothing is computed until it's required either for output or as input to another step. This model allows real-time feedback and preview during algorithm development, supporting a rapid algorithm development, test, and improvement cycle that scales seamlessly to large-scale production data processing. Through integration with a variety of other services, Earth Engine is able to bring to bear considerable analytic and technical firepower in a transparent fashion, including: AI-based classification via integration with Google's machine learning infrastructure, publishing and distribution at Google scale through integration with the Google Maps API, Maps Engine and Google Earth, and support for in-the-field activities such as validation, ground-truthing, crowd-sourcing and citizen science though the Android Open Data Kit.
The Chandra Source Catalog : Google Earth Interface
NASA Astrophysics Data System (ADS)
Glotfelty, Kenny; McLaughlin, W.; Evans, I.; Evans, J.; Anderson, C. S.; Bonaventura, N. R.; Davis, J. E.; Doe, S. M.; Fabbiano, G.; Galle, E. C.; Gibbs, D. G., II; Grier, J. D.; Hain, R.; Hall, D. M.; Harbo, P. N.; He, H.; Houck, J. C.; Karovska, M.; Kashyap, V. L.; Lauer, J.; McCollough, M. L.; McDowell, J. C.; Miller, J. B.; Mitschang, A. W.; Morgan, D. L.; Mossman, A. E.; Nichols, J. S.; Nowak, M. A.; Plummer, D. A.; Primini, F. A.; Refsdal, B. L.; Rots, A. R.; Siemiginowska, A. L.; Sundheim, B. A.; Tibbetts, M. S.; van Stone, D. W.; Winkelman, S. L.; Zografou, P.
2009-09-01
The Chandra Source Catalog (CSC) contains multi-resolution, exposure corrected, background subtracted, full-field images that are stored as individual FITS files and as three-color JPEG files. In this poster we discuss how we took these data and were able to, with relatively minimal effort, convert them for use with the Google Earth application in its ``Sky'' mode. We will highlight some of the challenges which include converting the data to the required Mercator projection, reworking the 3-color algorithm for pipeline processing, and ways to reduce the data volume through re-binning, using color-maps, and special Keyhole Markup Language (kml) tags to only load images on-demand. The result is a collection of some 11,000 3-color images that are available for all the individual observation in the CSC Release 1. We also have made available all ˜4000 Field-of-View outlines (with per-chip regions), which turns out are trivial to produce starting with a simple dmlist command. In the first week of release, approximately 40% of the images have been accessed at least once through some 50,000 individual web hits which have served over 4Gb of data to roughly 750 users in 60+ countries. We will also highlight some future directions we are exploring, including real-time catalog access to individual source properties and eventual access to file based products such as FITS images, spectra, and light-curves.
Tapir: A web interface for transit/eclipse observability
NASA Astrophysics Data System (ADS)
Jensen, Eric
2013-06-01
Tapir is a set of tools, written in Perl, that provides a web interface for showing the observability of periodic astronomical events, such as exoplanet transits or eclipsing binaries. The package provides tools for creating finding charts for each target and airmass plots for each event. The code can access target lists that are stored on-line in a Google spreadsheet or in a local text file.
Distributed Kernelized Locality-Sensitive Hashing for Faster Image Based Navigation
2015-03-26
Facebook, Google, and Yahoo !. Current methods for image retrieval become problematic when implemented on image datasets that can easily reach billions of...correlations. Tech industry leaders like Facebook, Google, and Yahoo ! sort and index even larger volumes of “big data” daily. When attempting to process...open source implementation of Google’s MapReduce programming paradigm [13] which has been used for many different things. Using Apache Hadoop, Yahoo
LLMapReduce: Multi-Lingual Map-Reduce for Supercomputing Environments
2015-11-20
1990s. Popularized by Google [36] and Apache Hadoop [37], map-reduce has become a staple technology of the ever- growing big data community...Lexington, MA, U.S.A Abstract— The map-reduce parallel programming model has become extremely popular in the big data community. Many big data ...to big data users running on a supercomputer. LLMapReduce dramatically simplifies map-reduce programming by providing simple parallel programming
Natural Language Search Interfaces: Health Data Needs Single-Field Variable Search.
Jay, Caroline; Harper, Simon; Dunlop, Ian; Smith, Sam; Sufi, Shoaib; Goble, Carole; Buchan, Iain
2016-01-14
Data discovery, particularly the discovery of key variables and their inter-relationships, is key to secondary data analysis, and in-turn, the evolving field of data science. Interface designers have presumed that their users are domain experts, and so they have provided complex interfaces to support these "experts." Such interfaces hark back to a time when searches needed to be accurate first time as there was a high computational cost associated with each search. Our work is part of a governmental research initiative between the medical and social research funding bodies to improve the use of social data in medical research. The cross-disciplinary nature of data science can make no assumptions regarding the domain expertise of a particular scientist, whose interests may intersect multiple domains. Here we consider the common requirement for scientists to seek archived data for secondary analysis. This has more in common with search needs of the "Google generation" than with their single-domain, single-tool forebears. Our study compares a Google-like interface with traditional ways of searching for noncomplex health data in a data archive. Two user interfaces are evaluated for the same set of tasks in extracting data from surveys stored in the UK Data Archive (UKDA). One interface, Web search, is "Google-like," enabling users to browse, search for, and view metadata about study variables, whereas the other, traditional search, has standard multioption user interface. Using a comprehensive set of tasks with 20 volunteers, we found that the Web search interface met data discovery needs and expectations better than the traditional search. A task × interface repeated measures analysis showed a main effect indicating that answers found through the Web search interface were more likely to be correct (F1,19=37.3, P<.001), with a main effect of task (F3,57=6.3, P<.001). Further, participants completed the task significantly faster using the Web search interface (F1,19=18.0, P<.001). There was also a main effect of task (F2,38=4.1, P=.025, Greenhouse-Geisser correction applied). Overall, participants were asked to rate learnability, ease of use, and satisfaction. Paired mean comparisons showed that the Web search interface received significantly higher ratings than the traditional search interface for learnability (P=.002, 95% CI [0.6-2.4]), ease of use (P<.001, 95% CI [1.2-3.2]), and satisfaction (P<.001, 95% CI [1.8-3.5]). The results show superior cross-domain usability of Web search, which is consistent with its general familiarity and with enabling queries to be refined as the search proceeds, which treats serendipity as part of the refinement. The results provide clear evidence that data science should adopt single-field natural language search interfaces for variable search supporting in particular: query reformulation; data browsing; faceted search; surrogates; relevance feedback; summarization, analytics, and visual presentation.
Natural Language Search Interfaces: Health Data Needs Single-Field Variable Search
Smith, Sam; Sufi, Shoaib; Goble, Carole; Buchan, Iain
2016-01-01
Background Data discovery, particularly the discovery of key variables and their inter-relationships, is key to secondary data analysis, and in-turn, the evolving field of data science. Interface designers have presumed that their users are domain experts, and so they have provided complex interfaces to support these “experts.” Such interfaces hark back to a time when searches needed to be accurate first time as there was a high computational cost associated with each search. Our work is part of a governmental research initiative between the medical and social research funding bodies to improve the use of social data in medical research. Objective The cross-disciplinary nature of data science can make no assumptions regarding the domain expertise of a particular scientist, whose interests may intersect multiple domains. Here we consider the common requirement for scientists to seek archived data for secondary analysis. This has more in common with search needs of the “Google generation” than with their single-domain, single-tool forebears. Our study compares a Google-like interface with traditional ways of searching for noncomplex health data in a data archive. Methods Two user interfaces are evaluated for the same set of tasks in extracting data from surveys stored in the UK Data Archive (UKDA). One interface, Web search, is “Google-like,” enabling users to browse, search for, and view metadata about study variables, whereas the other, traditional search, has standard multioption user interface. Results Using a comprehensive set of tasks with 20 volunteers, we found that the Web search interface met data discovery needs and expectations better than the traditional search. A task × interface repeated measures analysis showed a main effect indicating that answers found through the Web search interface were more likely to be correct (F 1,19=37.3, P<.001), with a main effect of task (F 3,57=6.3, P<.001). Further, participants completed the task significantly faster using the Web search interface (F 1,19=18.0, P<.001). There was also a main effect of task (F 2,38=4.1, P=.025, Greenhouse-Geisser correction applied). Overall, participants were asked to rate learnability, ease of use, and satisfaction. Paired mean comparisons showed that the Web search interface received significantly higher ratings than the traditional search interface for learnability (P=.002, 95% CI [0.6-2.4]), ease of use (P<.001, 95% CI [1.2-3.2]), and satisfaction (P<.001, 95% CI [1.8-3.5]). The results show superior cross-domain usability of Web search, which is consistent with its general familiarity and with enabling queries to be refined as the search proceeds, which treats serendipity as part of the refinement. Conclusions The results provide clear evidence that data science should adopt single-field natural language search interfaces for variable search supporting in particular: query reformulation; data browsing; faceted search; surrogates; relevance feedback; summarization, analytics, and visual presentation. PMID:26769334
Crop classification and mapping based on Sentinel missions data in cloud environment
NASA Astrophysics Data System (ADS)
Lavreniuk, M. S.; Kussul, N.; Shelestov, A.; Vasiliev, V.
2017-12-01
Availability of high resolution satellite imagery (Sentinel-1/2/3, Landsat) over large territories opens new opportunities in agricultural monitoring. In particular, it becomes feasible to solve crop classification and crop mapping task at country and regional scale using time series of heterogenous satellite imagery. But in this case, we face with the problem of Big Data. Dealing with time series of high resolution (10 m) multispectral imagery we need to download huge volumes of data and then process them. The solution is to move "processing chain" closer to data itself to drastically shorten time for data transfer. One more advantage of such approach is the possibility to parallelize data processing workflow and efficiently implement machine learning algorithms. This could be done with cloud platform where Sentinel imagery are stored. In this study, we investigate usability and efficiency of two different cloud platforms Amazon and Google for crop classification and crop mapping problems. Two pilot areas were investigated - Ukraine and England. Google provides user friendly environment Google Earth Engine for Earth observation applications with a lot of data processing and machine learning tools already deployed. At the same time with Amazon one gets much more flexibility in implementation of his own workflow. Detailed analysis of pros and cons will be done in the presentation.
SECURE INTERNET OF THINGS-BASED CLOUD FRAMEWORK TO CONTROL ZIKA VIRUS OUTBREAK.
Sareen, Sanjay; Sood, Sandeep K; Gupta, Sunil Kumar
2017-01-01
Zika virus (ZikaV) is currently one of the most important emerging viruses in the world which has caused outbreaks and epidemics and has also been associated with severe clinical manifestations and congenital malformations. Traditional approaches to combat the ZikaV outbreak are not effective for detection and control. The aim of this study is to propose a cloud-based system to prevent and control the spread of Zika virus disease using integration of mobile phones and Internet of Things (IoT). A Naive Bayesian Network (NBN) is used to diagnose the possibly infected users, and Google Maps Web service is used to provide the geographic positioning system (GPS)-based risk assessment to prevent the outbreak. It is used to represent each ZikaV infected user, mosquito-dense sites, and breeding sites on the Google map that helps the government healthcare authorities to control such risk-prone areas effectively and efficiently. The performance and accuracy of the proposed system are evaluated using dataset for 2 million users. Our system provides high accuracy for initial diagnosis of different users according to their symptoms and appropriate GPS-based risk assessment. The cloud-based proposed system contributed to the accurate NBN-based classification of infected users and accurate identification of risk-prone areas using Google Maps.
Reaching the Next Generation of College Students via Their Digital Devices.
NASA Astrophysics Data System (ADS)
Whitmeyer, S. J.; De Paor, D. G.; Bentley, C.
2015-12-01
Current college students attended school during a decade in which many school districts banned cellphones from the classroom or even from school grounds. These students are used to being told to put away their mobile devices and concentrate on traditional classroom activities such as watching PowerPoint presentations or calculating with pencil and paper. However, due to a combination of parental security concerns and recent education research, schools are rapidly changing policy and embracing mobile devices for ubiquitous learning opportunities inside and outside of the classroom. Consequently, many of the next generation of college students will have expectations of learning via mobile technology. We have developed a range of digital geology resources to aid mobile-based geoscience education at college level, including mapping on iPads and other tablets, "crowd-sourced" field projects, augmented reality-supported asynchronous field classes, 3D and 4D split-screen virtual reality tours, macroscopic and microscopic gigapixel imagery, 360° panoramas, assistive devices for inclusive field education, and game-style educational challenges. Class testing of virtual planetary tours shows modest short-term learning gains, but more work is needed to ensure long-term retention. Many of our resources rely on the Google Earth browser plug-in and application program interface (API). Because of security concerns, browser plug-ins in general are being phased out and the Google Earth API will not be supported in future browsers. However, a new plug-in-free API is promised by Google and an alternative open-source virtual globe called Cesium is undergoing rapid development. It already supports the main aspects of Keyhole Markup Language and has features of significant benefit to geoscience, including full support on mobile devices and sub-surface viewing and touring. The research team includes: Heather Almquist, Stephen Burgin, Cinzia Cervato, Filis Coba, Chloe Constants, Gene Cooper, Mladen Dordevic, Marissa Dudek, Brandon Fitzwater, Bridget Gomez, Tyler Hansen, Paul Karabinos, Terry Pavlis, Jen Piatek, Alan Pitts, Robin Rohrback, Bill Richards, Caroline Robinson, Jeff Rollins, Jeff Ryan, Ron Schott, Kristen St. John, and Barb Tewksbury. Supported by NSF DUE 1323419 and by Google Geo Curriculum Awards.
Cross-disciplinary Undergraduate Research: A Case Study in Digital Mapping, western Ireland
NASA Astrophysics Data System (ADS)
Whitmeyer, S. J.; de Paor, D. G.; Nicoletti, J.; Rivera, M.; Santangelo, B.; Daniels, J.
2008-12-01
As digital mapping technology becomes ever more advanced, field geologists spend a greater proportion of time learning digital methods relative to analyzing rocks and structures. To explore potential solutions to the time commitment implicit in learning digital field methods, we paired James Madison University (JMU) geology majors (experienced in traditional field techniques) with Worcester Polytechnic Institute (WPI) engineering students (experienced in computer applications) during a four week summer mapping project in Connemara, western Ireland. The project consisted of approximately equal parts digital field mapping (directed by the geology students), and lab-based map assembly, evaluation and formatting for virtual 3D terrains (directed by the engineering students). Students collected geologic data in the field using ruggedized handheld computers (Trimble GeoExplorer® series) with ArcPAD® software. Lab work initially focused on building geologic maps in ArcGIS® from the digital field data and then progressed to developing Google Earth-based visualizations of field data and maps. Challenges included exporting GIS data, such as locations and attributes, to KML tags for viewing in Google Earth, which we accomplished using a Linux bash script written by one of our engineers - a task outside the comfort zone of the average geology major. We also attempted to expand the scope of Google Earth by using DEMs of present-day geologically-induced landforms as representative models for paleo-geographic reconstructions of the western Ireland field area. As our integrated approach to digital field work progressed, we found that our digital field mapping produced data at a faster rate than could be effectively managed during our allotted time for lab work. This likely reflected the more developed methodology for digital field data collection, as compared with our lab-based attempts to develop new methods for 3D visualization of geologic maps. However, this experiment in cross-disciplinary undergraduate research was a big success, with an enthusiastic interchange of expertise between undergraduate geology and engineering students that produced new, cutting-edge methods for visualizing geologic data and maps.
What's New in the Ocean in Google Earth and Maps
NASA Astrophysics Data System (ADS)
Austin, J.; Sandwell, D. T.
2014-12-01
Jenifer Austin, Jamie Adams, Kurt Schwehr, Brian Sullivan, David Sandwell2, Walter Smith3, Vicki Ferrini4, and Barry Eakins5, 1 Google Inc., 1600 Amphitheatre Parkway, Mountain View, California, USA 2 University of California-San Diego, Scripps Institute of Oceanography, La Jolla, California ,USA3 NOAA Laboratory for Satellite Altimetry, College Park, Maryland, USA4 Lamont Doherty, Columbia University5 NOAAMore than two-thirds of Earth is covered by oceans. On the almost 6 year anniversary of launching an explorable ocean seafloor in Google Earth and Maps, we updated our global underwater terrain dataset in partnership with Lamont-Doherty at Columbia, the Scripps Institution of Oceanography, and NOAA. With this update to our ocean map, we'll reveal an additional 2% of the ocean in high resolution representing 2 years of work by Columbia, pulling in data from numerous institutions including the Campeche Escarpment in the Gulf of Mexico in partnership with Charlie Paul at MBARI and the Schmidt Ocean Institute. The Scripps Institution of Oceanography at UCSD has curated 30 years of data from more than 8,000 ship cruises and 135 different institutions to reveal 15 percent of the seafloor at 1 km resolution. In addition, explore new data from an automated pipeline built to make updates to our Ocean Map more scalable in partnership with NOAA's National Geophysical Data Center (link to http://www.ngdc.noaa.gov/mgg/bathymetry/) and the University of Colorado CIRES program (link to http://cires.colorado.edu/index.html).
NASA Astrophysics Data System (ADS)
Danladi, Iliya Bauchi; Kore, Basiru Mohammed; Gül, Murat
2017-10-01
Coastal areas are important regions in the world as they host huge population, diverse ecosystems and natural resources. However, owing to their settings, elevations and proximities to the sea, climate change (global warming) and human activities are threatening issues. Herein, we report the coastline changes and possible future threats related to sea level rise owing to global warming and human activities in the coastal region of Nigeria. Google earth images, Digital Elevation Model (DEM) and geological maps were used. Using google earth images, coastal changes for the past 43 years, 3 years prior to and after the construction of breakwaters along Goshen Beach Estate (Lekki) were examined. Additionally, coastline changes along Lekki Phase I from 2013 to 2016 were evaluated. The DEM map was used to delineate 0-2 m, 2-5 m and 5-10 m asl which correspond to undifferentiated sands and gravels to clays on the geological map. The results of the google earth images revealed remarkable erosion along both Lekki and Lekki Phase I, with the destruction of a lagoon in Lekki Phase I. Based on the result of the DEM map and geology, elevations of 0-2 m, 2-5 m and 5-10 m asl were interpreted as highly risky, moderately risky and risky respectively. Considering factors threatening coastal regions, the erosion and destruction of the lagoon along the Nigerian coast may be ascribed to sea level rise as a result of global warming and intense human activities respectively.
Transforming Polar Research with Google Glass Augmented Reality (Invited)
NASA Astrophysics Data System (ADS)
Ruthkoski, T.
2013-12-01
Augmented reality is a new technology with the potential to accelerate the advancement of science, particularly in geophysical research. Augmented reality is defined as a live, direct or indirect, view of a physical, real-world environment whose elements are augmented (or supplemented) by computer-generated sensory input such as sound, video, graphics or GPS data. When paired with advanced computing techniques on cloud resources, augmented reality has the potential to improve data collection techniques, visualizations, as well as in-situ analysis for many areas of research. Google is currently a pioneer of augmented reality technology and has released beta versions of their wearable computing device, Google Glass, to a select number of developers and beta testers. This community of 'Glass Explorers' is the vehicle from which Google shapes the future of their augmented reality device. Example applications of Google Glass in geophysical research range from use as a data gathering interface in harsh climates to an on-site visualization and analysis tool. Early participation in the shaping of the Google Glass device is an opportunity for researchers to tailor this new technology to their specific needs. The purpose of this presentation is to provide geophysical researchers with a hands-on first look at Google Glass and its potential as a scientific tool. Attendees will be given an overview of the technical specifications as well as a live demonstration of the device. Potential applications to geophysical research in polar regions will be the primary focus. The presentation will conclude with an open call to participate, during which attendees may indicate interest in developing projects that integrate Google Glass into their research. Application Mockup: Penguin Counter Google Glass Augmented Reality Device
Transforming Polar Research with Google Glass Augmented Reality (Invited)
NASA Astrophysics Data System (ADS)
Ramachandran, R.; McEniry, M.; Maskey, M.
2011-12-01
Augmented reality is a new technology with the potential to accelerate the advancement of science, particularly in geophysical research. Augmented reality is defined as a live, direct or indirect, view of a physical, real-world environment whose elements are augmented (or supplemented) by computer-generated sensory input such as sound, video, graphics or GPS data. When paired with advanced computing techniques on cloud resources, augmented reality has the potential to improve data collection techniques, visualizations, as well as in-situ analysis for many areas of research. Google is currently a pioneer of augmented reality technology and has released beta versions of their wearable computing device, Google Glass, to a select number of developers and beta testers. This community of 'Glass Explorers' is the vehicle from which Google shapes the future of their augmented reality device. Example applications of Google Glass in geophysical research range from use as a data gathering interface in harsh climates to an on-site visualization and analysis tool. Early participation in the shaping of the Google Glass device is an opportunity for researchers to tailor this new technology to their specific needs. The purpose of this presentation is to provide geophysical researchers with a hands-on first look at Google Glass and its potential as a scientific tool. Attendees will be given an overview of the technical specifications as well as a live demonstration of the device. Potential applications to geophysical research in polar regions will be the primary focus. The presentation will conclude with an open call to participate, during which attendees may indicate interest in developing projects that integrate Google Glass into their research. Application Mockup: Penguin Counter Google Glass Augmented Reality Device
NASA Astrophysics Data System (ADS)
Cuttler, R. T. H.; Tonner, T. W. W.; Al-Naimi, F. A.; Dingwall, L. M.; Al-Hemaidi, N.
2013-07-01
The development of the Qatar National Historic Environment Record (QNHER) by the Qatar Museums Authority and the University of Birmingham in 2008 was based on a customised, bilingual Access database and ArcGIS. While both platforms are stable and well supported, neither was designed for the documentation and retrieval of cultural heritage data. As a result it was decided to develop a custom application using Open Source code. The core module of this application is now completed and is orientated towards the storage and retrieval of geospatial heritage data for the curation of heritage assets. Based on MIDAS Heritage data standards and regionally relevant thesauri, it is a truly bilingual system. Significant attention has been paid to the user interface, which is userfriendly and intuitive. Based on a suite of web services and accessed through a web browser, the system makes full use of internet resources such as Google Maps and Bing Maps. The application avoids long term vendor ''tie-ins'' and as a fully integrated data management system, is now an important tool for both cultural resource managers and heritage researchers in Qatar.
Implementing a geographical information system to assess endemic fluoride areas in Lamphun, Thailand
Theerawasttanasiri, Nonthaphat; Taneepanichskul, Surasak; Pingchai, Wichain; Nimchareon, Yuwaree; Sriwichai, Sangworn
2018-01-01
Introduction Many studies have shown that fluoride can cross the placenta and that exposure to high fluoride during pregnancy may result in premature birth and/or a low birth weight. Lamphun is one of six provinces in Thailand where natural water fluoride (WF) concentrations >10.0 mg/L were found, and it was also found that >50% of households used water with high fluoride levels. Nevertheless, geographical information system (GIS) and maps of endemic fluoride areas are lacking. We aimed to measure the fluoride level of village water supplies to assess endemic fluoride areas and present GIS with maps in Google Maps. Methods A cross-sectional survey was conducted from July 2016 to January 2017. Purpose sampling was used to identify villages of districts with WF >10.0 mg/L in the Mueang Lamphun, Pasang, and Ban Thi districts. Water samples were collected with the geolocation measured by Smart System Info. Fluoride was analyzed with an ion-selective electrode instrument using a total ionic strength adjustment buffer. WF >0.70 mg/L was used to identify unsafe drinking water and areas with high endemic fluoride levels. Descriptive statistics were used to describe the findings, and MS Excel was used to create the GIS database. Maps were created in Google Earth and presented in Google Maps. Results We found that WF concentrations ranged between 0.10–13.60 mg/L. Forty-four percent (n=439) of samples were at unsafe levels (>0.70 mg/L), and. 54% (n=303) of villages and 46% (n=79,807) of households used the unsafe drinking water. Fifty percent (n=26) of subdistricts were classified as being endemic fluoride areas. Five subdistricts were endemic fluoride areas, and in those, there were two subdistricts in which every household used unsafe drinking water. Conclusion These findings show the distribution of endemic fluoride areas and unsafe drinking water in Lamphun. This is useful for health policy authorities, local governments, and villagers and enables collaboration to resolve these issues. The GIS data are available at https://drive.google.com/open?id=1mi4Pvomf5xHZ1MQjK44pdp2xXFw&usp=sharing. PMID:29398924
Theerawasttanasiri, Nonthaphat; Taneepanichskul, Surasak; Pingchai, Wichain; Nimchareon, Yuwaree; Sriwichai, Sangworn
2018-01-01
Many studies have shown that fluoride can cross the placenta and that exposure to high fluoride during pregnancy may result in premature birth and/or a low birth weight. Lamphun is one of six provinces in Thailand where natural water fluoride (WF) concentrations >10.0 mg/L were found, and it was also found that >50% of households used water with high fluoride levels. Nevertheless, geographical information system (GIS) and maps of endemic fluoride areas are lacking. We aimed to measure the fluoride level of village water supplies to assess endemic fluoride areas and present GIS with maps in Google Maps. A cross-sectional survey was conducted from July 2016 to January 2017. Purpose sampling was used to identify villages of districts with WF >10.0 mg/L in the Mueang Lamphun, Pasang, and Ban Thi districts. Water samples were collected with the geolocation measured by Smart System Info. Fluoride was analyzed with an ion-selective electrode instrument using a total ionic strength adjustment buffer. WF >0.70 mg/L was used to identify unsafe drinking water and areas with high endemic fluoride levels. Descriptive statistics were used to describe the findings, and MS Excel was used to create the GIS database. Maps were created in Google Earth and presented in Google Maps. We found that WF concentrations ranged between 0.10-13.60 mg/L. Forty-four percent (n=439) of samples were at unsafe levels (>0.70 mg/L), and. 54% (n=303) of villages and 46% (n=79,807) of households used the unsafe drinking water. Fifty percent (n=26) of subdistricts were classified as being endemic fluoride areas. Five subdistricts were endemic fluoride areas, and in those, there were two subdistricts in which every household used unsafe drinking water. These findings show the distribution of endemic fluoride areas and unsafe drinking water in Lamphun. This is useful for health policy authorities, local governments, and villagers and enables collaboration to resolve these issues. The GIS data are available at https://drive.google.com/open?id=1mi4Pvomf5xHZ1MQjK44pdp2xXFw&usp=sharing.
Moonshot Laboratories' Lava Relief Google Mapping Project
NASA Astrophysics Data System (ADS)
Brennan, B.; Tomita, M.
2016-12-01
The Moonshot Laboratories were conceived at the University Laboratory School (ULS) on Oahu, Hawaii as way to develop creative problem solvers able to resourcefully apply 21st century technologies to respond to the problems and needs of their communities. One example of this was involved students from ULS using modern mapping and imaging technologies to assist peers who had been displaced from their own school in Pahoe on the Big Island of Hawaii. During 2015, lava flows from the eruption of Kilauea Volcano were slowly encroaching into the district of Puna in 2015. The lava flow was cutting the main town of Pahoa in half, leaving no safe routes of passage into or out of the town. One elementary school in the path of the flow was closed entirely and a new one was erected north of the flow for students living on that side. Pahoa High School students and teachers living to the north were been forced to leave their school and transfer to Kea'au High School. These students were separated from friends, family and the community they grew up in and were being thrust into a foreign environment that until then had been their local rival. Using Google Mapping technologies, Moonshot Laboratories students created a dynamic map to introduce the incoming Pahoa students to their new school in Kea'au. Elements included a stylized My Maps basemap, YouTube video descriptions of the building, videos recorded by Google Glass showing first person experiences, and immersive images of classrooms were created using 360 cameras. During the first day of orientation at Kea'au for the 200 Pahoa students, each of them were given a tablet to view the map as they toured and got to know their new campus. The methods and technologies, and more importantly innovative thinking, used to create this map have enormous potential for how to educate all students about the world around us, and the issues facing it. http://www.moonshotincubator.com/
Visualizing Mars data and imagery with Google Earth
NASA Astrophysics Data System (ADS)
Beyer, R. A.; Broxton, M.; Gorelick, N.; Hancher, M.; Lundy, M.; Kolb, E.; Moratto, Z.; Nefian, A.; Scharff, T.; Weiss-Malik, M.
2009-12-01
There is a vast store of planetary geospatial data that has been collected by NASA but is difficult to access and visualize. Virtual globes have revolutionized the way we visualize and understand the Earth, but other planetary bodies including Mars and the Moon can be visualized in similar ways. Extraterrestrial virtual globes are poised to revolutionize planetary science, bring an exciting new dimension to science education, and allow ordinary users to explore imagery being sent back to Earth by planetary science satellites. The original Google Mars Web site allowed users to view base maps of Mars via the Web, but it did not have the full features of the 3D Google Earth client. We have previously demonstrated the use of Google Earth to display Mars imagery, but now with the launch of Mars in Google Earth, there is a base set of Mars data available for anyone to work from and add to. There are a variety of global maps to choose from and display. The Terrain layer has the MOLA gridded data topography, and where available, HRSC terrain models are mosaicked into the topography. In some locations there is also meter-scale terrain derived from HiRISE stereo imagery. There is rich information in the form of the IAU nomenclature database, data for the rovers and landers on the surface, and a Spacecraft Imagery layer which contains the image outlines for all HiRISE, CTX, CRISM, HRSC, and MOC image data released to the PDS and links back to their science data. There are also features like the Traveler's Guide to Mars, Historic Maps, Guided Tours, as well as the 'Live from Mars' feature, which shows the orbital tracks of both the Mars Odyssey and Mars Reconnaissance Orbiter for a few days in the recent past. It shows where they have acquired imagery, and also some preview image data. These capabilities have obvious public outreach and education benefits, but the potential benefits of allowing planetary scientists to rapidly explore these large and varied data collections—in geological context and within a single user interface—are also becoming evident. Because anyone can produce additional KML content for use in Google Earth, scientists can customize the environment to their needs as well as publish their own processed data and results for others to use. Many scientists and organizations have begun to do this already, resulting in a useful and growing collection of planetary-science-oriented Google Earth layers.
A campus-based course in field geology
NASA Astrophysics Data System (ADS)
Richard, G. A.; Hanson, G. N.
2009-12-01
GEO 305: Field Geology offers students practical experience in the field and in the computer laboratory conducting geological field studies on the Stony Brook University campus. Computer laboratory exercises feature mapping techniques and field studies of glacial and environmental geology, and include geophysical and hydrological analysis, interpretation, and mapping. Participants learn to use direct measurement and mathematical techniques to compute the location and geometry of features and gain practical experience in representing raster imagery and vector geographic data as features on maps. Data collecting techniques in the field include the use of hand-held GPS devices, compasses, ground-penetrating radar, tape measures, pacing, and leveling devices. Assignments that utilize these skills and techniques include mapping campus geology with GPS, using Google Earth to explore our geologic context, data file management and ArcGIS, tape and compass mapping of woodland trails, pace and compass mapping of woodland trails, measuring elevation differences on a hillside, measuring geologic sections and cores, drilling through glacial deposits, using ground penetrating radar on glaciotectonic topography, mapping the local water table, and the identification and mapping of boulders. Two three-hour sessions are offered per week, apportioned as needed between lecture; discussion; guided hands-on instruction in geospatial and other software such as ArcGIS, Google Earth, spreadsheets, and custom modules such as an arc intersection calculator; outdoor data collection and mapping; and writing of illustrated reports.
A Java-based tool for creating KML files from GPS waypoints
NASA Astrophysics Data System (ADS)
Kinnicutt, P. G.; Rivard, C.; Rimer, S.
2008-12-01
Google Earth provides a free tool with powerful capabilities for visualizing geoscience images and data. Commercial software tools exist for doing sophisticated digitizing and spatial modeling , but for the purposes of presentation, visualization and overlaying aerial images with data Google Earth provides much of the functionality. Likewise, with current technologies in GPS (Global Positioning System) systems and with Google Earth Plus, it is possible to upload GPS waypoints, tracks and routes directly into Google Earth for visualization. However, older technology GPS units and even low-cost GPS units found today may lack the necessary communications interface to a computer (e.g. no Bluetooth, no WiFi, no USB, no Serial, etc.) or may have an incompatible interface, such as a Serial port but no USB adapter available. In such cases, any waypoints, tracks and routes saved in the GPS unit or recorded in a field notebook must be manually transferred to a computer for use in a GIS system or other program. This presentation describes a Java-based tool developed by the author which enables users to enter GPS coordinates in a user-friendly manner, then save these coordinates in a Keyhole MarkUp Language (KML) file format, for visualization in Google Earth. This tool either accepts user-interactive input or accepts input from a CSV (Comma Separated Value) file, which can be generated from any spreadsheet program. This tool accepts input in the form of lat/long or UTM (Universal Transverse Mercator) coordinates. This presentation describes this system's applicability through several small case studies. This free and lightweight tool simplifies the task of manually inputting GPS data into Google Earth for people working in the field without an automated mechanism for uploading the data; for instance, the user may not have internet connectivity or may not have the proper hardware or software. Since it is a Java application and not a web- based tool, it can be installed on one's field laptop and the GPS data can be manually entered without the need for internet connectivity. This tool provides a table view of the GPS data, but lacks a KML viewer to view the data overlain on top of an aerial view, as this viewer functionality is provided in Google Earth. The tool's primary contribution lies in its more convenient method for entering the GPS data manually when automated technologies are not available.
Integration of Apollo Lunar Sample Data into Google Moon
NASA Technical Reports Server (NTRS)
Dawson, Melissa D.; Todd, Nancy S.; Lofgren, Gary
2010-01-01
The Google Moon Apollo Lunar Sample Data Integration project is a continuation of the Apollo 15 Google Moon Add-On project, which provides a scientific and educational tool for the study of the Moon and its geologic features. The main goal of this project is to provide a user-friendly interface for an interactive and educational outreach and learning tool for the Apollo missions. Specifically, this project?s focus is the dissemination of information about the lunar samples collected during the Apollo missions by providing any additional information needed to enhance the Apollo mission data on Google Moon. Apollo missions 15 and 16 were chosen to be completed first due to the availability of digitized lunar sample photographs and the amount of media associated with these missions. The user will be able to learn about the lunar samples collected in these Apollo missions, as well as see videos, pictures, and 360 degree panoramas of the lunar surface depicting the lunar samples in their natural state, following collection and during processing at NASA. Once completed, these interactive data layers will be submitted for inclusion into the Apollo 15 and 16 missions on Google Moon.
Aanensen, David M; Huntley, Derek M; Feil, Edward J; al-Own, Fada'a; Spratt, Brian G
2009-09-16
Epidemiologists and ecologists often collect data in the field and, on returning to their laboratory, enter their data into a database for further analysis. The recent introduction of mobile phones that utilise the open source Android operating system, and which include (among other features) both GPS and Google Maps, provide new opportunities for developing mobile phone applications, which in conjunction with web applications, allow two-way communication between field workers and their project databases. Here we describe a generic framework, consisting of mobile phone software, EpiCollect, and a web application located within www.spatialepidemiology.net. Data collected by multiple field workers can be submitted by phone, together with GPS data, to a common web database and can be displayed and analysed, along with previously collected data, using Google Maps (or Google Earth). Similarly, data from the web database can be requested and displayed on the mobile phone, again using Google Maps. Data filtering options allow the display of data submitted by the individual field workers or, for example, those data within certain values of a measured variable or a time period. Data collection frameworks utilising mobile phones with data submission to and from central databases are widely applicable and can give a field worker similar display and analysis tools on their mobile phone that they would have if viewing the data in their laboratory via the web. We demonstrate their utility for epidemiological data collection and display, and briefly discuss their application in ecological and community data collection. Furthermore, such frameworks offer great potential for recruiting 'citizen scientists' to contribute data easily to central databases through their mobile phone.
Perils of using speed zone data to assess real-world compliance to speed limits.
Chevalier, Anna; Clarke, Elizabeth; Chevalier, Aran John; Brown, Julie; Coxon, Kristy; Ivers, Rebecca; Keay, Lisa
2017-11-17
Real-world driving studies, including those involving speeding alert devices and autonomous vehicles, can gauge an individual vehicle's speeding behavior by comparing measured speed with mapped speed zone data. However, there are complexities with developing and maintaining a database of mapped speed zones over a large geographic area that may lead to inaccuracies within the data set. When this approach is applied to large-scale real-world driving data or speeding alert device data to determine speeding behavior, these inaccuracies may result in invalid identification of speeding. We investigated speeding events based on service provider speed zone data. We compared service provider speed zone data (Speed Alert by Smart Car Technologies Pty Ltd., Ultimo, NSW, Australia) against a second set of speed zone data (Google Maps Application Programming Interface [API] mapped speed zones). We found a systematic error in the zones where speed limits of 50-60 km/h, typical of local roads, were allocated to high-speed motorways, which produced false speed limits in the speed zone database. The result was detection of false-positive high-range speeding. Through comparison of the service provider speed zone data against a second set of speed zone data, we were able to identify and eliminate data most affected by this systematic error, thereby establishing a data set of speeding events with a high level of sensitivity (a true positive rate of 92% or 6,412/6,960). Mapped speed zones can be a source of error in real-world driving when examining vehicle speed. We explored the types of inaccuracies found within speed zone data and recommend that a second set of speed zone data be utilized when investigating speeding behavior or developing mapped speed zone data to minimize inaccuracy in estimates of speeding.
NASA Astrophysics Data System (ADS)
Shelestov, Andrii; Lavreniuk, Mykola; Kussul, Nataliia; Novikov, Alexei; Skakun, Sergii
2017-02-01
Many applied problems arising in agricultural monitoring and food security require reliable crop maps at national or global scale. Large scale crop mapping requires processing and management of large amount of heterogeneous satellite imagery acquired by various sensors that consequently leads to a “Big Data” problem. The main objective of this study is to explore efficiency of using the Google Earth Engine (GEE) platform when classifying multi-temporal satellite imagery with potential to apply the platform for a larger scale (e.g. country level) and multiple sensors (e.g. Landsat-8 and Sentinel-2). In particular, multiple state-of-the-art classifiers available in the GEE platform are compared to produce a high resolution (30 m) crop classification map for a large territory ( 28,100 km2 and 1.0 M ha of cropland). Though this study does not involve large volumes of data, it does address efficiency of the GEE platform to effectively execute complex workflows of satellite data processing required with large scale applications such as crop mapping. The study discusses strengths and weaknesses of classifiers, assesses accuracies that can be achieved with different classifiers for the Ukrainian landscape, and compares them to the benchmark classifier using a neural network approach that was developed in our previous studies. The study is carried out for the Joint Experiment of Crop Assessment and Monitoring (JECAM) test site in Ukraine covering the Kyiv region (North of Ukraine) in 2013. We found that Google Earth Engine (GEE) provides very good performance in terms of enabling access to the remote sensing products through the cloud platform and providing pre-processing; however, in terms of classification accuracy, the neural network based approach outperformed support vector machine (SVM), decision tree and random forest classifiers available in GEE.
Large-scale virtual screening on public cloud resources with Apache Spark.
Capuccini, Marco; Ahmed, Laeeq; Schaal, Wesley; Laure, Erwin; Spjuth, Ola
2017-01-01
Structure-based virtual screening is an in-silico method to screen a target receptor against a virtual molecular library. Applying docking-based screening to large molecular libraries can be computationally expensive, however it constitutes a trivially parallelizable task. Most of the available parallel implementations are based on message passing interface, relying on low failure rate hardware and fast network connection. Google's MapReduce revolutionized large-scale analysis, enabling the processing of massive datasets on commodity hardware and cloud resources, providing transparent scalability and fault tolerance at the software level. Open source implementations of MapReduce include Apache Hadoop and the more recent Apache Spark. We developed a method to run existing docking-based screening software on distributed cloud resources, utilizing the MapReduce approach. We benchmarked our method, which is implemented in Apache Spark, docking a publicly available target receptor against [Formula: see text]2.2 M compounds. The performance experiments show a good parallel efficiency (87%) when running in a public cloud environment. Our method enables parallel Structure-based virtual screening on public cloud resources or commodity computer clusters. The degree of scalability that we achieve allows for trying out our method on relatively small libraries first and then to scale to larger libraries. Our implementation is named Spark-VS and it is freely available as open source from GitHub (https://github.com/mcapuccini/spark-vs).Graphical abstract.
Using Google Earth for Submarine Operations at Pavilion Lake
NASA Astrophysics Data System (ADS)
Deans, M. C.; Lees, D. S.; Fong, T.; Lim, D. S.
2009-12-01
During the July 2009 Pavilion Lake field test, we supported submarine "flight" operations using Google Earth. The Intelligent Robotics Group at NASA Ames has experience with ground data systems for NASA missions, earth analog field tests, disaster response, and the Gigapan camera system. Leveraging this expertise and existing software, we put together a set of tools to support sub tracking and mapping, called the "Surface Data System." This system supports flight planning, real time flight operations, and post-flight analysis. For planning, we make overlays of the regional bedrock geology, sonar bathymetry, and sonar backscatter maps that show geology, depth, and structure of the bottom. Placemarks show the mooring locations for start and end points. Flight plans are shown as polylines with icons for waypoints. Flight tracks and imagery from previous field seasons are embedded in the map for planning follow-on activities. These data provide context for flight planning. During flights, sub position is updated every 5 seconds from the nav computer on the chase boat. We periodically update tracking KML files and refresh them with network links. A sub icon shows current location of the sub. A compass rose shows bearings to indicate heading to the next waypoint. A "Science Stenographer" listens on the voice loop and transcribes significant observations in real time. Observations called up to the surface immediately appear on the map as icons with date, time, position, and what was said. After each flight, the science back room immediately has the flight track and georeferenced notes from the pilots. We add additional information in post-processing. The submarines record video continuously, with "event" timestamps marked by the pilot. We cross-correlate the event timestamps with position logs to geolocate events and put a preview image and compressed video clip into the map. Animated flight tracks are also generated, showing timestamped position and providing timelapse playback of the flight. Neogeography tools are increasing in popularity and offer an excellent platform for geoinformatics. The scientists on the team are already familiar with Google Earth, eliminating up-front training on new tools. The flight maps and archived data are available immediately and in a usable format. Google Earth provides lots of measurement tools, annotation tools, and other built-in functions that we can use to create and analyze the map. All of this information is saved to a shared filesystem so that everyone on the team has access to all of the same map data. After the field season, the map data will be used by the team to analyse and correlate information from across the lake and across different flights to support their research, and to plan next year's activities.
Streets? Where We're Going, We Don't Need Streets
NASA Astrophysics Data System (ADS)
Bailey, J.
2017-12-01
In 2007 Google Street View started as a project to provide 360-degree imagery along streets, but in the decade since has evolved into a platform through which to explore everywhere from the slope of everest, to the middle of the Amazon rainforest to under the ocean. As camera technology has evolved it has also become a tool for ground truthing maps, and provided scientific observations, storytelling and education. The Google Street View "special collects" team has undertaken increasingly more challenging projects across 80+ countries and every continent. All of which culminated in possibly the most ambitious collection yet, the capture of Street View on board the International Space Station. Learn about the preparation and obstacles behind this and other special collects. Explore these datasets through both Google Earth and Google Expeditions VR, an educational tool to take students on virtual field trips using 360 degree imagery.
Scales, David; Zelenev, Alexei; Brownstein, John S.
2013-01-01
Background This is the first study quantitatively evaluating the effect that media-related limitations have on data from an automated epidemic intelligence system. Methods We modeled time series of HealthMap's two main data feeds, Google News and Moreover, to test for evidence of two potential limitations: first, human resources constraints, and second, high-profile outbreaks “crowding out” coverage of other infectious diseases. Results Google News events declined by 58.3%, 65.9%, and 14.7% on Saturday, Sunday and Monday, respectively, relative to other weekdays. Events were reduced by 27.4% during Christmas/New Years weeks and 33.6% lower during American Thanksgiving week than during an average week for Google News. Moreover data yielded similar results with the addition of Memorial Day (US) being associated with a 36.2% reduction in events. Other holiday effects were not statistically significant. We found evidence for a crowd out phenomenon for influenza/H1N1, where a 50% increase in influenza events corresponded with a 4% decline in other disease events for Google News only. Other prominent diseases in this database – avian influenza (H5N1), cholera, or foodborne illness – were not associated with a crowd out phenomenon. Conclusions These results provide quantitative evidence for the limited impact of editorial biases on HealthMap's web-crawling epidemic intelligence. PMID:24206612
Predicting the performance of local seismic networks using Matlab and Google Earth.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Chael, Eric Paul
2009-11-01
We have used Matlab and Google Earth to construct a prototype application for modeling the performance of local seismic networks for monitoring small, contained explosions. Published equations based on refraction experiments provide estimates of peak ground velocities as a function of event distance and charge weight. Matlab routines implement these relations to calculate the amplitudes across a network of stations from sources distributed over a geographic grid. The amplitudes are then compared to ambient noise levels at the stations, and scaled to determine the smallest yield that could be detected at each source location by a specified minimum number ofmore » stations. We use Google Earth as the primary user interface, both for positioning the stations of a hypothetical local network, and for displaying the resulting detection threshold contours.« less
NASA Astrophysics Data System (ADS)
Quang Truong, Xuan; Luan Truong, Xuan; Nguyen, Tuan Anh; Nguyen, Dinh Tuan; Cong Nguyen, Chi
2017-12-01
The objective of this study is to design and implement a WebGIS Decision Support System (WDSS) for reducing uncertainty and supporting to improve the quality of exploration decisions in the Sin-Quyen copper mine, northern Vietnam. The main distinctive feature of the Sin-Quyen deposit is an unusual composition of ores. Computer and software applied to the exploration problem have had a significant impact on the exploration process over the past 25 years, but up until now, no online system has been undertaken. The system was completely built on open source technology and the Open Geospatial Consortium Web Services (OWS). The input data includes remote sensing (RS), Geographical Information System (GIS) and data from drillhole explorations, the drillhole exploration data sets were designed as a geodatabase and stored in PostgreSQL. The WDSS must be able to processed exploration data and support users to access 2-dimensional (2D) or 3-dimensional (3D) cross-sections and map of boreholles exploration data and drill holes. The interface was designed in order to interact with based maps (e.g., Digital Elevation Model, Google Map, OpenStreetMap) and thematic maps (e.g., land use and land cover, administrative map, drillholes exploration map), and to provide GIS functions (such as creating a new map, updating an existing map, querying and statistical charts). In addition, the system provides geological cross-sections of ore bodies based on Inverse Distance Weighting (IDW), nearest neighbour interpolation and Kriging methods (e.g., Simple Kriging, Ordinary Kriging, Indicator Kriging and CoKriging). The results based on data available indicate that the best estimation method (of 23 borehole exploration data sets) for estimating geological cross-sections of ore bodies in Sin-Quyen copper mine is Ordinary Kriging. The WDSS could provide useful information to improve drilling efficiency in mineral exploration and for management decision making.
Using GeoRSS feeds to distribute house renting and selling information based on Google map
NASA Astrophysics Data System (ADS)
Nong, Yu; Wang, Kun; Miao, Lei; Chen, Fei
2007-06-01
Geographically Encoded Objects RSS (GeoRSS) is a way to encode location in RSS feeds. RSS is a widely supported format for syndication of news and weblogs, and is extendable to publish any sort of itemized data. When Weblogs explode since RSS became new portals, Geo-tagged feed is necessary to show the location that story tells. Geographically Encoded Objects adopts the core of RSS framework, making itself the map annotations specified in the RSS XML format. The case studied illuminates that GeoRSS could be maximally concise in representation and conception, so it's simple to manipulate generation and then mashup GeoRSS feeds with Google Map through API to show the real estate information with other attribute in the information window. After subscribe to feeds of concerned subjects, users could easily check for new bulletin showing on map through syndication. The primary design goal of GeoRSS is to make spatial data creation as easy as regular Web content development. However, it does more for successfully bridging the gap between traditional GIS professionals and amateurs, Web map hackers, and numerous services that enable location-based content for its simplicity and effectiveness.
Real-time bus location monitoring using Arduino
NASA Astrophysics Data System (ADS)
Ibrahim, Mohammad Y. M.; Audah, Lukman
2017-09-01
The Internet of Things (IoT) is the network of objects, such as a vehicles, mobile devices, and buildings that have electronic components, software, and network connectivity that enable them to collect data, run commands, and be controlled through the Internet. Controlling physical items from the Internet will increase efficiency and save time. The growing number of devices used by people increases the practicality of having IoT devices on the market. The IoT is also an opportunity to develop products that can save money and time and increase work efficiency. Initially, they need more efficiency for real-time bus location systems, especially in university campuses. This system can easily find the accurate locations of and distances between each bus stop and the estimated time to reach a new location. This system has been separated into two parts, which are the hardware and the software. The hardware parts are the Arduino Uno and the Global Positioning System (GPS), while Google Earth and GpsGate are the software parts. The GPS continuously takes input data from the satellite and stores the latitude and longitude values in the Arduino Uno. If we want to track the vehicle, we need to send the longitude and latitude as a message to the Google Earth software to convert these into maps for navigation. Once the Arduino Uno is activated, it takes the last received latitude and longitude positions' values from GpsGate and sends a message to Google Earth. Once the message has been sent to Google Earth, the current location will be shown, and navigation will be activated automatically. Then it will be broadcast using ManyCam, Google+ Hangouts, and YouTube, as well as Facebook, and appear to users. The additional features use Google Forms for determining problems faced by students, who can also take immediate action against the responsible department. Then after several successful simulations, the results will be shown in real time on a map.
Embracing Open Software Development in Solar Physics
NASA Astrophysics Data System (ADS)
Hughitt, V. K.; Ireland, J.; Christe, S.; Mueller, D.
2012-12-01
We discuss two ongoing software projects in solar physics that have adopted best practices of the open source software community. The first, the Helioviewer Project, is a powerful data visualization tool which includes online and Java interfaces inspired by Google Maps (tm). This effort allows users to find solar features and events of interest, and download the corresponding data. Having found data of interest, the user now has to analyze it. The dominant solar data analysis platform is an open-source library called SolarSoft (SSW). Although SSW itself is open-source, the programming language used is IDL, a proprietary language with licensing costs that are prohibative for many institutions and individuals. SSW is composed of a collection of related scripts written by missions and individuals for solar data processing and analysis, without any consistent data structures or common interfaces. Further, at the time when SSW was initially developed, many of the best software development processes of today (mirrored and distributed version control, unit testing, continuous integration, etc.) were not standard, and have not since been adopted. The challenges inherent in developing SolarSoft led to a second software project known as SunPy. SunPy is an open-source Python-based library which seeks to create a unified solar data analysis environment including a number of core datatypes such as Maps, Lightcurves, and Spectra which have consistent interfaces and behaviors. By taking advantage of the large and sophisticated body of scientific software already available in Python (e.g. SciPy, NumPy, Matplotlib), and by adopting many of the best practices refined in open-source software development, SunPy has been able to develop at a very rapid pace while still ensuring a high level of reliability. The Helioviewer Project and SunPy represent two pioneering technologies in solar physics - simple yet flexible data visualization and a powerful, new data analysis environment. We discuss the development of both these efforts and how they are beginning to influence the solar physics community.
Chang, Aileen Y; Parrales, Maria E; Jimenez, Javier; Sobieszczyk, Magdalena E; Hammer, Scott M; Copenhaver, David J; Kulkarni, Rajan P
2009-01-01
Background Dengue fever is a mosquito-borne illness that places significant burden on tropical developing countries with unplanned urbanization. A surveillance system using Google Earth and GIS mapping technologies was developed in Nicaragua as a management tool. Methods and Results Satellite imagery of the town of Bluefields, Nicaragua captured from Google Earth was used to create a base-map in ArcGIS 9. Indices of larval infestation, locations of tire dumps, cemeteries, large areas of standing water, etc. that may act as larval development sites, and locations of the homes of dengue cases collected during routine epidemiologic surveying were overlaid onto this map. Visual imagery of the location of dengue cases, larval infestation, and locations of potential larval development sites were used by dengue control specialists to prioritize specific neighborhoods for targeted control interventions. Conclusion This dengue surveillance program allows public health workers in resource-limited settings to accurately identify areas with high indices of mosquito infestation and interpret the spatial relationship of these areas with potential larval development sites such as garbage piles and large pools of standing water. As a result, it is possible to prioritize control strategies and to target interventions to highest risk areas in order to eliminate the likely origin of the mosquito vector. This program is well-suited for resource-limited settings since it utilizes readily available technologies that do not rely on Internet access for daily use and can easily be implemented in many developing countries for very little cost. PMID:19627614
Kamadjeu, Raoul
2009-01-01
Background The use of GIS in public health is growing, a consequence of a rapidly evolving technology and increasing accessibility to a wider audience. Google Earth™ (GE) is becoming an important mapping infrastructure for public health. However, generating traditional public health maps for GE is still beyond the reach of most public health professionals. In this paper, we explain, through the example of polio eradication activities in the Democratic Republic of Congo, how we used GE Earth as a planning tool and we share the methods used to generate public health maps. Results The use of GE improved field operations and resulted in better dispatch of vaccination teams and allocation of resources. It also allowed the creation of maps of high quality for advocacy, training and to help understand the spatiotemporal relationship between all the entities involved in the polio outbreak and response. Conclusion GE has the potential of making mapping available to a new set of public health users in developing countries. High quality and free satellite imagery, rich features including Keyhole Markup Language or image overlay provide a flexible but yet powerful platform that set it apart from traditional GIS tools and this power is still to be fully harnessed by public health professionals. PMID:19161606
Franklin, Erik C; Stat, Michael; Pochon, Xavier; Putnam, Hollie M; Gates, Ruth D
2012-03-01
The genus Symbiodinium encompasses a group of unicellular, photosynthetic dinoflagellates that are found free living or in hospite with a wide range of marine invertebrate hosts including scleractinian corals. We present GeoSymbio, a hybrid web application that provides an online, easy to use and freely accessible interface for users to discover, explore and utilize global geospatial bioinformatic and ecoinformatic data on Symbiodinium-host symbioses. The novelty of this application lies in the combination of a variety of query and visualization tools, including dynamic searchable maps, data tables with filter and grouping functions, and interactive charts that summarize the data. Importantly, this application is hosted remotely or 'in the cloud' using Google Apps, and therefore does not require any specialty GIS, web programming or data programming expertise from the user. The current version of the application utilizes Symbiodinium data based on the ITS2 genetic marker from PCR-based techniques, including denaturing gradient gel electrophoresis, sequencing and cloning of specimens collected during 1982-2010. All data elements of the application are also downloadable as spatial files, tables and nucleic acid sequence files in common formats for desktop analysis. The application provides a unique tool set to facilitate research on the basic biology of Symbiodinium and expedite new insights into their ecology, biogeography and evolution in the face of a changing global climate. GeoSymbio can be accessed at https://sites.google.com/site/geosymbio/. © 2011 Blackwell Publishing Ltd.
TouchTerrain: A simple web-tool for creating 3D-printable topographic models
NASA Astrophysics Data System (ADS)
Hasiuk, Franciszek J.; Harding, Chris; Renner, Alex Raymond; Winer, Eliot
2017-12-01
An open-source web-application, TouchTerrain, was developed to simplify the production of 3D-printable terrain models. Direct Digital Manufacturing (DDM) using 3D Printers can change how geoscientists, students, and stakeholders interact with 3D data, with the potential to improve geoscience communication and environmental literacy. No other manufacturing technology can convert digital data into tangible objects quickly at relatively low cost; however, the expertise necessary to produce a 3D-printed terrain model can be a substantial burden: knowledge of geographical information systems, computer aided design (CAD) software, and 3D printers may all be required. Furthermore, printing models larger than the build volume of a 3D printer can pose further technical hurdles. The TouchTerrain web-application simplifies DDM for elevation data by generating digital 3D models customized for a specific 3D printer's capabilities. The only required user input is the selection of a region-of-interest using the provided web-application with a Google Maps-style interface. Publically available digital elevation data is processed via the Google Earth Engine API. To allow the manufacture of 3D terrain models larger than a 3D printer's build volume the selected area can be split into multiple tiles without third-party software. This application significantly reduces the time and effort required for a non-expert like an educator to obtain 3D terrain models for use in class. The web application is deployed at http://touchterrain.geol.iastate.edu/.
A Google Glass navigation system for ultrasound and fluorescence dual-mode image-guided surgery
NASA Astrophysics Data System (ADS)
Zhang, Zeshu; Pei, Jing; Wang, Dong; Hu, Chuanzhen; Ye, Jian; Gan, Qi; Liu, Peng; Yue, Jian; Wang, Benzhong; Shao, Pengfei; Povoski, Stephen P.; Martin, Edward W.; Yilmaz, Alper; Tweedle, Michael F.; Xu, Ronald X.
2016-03-01
Surgical resection remains the primary curative intervention for cancer treatment. However, the occurrence of a residual tumor after resection is very common, leading to the recurrence of the disease and the need for re-resection. We develop a surgical Google Glass navigation system that combines near infrared fluorescent imaging and ultrasonography for intraoperative detection of sites of tumor and assessment of surgical resection boundaries, well as for guiding sentinel lymph node (SLN) mapping and biopsy. The system consists of a monochromatic CCD camera, a computer, a Google Glass wearable headset, an ultrasonic machine and an array of LED light sources. All the above components, except the Google Glass, are connected to a host computer by a USB or HDMI port. Wireless connection is established between the glass and the host computer for image acquisition and data transport tasks. A control program is written in C++ to call OpenCV functions for image calibration, processing and display. The technical feasibility of the system is tested in both tumor simulating phantoms and in a human subject. When the system is used for simulated phantom resection tasks, the tumor boundaries, invisible to the naked eye, can be clearly visualized with the surgical Google Glass navigation system. This system has also been used in an IRB approved protocol in a single patient during SLN mapping and biopsy in the First Affiliated Hospital of Anhui Medical University, demonstrating the ability to successfully localize and resect all apparent SLNs. In summary, our tumor simulating phantom and human subject studies have demonstrated the technical feasibility of successfully using the proposed goggle navigation system during cancer surgery.
Towards a geospatial wikipedia
NASA Astrophysics Data System (ADS)
Fritz, S.; McCallum, I.; Schill, C.; Perger, C.; Kraxner, F.; Obersteiner, M.
2009-04-01
Based on the Google Earth (http://earth.google.com) platform we have developed a geospatial Wikipedia (geo-wiki.org). The tool allows everybody in the world to contribute to spatial validation and is made available to the internet community interested in that task. We illustrate how this tool can be used for different applications. In our first application we combine uncertainty hotspot information from three global land cover datasets (GLC, MODIS, GlobCover). With an ever increasing amount of high resolution images available on Google Earth, it is becoming increasingly possible to distinguish land cover features with a high degree of accuracy. We first direct the land cover validation community to certain hotspots of land cover uncertainty and then ask them to fill in a small popup menu on type of land cover, possibly a picture at that location with the different cardinal points as well as date and what type of validation was chosen (google earth imagery/panoramio or if the person has ground truth data). We have implemented the tool via a land cover validation community at FACEBOOK which is based on a snowball system which allows the tracking of individuals and the possibility to ignore users which misuse the system. In a second application we illustrate how the tool could possibly be used for mapping malaria occurrence and small water bodies as well as overall malaria risk. For this application we have implemented a polygon as well as attribute function using Google maps as along with virtual earth using openlayers. The third application deals with illegal logging and how an alert system for illegal logging detection within a certain land tenure system could be implemented. Here we show how the tool can be used to document illegal logging via a YouTube video.
NASA Astrophysics Data System (ADS)
Erickson, T. A.; Granger, B.; Grout, J.; Corlay, S.
2017-12-01
The volume of Earth science data gathered from satellites, aircraft, drones, and field instruments continues to increase. For many scientific questions in the Earth sciences, managing this large volume of data is a barrier to progress, as it is difficult to explore and analyze large volumes of data using the traditional paradigm of downloading datasets to a local computer for analysis. Furthermore, methods for communicating Earth science algorithms that operate on large datasets in an easily understandable and reproducible way are needed. Here we describe a system for developing, interacting, and sharing well-documented Earth Science algorithms that combines existing software components: Jupyter Notebook: An open-source, web-based environment that supports documents that combine code and computational results with text narrative, mathematics, images, and other media. These notebooks provide an environment for interactive exploration of data and development of well documented algorithms. Jupyter Widgets / ipyleaflet: An architecture for creating interactive user interface controls (such as sliders, text boxes, etc.) in Jupyter Notebooks that communicate with Python code. This architecture includes a default set of UI controls (sliders, dropboxes, etc.) as well as APIs for building custom UI controls. The ipyleaflet project is one example that offers a custom interactive map control that allows a user to display and manipulate geographic data within the Jupyter Notebook. Google Earth Engine: A cloud-based geospatial analysis platform that provides access to petabytes of Earth science data via a Python API. The combination of Jupyter Notebooks, Jupyter Widgets, ipyleaflet, and Google Earth Engine makes it possible to explore and analyze massive Earth science datasets via a web browser, in an environment suitable for interactive exploration, teaching, and sharing. Using these environments can make Earth science analyses easier to understand and reproducible, which may increase the rate of scientific discoveries and the transition of discoveries into real-world impacts.
NASA Astrophysics Data System (ADS)
Brumana, R.; Santana Quintero, M.; Barazzetti, L.; Previtali, M.; Banfi, F.; Oreni, D.; Roels, D.; Roncoroni, F.
2015-08-01
Landscapes are dynamic entities, stretching and transforming across space and time, and need to be safeguarded as living places for the future, with interaction of human, social and economic dimensions. To have a comprehensive landscape evaluation several open data are needed, each one characterized by its own protocol, service interface, limiting or impeding this way interoperability and their integration. Indeed, nowadays the development of websites targeted to landscape assessment and touristic purposes requires many resources in terms of time, cost and IT skills to be implemented at different scales. For this reason these applications are limited to few cases mainly focusing on worldwide known touristic sites. The capability to spread the development of web-based multimedia virtual museum based on geospatial data relies for the future being on the possibility to discover the needed geo-spatial data through a single point of access in an homogenous way. In this paper the proposed innovative approach may facilitate the access to open data in a homogeneous way by means of specific components (the brokers) performing interoperability actions required to interconnect heterogeneous data sources. In the specific case study here analysed it has been implemented an interface to migrate a geo-swat chart based on local and regional geographic information into an user friendly Google Earth©-based infrastructure, integrating ancient cadastres and modern cartography, accessible by professionals and tourists via web and also via portable devices like tables and smartphones. The general aim of this work on the case study on the Lake of Como (Tremezzina municipality), is to boost the integration of assessment methodologies with digital geo-based technologies of map correlation for the multimedia ecomuseum system accessible via web. The developed WebGIS system integrates multi-scale and multi-temporal maps with different information (cultural, historical, landscape levels) represented by thematic icons allowing to transfer the richness of the landscape value to both tourists and professionals.
3D Orbit Visualization for Earth-Observing Missions
NASA Technical Reports Server (NTRS)
Jacob, Joseph C.; Plesea, Lucian; Chafin, Brian G.; Weiss, Barry H.
2011-01-01
This software visualizes orbit paths for the Orbiting Carbon Observatory (OCO), but was designed to be general and applicable to any Earth-observing mission. The software uses the Google Earth user interface to provide a visual mechanism to explore spacecraft orbit paths, ground footprint locations, and local cloud cover conditions. In addition, a drill-down capability allows for users to point and click on a particular observation frame to pop up ancillary information such as data product filenames and directory paths, latitude, longitude, time stamp, column-average dry air mole fraction of carbon dioxide, and solar zenith angle. This software can be integrated with the ground data system for any Earth-observing mission to automatically generate daily orbit path data products in Google Earth KML format. These KML data products can be directly loaded into the Google Earth application for interactive 3D visualization of the orbit paths for each mission day. Each time the application runs, the daily orbit paths are encapsulated in a KML file for each mission day since the last time the application ran. Alternatively, the daily KML for a specified mission day may be generated. The application automatically extracts the spacecraft position and ground footprint geometry as a function of time from a daily Level 1B data product created and archived by the mission s ground data system software. In addition, ancillary data, such as the column-averaged dry air mole fraction of carbon dioxide and solar zenith angle, are automatically extracted from a Level 2 mission data product. Zoom, pan, and rotate capability are provided through the standard Google Earth interface. Cloud cover is indicated with an image layer from the MODIS (Moderate Resolution Imaging Spectroradiometer) aboard the Aqua satellite, which is automatically retrieved from JPL s OnEarth Web service.
NASA Astrophysics Data System (ADS)
Landsfeld, M. F.; Hegewisch, K.; Daudert, B.; Morton, C.; Husak, G. J.; Friedrichs, M.; Funk, C. C.; Huntington, J. L.; Abatzoglou, J. T.; Verdin, J. P.
2016-12-01
The Famine Early Warning Systems Network (FEWS NET) focuses on food insecurity in developing nations and provides objective, evidence-based analysis to help government decision-makers and relief agencies plan for and respond to humanitarian emergencies. The network of FEWS NET analysts and scientists require flexible, interactive tools to aid in their monitoring and research efforts. Because they often work in bandwidth-limited regions, lightweight Internet tools and services that bypass the need for downloading massive datasets are preferred for their work. To support food security analysis FEWS NET developed a custom interface for the Google Earth Engine (GEE). GEE is a platform developed by Google to support scientific analysis of environmental data in their cloud computing environment. This platform allows scientists and independent researchers to mine massive collections of environmental data, leveraging Google's vast computational resources for purposes of detecting changes and monitoring the Earth's surface and climate. GEE hosts an enormous amount of satellite imagery and climate archives, one of which is the Climate Hazards Group Infrared Precipitation with Stations dataset (CHIRPS). CHIRPS precipitation dataset is a key input for FEWS NET monitoring and forecasting efforts. In this talk we introduce the FEWS Engine interface. We present an application that highlights the utility of FEWS Engine for forecasting the upcoming seasonal precipitation of southern Africa. Specifically, the current state of ENSO is assessed and used to identify similar historical seasons. The FEWS Engine compositing tool is used to examine rainfall and other environmental data for these analog seasons. The application illustrates the unique benefits of using FEWS Engine for on-the-fly food security scenario development.
NASA Astrophysics Data System (ADS)
Gabriel, Patrik; Backhaus, Udo
2013-04-01
Nearly every smartphone is now GPS capable. The widespread use of GPS navigation has developed alongside less expensive hardware and user-friendly software interfaces, which may help to bring scientific research and teaching closer to real life.
Gopalakrishnan, V; Baskaran, R; Venkatraman, B
2016-08-01
A decision support system (DSS) is implemented in Radiological Safety Division, Indira Gandhi Centre for Atomic Research for providing guidance for emergency decision making in case of an inadvertent nuclear accident. Real time gamma dose rate measurement around the stack is used for estimating the radioactive release rate (source term) by using inverse calculation. Wireless gamma dose logging network is designed, implemented, and installed around the Madras Atomic Power Station reactor stack to continuously acquire the environmental gamma dose rate and the details are presented in the paper. The network uses XBee-Pro wireless modules and PSoC controller for wireless interfacing, and the data are logged at the base station. A LabView based program is developed to receive the data, display it on the Google Map, plot the data over the time scale, and register the data in a file to share with DSS software. The DSS at the base station evaluates the real time source term to assess radiation impact.
Juicebox.js Provides a Cloud-Based Visualization System for Hi-C Data.
Robinson, James T; Turner, Douglass; Durand, Neva C; Thorvaldsdóttir, Helga; Mesirov, Jill P; Aiden, Erez Lieberman
2018-02-28
Contact mapping experiments such as Hi-C explore how genomes fold in 3D. Here, we introduce Juicebox.js, a cloud-based web application for exploring the resulting datasets. Like the original Juicebox application, Juicebox.js allows users to zoom in and out of such datasets using an interface similar to Google Earth. Juicebox.js also has many features designed to facilitate data reproducibility and sharing. Furthermore, Juicebox.js encodes the exact state of the browser in a shareable URL. Creating a public browser for a new Hi-C dataset does not require coding and can be accomplished in under a minute. The web app also makes it possible to create interactive figures online that can complement or replace ordinary journal figures. When combined with Juicer, this makes the entire process of data analysis transparent, insofar as every step from raw reads to published figure is publicly available as open source code. Copyright © 2018 The Authors. Published by Elsevier Inc. All rights reserved.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Gopalakrishnan, V.; Baskaran, R.; Venkatraman, B.
A decision support system (DSS) is implemented in Radiological Safety Division, Indira Gandhi Centre for Atomic Research for providing guidance for emergency decision making in case of an inadvertent nuclear accident. Real time gamma dose rate measurement around the stack is used for estimating the radioactive release rate (source term) by using inverse calculation. Wireless gamma dose logging network is designed, implemented, and installed around the Madras Atomic Power Station reactor stack to continuously acquire the environmental gamma dose rate and the details are presented in the paper. The network uses XBee–Pro wireless modules and PSoC controller for wireless interfacing,more » and the data are logged at the base station. A LabView based program is developed to receive the data, display it on the Google Map, plot the data over the time scale, and register the data in a file to share with DSS software. The DSS at the base station evaluates the real time source term to assess radiation impact.« less
The use of mobile technology for tourism development (case study: banyumas regency)
NASA Astrophysics Data System (ADS)
Gunawan, H.
2018-01-01
Banyumas is regency in Central Java Province and Purwokerto Capital City. Banyumas has a lot of tourism areas, but not all of the areas are known by tourists because the lack of information. The information on Banyumas government’s website is not updated and uncomplete, so that it needs other information services for the tourism information in Banyumas. An android-based application Banyumas Travel Guide (BMSGuide) is a location-based application to help the people can access the information whenever and wherever they are. The service in this application is using Android which is platform that is now developing rapidly with interface User Friendly and the price is affordable. The application will access Google Maps and show the user location, destination location along with the information, and the navigation to the location. The information is gotten by accessing the satellite with GPS (Global Positioning System) tool of the user’s headset. By using BMSGuide application, the information service of tourism location and the supporting place around is served.
Smart Garbage Based on Internet of Things (IoT)
NASA Astrophysics Data System (ADS)
Joni, K.; Haryanto; Rohim, D. F.
2018-01-01
Trash is one of serious problems in Indonesia. It leads to another problems which deal with social, culture and economic, most of cities inindonesia have some problems in managing the trush. It is because the government doesn’t have the standard of trash management. The schedule to take the trash usually discover a problems with a lateness. The lack of trashmanagement and people’s habbit who don’t care about the trash impact effect to more serious problems that dealwith the trash. Such as, enviromental problems. Based on the problems the researcher makes a system called “Smart Garbage Based On Internet of things”. The system applies the technology of temporary garbage used internal system. This system creates in order to fix the problem of punctuality. Which makes the trash management more effective and afficient. The process of uploading the information in the internal awerness of based on the “real time” ativity this system can increase the peurity and health. Because the trash is controlled and not aculamulated arround people. The ultrasonic sensor and nodeMCU are used as a module of IoT to discover the temporary garbage. Grapical user Interface dekstop and android are also used in order to observe the level of trash in temporary garbage. The notification system can be access in mobile and dekstop interface. Furthermore the temporary garbage location can find by using google map the result of research shows that the system is running well. The system can onilne 12 hours/day along 2 weeks in exsperiment. The data sent by online based on “real time” activity and the result of mapping and tracking on android show the location of garbage.
Analysis of world terror networks from the reduced Google matrix of Wikipedia
NASA Astrophysics Data System (ADS)
El Zant, Samer; Frahm, Klaus M.; Jaffrès-Runser, Katia; Shepelyansky, Dima L.
2018-01-01
We apply the reduced Google matrix method to analyze interactions between 95 terrorist groups and determine their relationships and influence on 64 world countries. This is done on the basis of the Google matrix of the English Wikipedia (2017) composed of 5 416 537 articles which accumulate a great part of global human knowledge. The reduced Google matrix takes into account the direct and hidden links between a selection of 159 nodes (articles) appearing due to all paths of a random surfer moving over the whole network. As a result we obtain the network structure of terrorist groups and their relations with selected countries including hidden indirect links. Using the sensitivity of PageRank to a weight variation of specific links we determine the geopolitical sensitivity and influence of specific terrorist groups on world countries. The world maps of the sensitivity of various countries to influence of specific terrorist groups are obtained. We argue that this approach can find useful application for more extensive and detailed data bases analysis.
Local Air Quality Conditions and Forecasts
... Monitor Location Archived Maps by Region Canada Air Quality Air Quality on Google Earth Links A-Z About AirNow AirNow International Air Quality Action Days / Alerts AirCompare Air Quality Index (AQI) ...
Ahmetovic, Dragan; Manduchi, Roberto; Coughlan, James M.; Mascetti, Sergio
2016-01-01
In this paper we propose a computer vision-based technique that mines existing spatial image databases for discovery of zebra crosswalks in urban settings. Knowing the location of crosswalks is critical for a blind person planning a trip that includes street crossing. By augmenting existing spatial databases (such as Google Maps or OpenStreetMap) with this information, a blind traveler may make more informed routing decisions, resulting in greater safety during independent travel. Our algorithm first searches for zebra crosswalks in satellite images; all candidates thus found are validated against spatially registered Google Street View images. This cascaded approach enables fast and reliable discovery and localization of zebra crosswalks in large image datasets. While fully automatic, our algorithm could also be complemented by a final crowdsourcing validation stage for increased accuracy. PMID:26824080
Chien, Tsair-Wei; Chang, Yu; Wang, Hsien-Yi
2018-02-01
Many researchers used National Health Insurance database to publish medical papers which are often retrospective, population-based, and cohort studies. However, the author's research domain and academic characteristics are still unclear.By searching the PubMed database (Pubmed.com), we used the keyword of [Taiwan] and [National Health Insurance Research Database], then downloaded 2913 articles published from 1995 to 2017. Social network analysis (SNA), Gini coefficient, and Google Maps were applied to gather these data for visualizing: the most productive author; the pattern of coauthor collaboration teams; and the author's research domain denoted by abstract keywords and Pubmed MESH (medical subject heading) terms.Utilizing the 2913 papers from Taiwan's National Health Insurance database, we chose the top 10 research teams shown on Google Maps and analyzed one author (Dr. Kao) who published 149 papers in the database in 2015. In the past 15 years, we found Dr. Kao had 2987 connections with other coauthors from 13 research teams. The cooccurrence abstract keywords with the highest frequency are cohort study and National Health Insurance Research Database. The most coexistent MESH terms are tomography, X-ray computed, and positron-emission tomography. The strength of the author research distinct domain is very low (Gini < 0.40).SNA incorporated with Google Maps and Gini coefficient provides insight into the relationships between entities. The results obtained in this study can be applied for a comprehensive understanding of other productive authors in the field of academics.
GeoMapApp, Virtual Ocean, and other Free Data Resources for the 21st Century Classroom
NASA Astrophysics Data System (ADS)
Goodwillie, A. M.; Ryan, W.; Carbotte, S.; Melkonian, A.; Coplan, J.; Arko, R.; Ferrini, V.; O'Hara, S.; Leung, A.; Bonckzowski, J.
2008-12-01
With funding from the U.S. National Science Foundation, the Marine Geoscience Data System (MGDS) (http://www.marine-geo.org/) is developing GeoMapApp (http://www.geomapapp.org) - a computer application that provides wide-ranging map-based visualization and manipulation options for interdisciplinary geosciences research and education. The novelty comes from the use of this visual tool to discover and explore data, with seamless links to further discovery using traditional text-based approaches. Users can generate custom maps and grids and import their own data sets. Built-in functionality allows users to readily explore a broad suite of interactive data sets and interfaces. Examples include multi-resolution global digital models of topography, gravity, sediment thickness, and crustal ages; rock, fluid, biology and sediment sample information; research cruise underway geophysical and multibeam data; earthquake events; submersible dive photos of hydrothermal vents; geochemical analyses; DSDP/ODP core logs; seismic reflection profiles; contouring, shading, profiling of grids; and many more. On-line audio-visual tutorials lead users step-by-step through GeoMapApp functionality (http://www.geomapapp.org/tutorials/). Virtual Ocean (http://www.virtualocean.org/) integrates GeoMapApp with a 3-D earth browser based upon NASA WorldWind, providing yet more powerful capabilities. The searchable MGDS Media Bank (http://media.marine-geo.org/) supports viewing of remarkable images and video from the NSF Ridge 2000 and MARGINS programs. For users familiar with Google Earth (tm), KML files are available for viewing several MGDS data sets (http://www.marine-geo.org/education/kmls.php). Examples of accessing and manipulating a range of geoscience data sets from various NSF-funded programs will be shown. GeoMapApp, Virtual Ocean, the MGDS Media Bank and KML files are free MGDS data resources and work on any type of computer. They are currently used by educators, researchers, school teachers and the general public.
ERIC Educational Resources Information Center
McMahon, Don; Cihak, David F.; Wright, Rachel
2015-01-01
The purpose of this study was to examine the effects of location-based augmented reality navigation compared to Google Maps and paper maps as navigation aids for students with disabilities. The participants in this single subject study were three college students with intellectual disability and one college student with autism spectrum disorder.…
In the current study, three Google Street View cars were equipped with the Aclima Environmental Intelligence ™ Platform. The air pollutants of interest, including O3, NO, NO2, CO2, black carbon, and particle number in several size ranges, were measured using a suite of fast...
Usability Evaluation of Public Web Mapping Sites
NASA Astrophysics Data System (ADS)
Wang, C.
2014-04-01
Web mapping sites are interactive maps that are accessed via Webpages. With the rapid development of Internet and Geographic Information System (GIS) field, public web mapping sites are not foreign to people. Nowadays, people use these web mapping sites for various reasons, in that increasing maps and related map services of web mapping sites are freely available for end users. Thus, increased users of web mapping sites led to more usability studies. Usability Engineering (UE), for instance, is an approach for analyzing and improving the usability of websites through examining and evaluating an interface. In this research, UE method was employed to explore usability problems of four public web mapping sites, analyze the problems quantitatively and provide guidelines for future design based on the test results. Firstly, the development progress for usability studies were described, and simultaneously several usability evaluation methods such as Usability Engineering (UE), User-Centered Design (UCD) and Human-Computer Interaction (HCI) were generally introduced. Then the method and procedure of experiments for the usability test were presented in detail. In this usability evaluation experiment, four public web mapping sites (Google Maps, Bing maps, Mapquest, Yahoo Maps) were chosen as the testing websites. And 42 people, who having different GIS skills (test users or experts), gender (male or female), age and nationality, participated in this test to complete the several test tasks in different teams. The test comprised three parts: a pretest background information questionnaire, several test tasks for quantitative statistics and progress analysis, and a posttest questionnaire. The pretest and posttest questionnaires focused on gaining the verbal explanation of their actions qualitatively. And the design for test tasks targeted at gathering quantitative data for the errors and problems of the websites. Then, the results mainly from the test part were analyzed. The success rate from different public web mapping sites was calculated and compared, and displayed by the means of diagram. And the answers from questionnaires were also classified and organized in this part. Moreover, based on the analysis, this paper expands the discussion about the layout, map visualization, map tools, search logic and etc. Finally, this paper closed with some valuable guidelines and suggestions for the design of public web mapping sites. Also, limitations for this research stated in the end.
ERIC Educational Resources Information Center
McMillin, Bill; Gibson, Sally; MacDonald, Jean
2016-01-01
Animated maps of the library stacks were integrated into the catalog interface at Pratt Institute and into the EBSCO Discovery Service interface at Illinois State University. The mapping feature was developed for optimal automation of the update process to enable a range of library personnel to update maps and call-number ranges. The development…
Information Portals: The Next Generation Catalog
ERIC Educational Resources Information Center
Allison, DeeAnn
2010-01-01
Libraries today face an increasing challenge: to provide relevant information to diverse populations with differing needs while competing with Web search engines like Google. In 2009, a large group of libraries, including the University of Nebraska-Lincoln Libraries, joined with Innovative Interfaces as development partners to design a new type of…
Google Glass-Directed Monitoring and Control of Microfluidic Biosensors and Actuators
Zhang, Yu Shrike; Busignani, Fabio; Ribas, João; Aleman, Julio; Rodrigues, Talles Nascimento; Shaegh, Seyed Ali Mousavi; Massa, Solange; Rossi, Camilla Baj; Taurino, Irene; Shin, Su-Ryon; Calzone, Giovanni; Amaratunga, Givan Mark; Chambers, Douglas Leon; Jabari, Saman; Niu, Yuxi; Manoharan, Vijayan; Dokmeci, Mehmet Remzi; Carrara, Sandro; Demarchi, Danilo; Khademhosseini, Ali
2016-01-01
Google Glass is a recently designed wearable device capable of displaying information in a smartphone-like hands-free format by wireless communication. The Glass also provides convenient control over remote devices, primarily enabled by voice recognition commands. These unique features of the Google Glass make it useful for medical and biomedical applications where hands-free experiences are strongly preferred. Here, we report for the first time, an integral set of hardware, firmware, software, and Glassware that enabled wireless transmission of sensor data onto the Google Glass for on-demand data visualization and real-time analysis. Additionally, the platform allowed the user to control outputs entered through the Glass, therefore achieving bi-directional Glass-device interfacing. Using this versatile platform, we demonstrated its capability in monitoring physical and physiological parameters such as temperature, pH, and morphology of liver- and heart-on-chips. Furthermore, we showed the capability to remotely introduce pharmaceutical compounds into a microfluidic human primary liver bioreactor at desired time points while monitoring their effects through the Glass. We believe that such an innovative platform, along with its concept, has set up a premise in wearable monitoring and controlling technology for a wide variety of applications in biomedicine. PMID:26928456
Google Glass-Directed Monitoring and Control of Microfluidic Biosensors and Actuators
NASA Astrophysics Data System (ADS)
Zhang, Yu Shrike; Busignani, Fabio; Ribas, João; Aleman, Julio; Rodrigues, Talles Nascimento; Shaegh, Seyed Ali Mousavi; Massa, Solange; Rossi, Camilla Baj; Taurino, Irene; Shin, Su-Ryon; Calzone, Giovanni; Amaratunga, Givan Mark; Chambers, Douglas Leon; Jabari, Saman; Niu, Yuxi; Manoharan, Vijayan; Dokmeci, Mehmet Remzi; Carrara, Sandro; Demarchi, Danilo; Khademhosseini, Ali
2016-03-01
Google Glass is a recently designed wearable device capable of displaying information in a smartphone-like hands-free format by wireless communication. The Glass also provides convenient control over remote devices, primarily enabled by voice recognition commands. These unique features of the Google Glass make it useful for medical and biomedical applications where hands-free experiences are strongly preferred. Here, we report for the first time, an integral set of hardware, firmware, software, and Glassware that enabled wireless transmission of sensor data onto the Google Glass for on-demand data visualization and real-time analysis. Additionally, the platform allowed the user to control outputs entered through the Glass, therefore achieving bi-directional Glass-device interfacing. Using this versatile platform, we demonstrated its capability in monitoring physical and physiological parameters such as temperature, pH, and morphology of liver- and heart-on-chips. Furthermore, we showed the capability to remotely introduce pharmaceutical compounds into a microfluidic human primary liver bioreactor at desired time points while monitoring their effects through the Glass. We believe that such an innovative platform, along with its concept, has set up a premise in wearable monitoring and controlling technology for a wide variety of applications in biomedicine.
Google Glass-Directed Monitoring and Control of Microfluidic Biosensors and Actuators.
Zhang, Yu Shrike; Busignani, Fabio; Ribas, João; Aleman, Julio; Rodrigues, Talles Nascimento; Shaegh, Seyed Ali Mousavi; Massa, Solange; Baj Rossi, Camilla; Taurino, Irene; Shin, Su-Ryon; Calzone, Giovanni; Amaratunga, Givan Mark; Chambers, Douglas Leon; Jabari, Saman; Niu, Yuxi; Manoharan, Vijayan; Dokmeci, Mehmet Remzi; Carrara, Sandro; Demarchi, Danilo; Khademhosseini, Ali
2016-03-01
Google Glass is a recently designed wearable device capable of displaying information in a smartphone-like hands-free format by wireless communication. The Glass also provides convenient control over remote devices, primarily enabled by voice recognition commands. These unique features of the Google Glass make it useful for medical and biomedical applications where hands-free experiences are strongly preferred. Here, we report for the first time, an integral set of hardware, firmware, software, and Glassware that enabled wireless transmission of sensor data onto the Google Glass for on-demand data visualization and real-time analysis. Additionally, the platform allowed the user to control outputs entered through the Glass, therefore achieving bi-directional Glass-device interfacing. Using this versatile platform, we demonstrated its capability in monitoring physical and physiological parameters such as temperature, pH, and morphology of liver- and heart-on-chips. Furthermore, we showed the capability to remotely introduce pharmaceutical compounds into a microfluidic human primary liver bioreactor at desired time points while monitoring their effects through the Glass. We believe that such an innovative platform, along with its concept, has set up a premise in wearable monitoring and controlling technology for a wide variety of applications in biomedicine.
Leveraging Google Geo Tools for Interactive STEM Education: Insights from the GEODE Project
NASA Astrophysics Data System (ADS)
Dordevic, M.; Whitmeyer, S. J.; De Paor, D. G.; Karabinos, P.; Burgin, S.; Coba, F.; Bentley, C.; St John, K. K.
2016-12-01
Web-based imagery and geospatial tools have transformed our ability to immerse students in global virtual environments. Google's suite of geospatial tools, such as Google Earth (± Engine), Google Maps, and Street View, allow developers and instructors to create interactive and immersive environments, where students can investigate and resolve common misconceptions in STEM concepts and natural processes. The GEODE (.net) project is developing digital resources to enhance STEM education. These include virtual field experiences (VFEs), such as an interactive visualization of the breakup of the Pangaea supercontinent, a "Grand Tour of the Terrestrial Planets," and GigaPan-based VFEs of sites like the Canadian Rockies. Web-based challenges, such as EarthQuiz (.net) and the "Fold Analysis Challenge," incorporate scaffolded investigations of geoscience concepts. EarthQuiz features web-hosted imagery, such as Street View, Photo Spheres, GigaPans, and Satellite View, as the basis for guided inquiry. In the Fold Analysis Challenge, upper-level undergraduates use Google Earth to evaluate a doubly-plunging fold at Sheep Mountain, WY. GEODE.net also features: "Reasons for the Seasons"—a Google Earth-based visualization that addresses misconceptions that abound amongst students, teachers, and the public, many of whom believe that seasonality is caused by large variations in Earth's distance from the Sun; "Plate Euler Pole Finder," which helps students understand rotational motion of tectonic plates on the globe; and "Exploring Marine Sediments Using Google Earth," an exercise that uses empirical data to explore the surficial distribution of marine sediments in the modern ocean. The GEODE research team includes the authors and: Heather Almquist, Cinzia Cervato, Gene Cooper, Helen Crompton, Terry Pavlis, Jen Piatek, Bill Richards, Jeff Ryan, Ron Schott, Barb Tewksbury, and their students and collaborating colleagues. We are supported by NSF DUE 1323419 and a Google Geo Curriculum Award.
Wilber 3: A Python-Django Web Application For Acquiring Large-scale Event-oriented Seismic Data
NASA Astrophysics Data System (ADS)
Newman, R. L.; Clark, A.; Trabant, C. M.; Karstens, R.; Hutko, A. R.; Casey, R. E.; Ahern, T. K.
2013-12-01
Since 2001, the IRIS Data Management Center (DMC) WILBER II system has provided a convenient web-based interface for locating seismic data related to a particular event, and requesting a subset of that data for download. Since its launch, both the scale of available data and the technology of web-based applications have developed significantly. Wilber 3 is a ground-up redesign that leverages a number of public and open-source projects to provide an event-oriented data request interface with a high level of interactivity and scalability for multiple data types. Wilber 3 uses the IRIS/Federation of Digital Seismic Networks (FDSN) web services for event data, metadata, and time-series data. Combining a carefully optimized Google Map with the highly scalable SlickGrid data API, the Wilber 3 client-side interface can load tens of thousands of events or networks/stations in a single request, and provide instantly responsive browsing, sorting, and filtering of event and meta data in the web browser, without further reliance on the data service. The server-side of Wilber 3 is a Python-Django application, one of over a dozen developed in the last year at IRIS, whose common framework, components, and administrative overhead represent a massive savings in developer resources. Requests for assembled datasets, which may include thousands of data channels and gigabytes of data, are queued and executed using the Celery distributed Python task scheduler, giving Wilber 3 the ability to operate in parallel across a large number of nodes.
Mercury Toolset for Spatiotemporal Metadata
NASA Technical Reports Server (NTRS)
Wilson, Bruce E.; Palanisamy, Giri; Devarakonda, Ranjeet; Rhyne, B. Timothy; Lindsley, Chris; Green, James
2010-01-01
Mercury (http://mercury.ornl.gov) is a set of tools for federated harvesting, searching, and retrieving metadata, particularly spatiotemporal metadata. Version 3.0 of the Mercury toolset provides orders of magnitude improvements in search speed, support for additional metadata formats, integration with Google Maps for spatial queries, facetted type search, support for RSS (Really Simple Syndication) delivery of search results, and enhanced customization to meet the needs of the multiple projects that use Mercury. It provides a single portal to very quickly search for data and information contained in disparate data management systems, each of which may use different metadata formats. Mercury harvests metadata and key data from contributing project servers distributed around the world and builds a centralized index. The search interfaces then allow the users to perform a variety of fielded, spatial, and temporal searches across these metadata sources. This centralized repository of metadata with distributed data sources provides extremely fast search results to the user, while allowing data providers to advertise the availability of their data and maintain complete control and ownership of that data. Mercury periodically (typically daily) harvests metadata sources through a collection of interfaces and re-indexes these metadata to provide extremely rapid search capabilities, even over collections with tens of millions of metadata records. A number of both graphical and application interfaces have been constructed within Mercury, to enable both human users and other computer programs to perform queries. Mercury was also designed to support multiple different projects, so that the particular fields that can be queried and used with search filters are easy to configure for each different project.
Mercury Toolset for Spatiotemporal Metadata
NASA Astrophysics Data System (ADS)
Devarakonda, Ranjeet; Palanisamy, Giri; Green, James; Wilson, Bruce; Rhyne, B. Timothy; Lindsley, Chris
2010-06-01
Mercury (http://mercury.ornl.gov) is a set of tools for federated harvesting, searching, and retrieving metadata, particularly spatiotemporal metadata. Version 3.0 of the Mercury toolset provides orders of magnitude improvements in search speed, support for additional metadata formats, integration with Google Maps for spatial queries, facetted type search, support for RSS (Really Simple Syndication) delivery of search results, and enhanced customization to meet the needs of the multiple projects that use Mercury. It provides a single portal to very quickly search for data and information contained in disparate data management systems, each of which may use different metadata formats. Mercury harvests metadata and key data from contributing project servers distributed around the world and builds a centralized index. The search interfaces then allow the users to perform a variety of fielded, spatial, and temporal searches across these metadata sources. This centralized repository of metadata with distributed data sources provides extremely fast search results to the user, while allowing data providers to advertise the availability of their data and maintain complete control and ownership of that data. Mercury periodically (typically daily)harvests metadata sources through a collection of interfaces and re-indexes these metadata to provide extremely rapid search capabilities, even over collections with tens of millions of metadata records. A number of both graphical and application interfaces have been constructed within Mercury, to enable both human users and other computer programs to perform queries. Mercury was also designed to support multiple different projects, so that the particular fields that can be queried and used with search filters are easy to configure for each different project.
Mapping for the masses: using free remote sensing data for disaster management
NASA Astrophysics Data System (ADS)
Teeuw, R.; McWilliam, N.; Morris, N.; Saunders, C.
2009-04-01
We examine the uses of free satellite imagery and Digital Elevation Models (DEMs) for disaster management, targeting three data sources: the United Nations Charter on Space and Disasters, Google Earth and internet-based satellite data archives, such as the Global Land Cover Facility (GLCF). The research has assessed SRTM and ASTER DEM data, Landsat TM/ETM+ and ASTER imagery, as well as utilising datasets and basic GIS operations available via Google Earth. As an aid to Disaster Risk Reduction, four sets of maps can be produced from satellite data: (i) Multiple Geohazards: areas prone to slope instability, coastal inundation and fluvial flooding; (ii) Vulnerability: population density, habitation types, land cover types and infrastructure; (iii) Disaster Risk: produced by combining severity scores from (i) and (ii); (iv) Reconstruction: zones of rock/sediment with construction uses; areas of woodland (for fuel/construction) water sources; transport routes; zones suitable for re-settlement. This set of Disaster Risk Reduction maps are ideal for regional (1:50,000 to 1:250,000 scale) planning for in low-income countries: more detailed assessments require relatively expensive high resolution satellite imagery or aerial photography, although Google Earth has a good track record for posting high-res imagery of disaster zones (e.g. the 2008 Burma storm surge). The Disaster Risk maps highlight areas of maximum risk to a region's emergency planners and decision makers, enabling various types of public education and other disaster mitigation measures. The Reconstruction map also helps to save lives, by facilitating disaster recovery. Many problems have been identified. Access to the UN Charter imagery is fine after a disaster, but very difficult if assessing pre-disaster indicators: the data supplied also tends to be pre-processed, when some relief agencies would prefer to have raw data. The limited and expensive internet access in many developing countries limits access to archives of free satellite data, such as the GLCF. Finally, data integration, spatial/temporal analysis and map production are all hindered by the high price of most GIS software, making the development of suitable open-source software a priority.
Google Maps for Crowdsourced Emergency Routing
NASA Astrophysics Data System (ADS)
Nedkov, S.; Zlatanova, S.
2012-08-01
Gathering infrastructure data in emergency situations is challenging. The affected by a disaster areas are often large and the needed observations numerous. Spaceborne remote sensing techniques cover large areas but they are of limited use as their field of view may be blocked by clouds, smoke, buildings, highways, etc. Remote sensing products furthermore require specialists to collect and analyze the data. This contrasts the nature of the damage detection problem: almost everyone is capable of observing whether a street is usable or not. The crowd is fit for solving these challenges as its members are numerous, they are willing to help and are often in the vicinity of the disaster thereby forming a highly dispersed sensor network. This paper proposes and implements a small WebGIS application for performing shortest path calculations based on crowdsourced information about the infrastructure health. The application is built on top of Google Maps and uses its routing service to calculate the shortest distance between two locations. Impassable areas are indicated on a map by people performing in-situ observations on a mobile device, and by users on a desktop machine who consult a multitude of information sources.
In-field Access to Geoscientific Metadata through GPS-enabled Mobile Phones
NASA Astrophysics Data System (ADS)
Hobona, Gobe; Jackson, Mike; Jordan, Colm; Butchart, Ben
2010-05-01
Fieldwork is an integral part of much geosciences research. But whilst geoscientists have physical or online access to data collections whilst in the laboratory or at base stations, equivalent in-field access is not standard or straightforward. The increasing availability of mobile internet and GPS-supported mobile phones, however, now provides the basis for addressing this issue. The SPACER project was commissioned by the Rapid Innovation initiative of the UK Joint Information Systems Committee (JISC) to explore the potential for GPS-enabled mobile phones to access geoscientific metadata collections. Metadata collections within the geosciences and the wider geospatial domain can be disseminated through web services based on the Catalogue Service for Web(CSW) standard of the Open Geospatial Consortium (OGC) - a global grouping of over 380 private, public and academic organisations aiming to improve interoperability between geospatial technologies. CSW offers an XML-over-HTTP interface for querying and retrieval of geospatial metadata. By default, the metadata returned by CSW is based on the ISO19115 standard and encoded in XML conformant to ISO19139. The SPACER project has created a prototype application that enables mobile phones to send queries to CSW containing user-defined keywords and coordinates acquired from GPS devices built-into the phones. The prototype has been developed using the free and open source Google Android platform. The mobile application offers views for listing titles, presenting multiple metadata elements and a Google Map with an overlay of bounding coordinates of datasets. The presentation will describe the architecture and approach applied in the development of the prototype.
Do, Nhan V; Barnhill, Rick; Heermann-Do, Kimberly A; Salzman, Keith L; Gimbel, Ronald W
2011-01-01
To design, build, implement, and evaluate a personal health record (PHR), tethered to the Military Health System, that leverages Microsoft® HealthVault and Google® Health infrastructure based on user preference. A pilot project was conducted in 2008-2009 at Madigan Army Medical Center in Tacoma, Washington. Our PHR was architected to a flexible platform that incorporated standards-based models of Continuity of Document and Continuity of Care Record to map Department of Defense-sourced health data, via a secure Veterans Administration data broker, to Microsoft® HealthVault and Google® Health based on user preference. The project design and implementation were guided by provider and patient advisory panels with formal user evaluation. The pilot project included 250 beneficiary users. Approximately 73.2% of users were < 65 years of age, and 38.4% were female. Of the users, 169 (67.6%) selected Microsoft® HealthVault, and 81 (32.4%) selected Google® Health as their PHR of preference. Sample evaluation of users reflected 100% (n = 60) satisfied with convenience of record access and 91.7% (n = 55) satisfied with overall functionality of PHR. Key lessons learned related to data-transfer decisions (push vs pull), purposeful delays in reporting sensitive information, understanding and mapping PHR use and clinical workflow, and decisions on information patients may choose to share with their provider. Currently PHRs are being viewed as empowering tools for patient activation. Design and implementation issues (eg, technical, organizational, information security) are substantial and must be thoughtfully approached. Adopting standards into design can enhance the national goal of portability and interoperability.
Use of Openly Available Satellite Images for Remote Sensing Education
NASA Astrophysics Data System (ADS)
Wang, C.-K.
2011-09-01
With the advent of Google Earth, Google Maps, and Microsoft Bing Maps, high resolution satellite imagery are becoming more easily accessible than ever. It have been the case that the college students may already have wealth experiences with the high resolution satellite imagery by using these software and web services prior to any formal remote sensing education. It is obvious that the remote sensing education should be adjusted to the fact that the audience are already the customers of remote sensing products (through the use of the above mentioned services). This paper reports the use of openly available satellite imagery in an introductory-level remote sensing course in the Department of Geomatics of National Cheng Kung University as a term project. From the experience learned from the fall of 2009 and 2010, it shows that this term project has effectively aroused the students' enthusiastic toward Remote Sensing.
In campus location finder using mobile application services
NASA Astrophysics Data System (ADS)
Fai, Low Weng; Audah, Lukman
2017-09-01
Navigation services become very common in this era, the application include Google Map, Waze and etc. Although navigation application contains the main routing service in open area but not all of the buildings are recorded in the database. In this project, an application is made for the indoor and outdoor navigation in Universiti Tun Hussein Onn Malaysia (UTHM). It is used to help outsider and new incoming students by navigating them from their current location to destination using mobile application name "U Finder". Thunkable website has been used to build the application for outdoor and indoor navigation. Outdoor navigation is linked to the Google Map and indoor navigation is using the QR code for positioning and routing picture for navigation. The outdoor navigation can route user to the main faculties in UTHM and indoor navigation is only done for the G1 building in UTHM.
NASA Astrophysics Data System (ADS)
Schmaltz, J. E.; Ilavajhala, S.; Plesea, L.; Hall, J. R.; Boller, R. A.; Chang, G.; Sadaqathullah, S.; Kim, R.; Murphy, K. J.; Thompson, C. K.
2012-12-01
Expedited processing of imagery from NASA satellites for near-real time use by non-science applications users has a long history, especially since the beginning of the Terra and Aqua missions. Several years ago, the Land Atmosphere Near-real-time Capability for EOS (LANCE) was created to greatly expand the range of near-real time data products from a variety of Earth Observing System (EOS) instruments. NASA's Earth Observing System Data and Information System (EOSDIS) began exploring methods to distribute these data as imagery in an intuitive, geo-referenced format, which would be available within three hours of acquisition. Toward this end, EOSDIS has developed the Global Imagery Browse Services (GIBS, http://earthdata.nasa.gov/gibs) to provide highly responsive, scalable, and expandable imagery services. The baseline technology chosen for GIBS was a Tiled Web Mapping Service (TWMS) developed at the Jet Propulsion Laboratory. Using this, global images and mosaics are divided into tiles with fixed bounding boxes for a pyramid of fixed resolutions. Initially, the satellite imagery is created at the existing data systems for each sensor, ensuring the oversight of those most knowledgeable about the science. There, the satellite data is geolocated and converted to an image format such as JPEG, TIFF, or PNG. The GIBS ingest server retrieves imagery from the various data systems and converts them into image tiles, which are stored in a highly-optimized raster format named Meta Raster Format (MRF). The image tiles are then served to users via HTTP by means of an Apache module. Services are available for the entire globe (lat-long projection) and for both polar regions (polar stereographic projection). Requests to the services can be made with the non-standard, but widely known, TWMS format or via the well-known OGC Web Map Tile Service (WMTS) standard format. Standard OGC Web Map Service (WMS) access to the GIBS server is also available. In addition, users may request a KML pyramid. This variety of access methods allows stakeholders to develop visualization/browse clients for a diverse variety of specific audiences. Currently, EOSDIS is providing an OpenLayers web client, Worldview (http://earthdata.nasa.gov/worldview), as an interface to GIBS. A variety of other existing clients can also be developed using such tools as Google Earth, Google Earth browser Plugin, ESRI's Adobe Flash/Flex Client Library, NASA World Wind, Perceptive Pixel Client, Esri's iOS Client Library, and OpenLayers for Mobile. The imagery browse capabilities from GIBS can be combined with other EOSDIS services (i.e. ECHO OpenSearch) via a client that ties them both together to provide an interface that enables data download from the onscreen imagery. Future plans for GIBS include providing imagery based on science quality data from the entire data record of these EOS instruments.
Participating in the Geospatial Web: Collaborative Mapping, Social Networks and Participatory GIS
NASA Astrophysics Data System (ADS)
Rouse, L. Jesse; Bergeron, Susan J.; Harris, Trevor M.
In 2005, Google, Microsoft and Yahoo! released free Web mapping applications that opened up digital mapping to mainstream Internet users. Importantly, these companies also released free APIs for their platforms, allowing users to geo-locate and map their own data. These initiatives have spurred the growth of the Geospatial Web and represent spatially aware online communities and new ways of enabling communities to share information from the bottom up. This chapter explores how the emerging Geospatial Web can meet some of the fundamental needs of Participatory GIS projects to incorporate local knowledge into GIS, as well as promote public access and collaborative mapping.
NASA Technical Reports Server (NTRS)
2010-01-01
Topics covered include: Situational Awareness from a Low-Cost Camera System; Data Acquisition System for Multi-Frequency Radar Flight Operations Preparation; Mercury Toolset for Spatiotemporal Metadata; Social Tagging of Mission Data; Integrating Radar Image Data with Google Maps; Demonstration of a Submillimeter-Wave HEMT Oscillator Module at 330 GHz; Flexible Peripheral Component Interconnect Input/Output Card; Interface Supports Lightweight Subsystem Routing for Flight Applications; MMIC Amplifiers and Wafer Probes for 350 to 500 GHz; Public Risk Assessment Program; Particle Swarm Optimization Toolbox; Telescience Support Center Data System Software; Update on PISCES; Ground and Space Radar Volume Matching and Comparison Software; Web-Based Interface for Command and Control of Network Sensors; Orbit Determination Toolbox; Distributed Observer Network; Computer-Automated Evolution of Spacecraft X-Band Antennas; Practical Loop-Shaping Design of Feedback Control Systems; Fully Printed High-Frequency Phased-Array Antenna on Flexible Substrate; Formula for the Removal and Remediation of Polychlorinated Biphenyls in Painted Structures; Integrated Solar Concentrator and Shielded Radiator; Water Membrane Evaporator; Modeling of Failure for Analysis of Triaxial Braided Carbon Fiber Composites; Catalyst for Carbon Monoxide Oxidation; Titanium Hydroxide - a Volatile Species at High Temperature; Selective Functionalization of Carbon Nanotubes: Part II; Steerable Hopping Six-Legged Robot; Launchable and Retrievable Tetherobot; Hybrid Heat Exchangers; Orbital Winch for High-Strength, Space-Survivable Tethers; Parameterized Linear Longitudinal Airship Model; and Physics of Life: A Model for Non-Newtonian Properties of Living Systems.
GLIMPSE: Google Glass interface for sensory feedback in myoelectric hand prostheses.
Markovic, Marko; Karnal, Hemanth; Graimann, Bernhard; Farina, Dario; Dosen, Strahinja
2017-06-01
Providing sensory feedback to the user of the prosthesis is an important challenge. The common approach is to use tactile stimulation, which is easy to implement but requires training and has limited information bandwidth. In this study, we propose an alternative approach based on augmented reality. We have developed the GLIMPSE, a Google Glass application which connects to the prosthesis via a Bluetooth interface and renders the prosthesis states (EMG signals, aperture, force and contact) using augmented reality (see-through display) and sound (bone conduction transducer). The interface was tested in healthy subjects that used the prosthesis with (FB group) and without (NFB group) feedback during a modified clothespins test that allowed us to vary the difficulty of the task. The outcome measures were the number of unsuccessful trials, the time to accomplish the task, and the subjective ratings of the relevance of the feedback. There was no difference in performance between FB and NFB groups in the case of a simple task (basic, same-color clothespins test), but the feedback significantly improved the performance in a more complex task (pins of different resistances). Importantly, the GLIMPSE feedback did not increase the time to accomplish the task. Therefore, the supplemental feedback might be useful in the tasks which are more demanding, and thereby less likely to benefit from learning and feedforward control. The subjects integrated the supplemental feedback with the intrinsic sources (vision and muscle proprioception), developing their own idiosyncratic strategies to accomplish the task. The present study demonstrates a novel self-contained, ready-to-deploy, wearable feedback interface. The interface was successfully tested and was proven to be feasible and functionally beneficial. The GLIMPSE can be used as a practical solution but also as a general and flexible instrument to investigate closed-loop prosthesis control.
Harnessing Satellite Imageries in Feature Extraction Using Google Earth Pro
NASA Astrophysics Data System (ADS)
Fernandez, Sim Joseph; Milano, Alan
2016-07-01
Climate change has been a long-time concern worldwide. Impending flooding, for one, is among its unwanted consequences. The Phil-LiDAR 1 project of the Department of Science and Technology (DOST), Republic of the Philippines, has developed an early warning system in regards to flood hazards. The project utilizes the use of remote sensing technologies in determining the lives in probable dire danger by mapping and attributing building features using LiDAR dataset and satellite imageries. A free mapping software named Google Earth Pro (GEP) is used to load these satellite imageries as base maps. Geotagging of building features has been done so far with the use of handheld Global Positioning System (GPS). Alternatively, mapping and attribution of building features using GEP saves a substantial amount of resources such as manpower, time and budget. Accuracy-wise, geotagging by GEP is dependent on either the satellite imageries or orthophotograph images of half-meter resolution obtained during LiDAR acquisition and not on the GPS of three-meter accuracy. The attributed building features are overlain to the flood hazard map of Phil-LiDAR 1 in order to determine the exposed population. The building features as obtained from satellite imageries may not only be used in flood exposure assessment but may also be used in assessing other hazards and a number of other uses. Several other features may also be extracted from the satellite imageries.
NASA Astrophysics Data System (ADS)
Carraro, Francesco
"Mars @ ASDC" is a project born with the goal of using the new web technologies to assist researches involved in the study of Mars. This project employs Mars map and javascript APIs provided by Google to visualize data acquired by space missions on the planet. So far, visualization of tracks acquired by MARSIS and regions observed by VIRTIS-Rosetta has been implemented. The main reason for the creation of this kind of tool is the difficulty in handling hundreds or thousands of acquisitions, like the ones from MARSIS, and the consequent difficulty in finding observations related to a particular region. This led to the development of a tool which allows to search for acquisitions either by defining the region of interest through a set of geometrical parameters or by manually selecting the region on the map through a few mouse clicks The system allows the visualization of tracks (acquired by MARSIS) or regions (acquired by VIRTIS-Rosetta) which intersect the user defined region. MARSIS tracks can be visualized both in Mercator and polar projections while the regions observed by VIRTIS can presently be visualized only in Mercator projection. The Mercator projection is the standard map provided by Google. The polar projections are provided by NASA and have been developed to be used in combination with APIs provided by Google The whole project has been developed following the "open source" philosophy: the client-side code which handles the functioning of the web page is written in javascript; the server-side code which executes the searches for tracks or regions is written in PHP and the DB which undergoes the system is MySQL.
Tags Help Make Libraries Del.icio.us: Social Bookmarking and Tagging Boost Participation
ERIC Educational Resources Information Center
Rethlefsen, Melissa L.
2007-01-01
Traditional library web products, whether online public access catalogs, library databases, or even library web sites, have long been rigidly controlled and difficult to use. Patrons regularly prefer Google's simple interface. Now social bookmarking and tagging tools help librarians bridge the gap between the library's need to offer authoritative,…
The jmzQuantML programming interface and validator for the mzQuantML data standard.
Qi, Da; Krishna, Ritesh; Jones, Andrew R
2014-03-01
The mzQuantML standard from the HUPO Proteomics Standards Initiative has recently been released, capturing quantitative data about peptides and proteins, following analysis of MS data. We present a Java application programming interface (API) for mzQuantML called jmzQuantML. The API provides robust bridges between Java classes and elements in mzQuantML files and allows random access to any part of the file. The API provides read and write capabilities, and is designed to be embedded in other software packages, enabling mzQuantML support to be added to proteomics software tools (http://code.google.com/p/jmzquantml/). The mzQuantML standard is designed around a multilevel validation system to ensure that files are structurally and semantically correct for different proteomics quantitative techniques. In this article, we also describe a Java software tool (http://code.google.com/p/mzquantml-validator/) for validating mzQuantML files, which is a formal part of the data standard. © 2014 The Authors. Proteomics published by Wiley-VCH Verlag GmbH & Co. KGaA, Weinheim.
Fleischman, Ross J.; Lundquist, Mark; Jui, Jonathan; Newgard, Craig D.; Warden, Craig
2014-01-01
Objective To derive and validate a model that accurately predicts ambulance arrival time that could be implemented as a Google Maps web application. Methods This was a retrospective study of all scene transports in Multnomah County, Oregon, from January 1 through December 31, 2008. Scene and destination hospital addresses were converted to coordinates. ArcGIS Network Analyst was used to estimate transport times based on street network speed limits. We then created a linear regression model to improve the accuracy of these street network estimates using weather, patient characteristics, use of lights and sirens, daylight, and rush-hour intervals. The model was derived from a 50% sample and validated on the remainder. Significance of the covariates was determined by p < 0.05 for a t-test of the model coefficients. Accuracy was quantified by the proportion of estimates that were within 5 minutes of the actual transport times recorded by computer-aided dispatch. We then built a Google Maps-based web application to demonstrate application in real-world EMS operations. Results There were 48,308 included transports. Street network estimates of transport time were accurate within 5 minutes of actual transport time less than 16% of the time. Actual transport times were longer during daylight and rush-hour intervals and shorter with use of lights and sirens. Age under 18 years, gender, wet weather, and trauma system entry were not significant predictors of transport time. Our model predicted arrival time within 5 minutes 73% of the time. For lights and sirens transports, accuracy was within 5 minutes 77% of the time. Accuracy was identical in the validation dataset. Lights and sirens saved an average of 3.1 minutes for transports under 8.8 minutes, and 5.3 minutes for longer transports. Conclusions An estimate of transport time based only on a street network significantly underestimated transport times. A simple model incorporating few variables can predict ambulance time of arrival to the emergency department with good accuracy. This model could be linked to global positioning system data and an automated Google Maps web application to optimize emergency department resource use. Use of lights and sirens had a significant effect on transport times. PMID:23865736
NASA Astrophysics Data System (ADS)
Zhang, X.; Wu, B.; Zhang, M.; Zeng, H.
2017-12-01
Rice is one of the main staple foods in East Asia and Southeast Asia, which has occupied more than half of the world's population with 11% of cultivated land. Study on rice can provide direct or indirect information on food security and water source management. Remote sensing has proven to be the most effective method to monitoring the cropland in large scale by using temporary and spectral information. There are two main kinds of satellite have been used to mapping rice including microwave and optical. Rice, as the main crop of paddy fields, the main feature different from other crops is flooding phenomenon at planning stage (Figure 1). Microwave satellites can penetrate through clouds and efficiency on monitoring flooding phenomenon. Meanwhile, the vegetation index based on optical satellite can well distinguish rice from other vegetation. Google Earth Engine is a cloud-based platform that makes it easy to access high-performance computing resources for processing very large geospatial datasets. Google has collected large number of remote sensing satellite data around the world, which providing researchers with the possibility of doing application by using multi-source remote sensing data in a large area. In this work, we map rice planting area in south China through integration of Landsat-8 OLI, Sentienl-2, and Sentinel-1 Synthetic Aperture Radar (SAR) images. The flowchart is shown in figure 2. First, a threshold method the VH polarized backscatter from SAR sensor and vegetation index including normalized difference vegetation index (NDVI) and enhanced vegetation index (EVI) from optical sensor were used the classify the rice extent map. The forest and water surface extent map provided by earth engine were used to mask forest and water. To overcome the problem of the "salt and pepper effect" by Pixel-based classification when the spatial resolution increased, we segment the optical image and use the pixel- based classification results to merge the object-oriented segmentation data, and finally get the rice extent map. At last, by using the time series analysis, the peak count was obtained for each rice area to ensure the crop intensity. In this work, the rice ground point from a GVG crowdsourcing smartphone and rice area statistical results from National Bureau of Statistics were used to validate and evaluate our result.
NASA Astrophysics Data System (ADS)
Manaud, Nicolas; Carter, John; Boix, Oriol
2016-10-01
The "Where On Mars?" project is essentially the evolution of an existing outreach product developed in collaboration between ESA and CartoDB; an interactive map visualisation of the ESA's ExoMars Rover candidate landing sites (whereonmars.co). Planetary imagery data and maps are increasingly produced by the scientific community, and shared typically as images, in scientific publications, presentations or public outreach websites. However, this media lacks of interactivity and contextual information available for further exploration, making it difficult for any audience to relate one location-based information to another. We believe that interactive web maps are a powerful way of telling stories, engaging with and educating people who, over the last decade, have become familiar with tools such as Google Maps. A few planetary web maps exist but they are either too complex for non-experts, or are closed-systems that do not allows anyone to publish and share content. The long-term vision for the project is to provide researchers, communicators, educators and a worldwide public with an open planetary mapping and social platform enabling them to create, share, communicate and consume research-based content. We aim for this platform to become the reference website everyone will go to learn about Mars and other planets in our Solar System; just like people head to Google Maps to find their bearings or any location-based information. The driver is clearly to create for people an emotional connection with Mars. The short-term objectives for the project are (1) to produce and curate an open repository of basemaps, geospatial data sets, map visualisations, and story maps; (2) to develop a beautifully crafted and engaging interactive map of Mars. Based on user-generated content, the underlying framework should (3) make it easy to create and share additional interactive maps telling specific stories.
Measuring Neighborhood Walkable Environments: A Comparison of Three Approaches
Chiang, Yen-Cheng; Sullivan, William; Larsen, Linda
2017-01-01
Multiple studies have revealed the impact of walkable environments on physical activity. Scholars attach considerable importance to leisure and health-related walking. Recent studies have used Google Street View as an instrument to assess city streets and walkable environments; however, no study has compared the validity of Google Street View assessments of walkable environment attributes to assessments made by local residents and compiled from field visits. In this study, we involved nearby residents and compared the extent to which Google Street View assessments of the walkable environment correlated with assessments from local residents and with field visits. We determined the assessment approaches (local resident or field visit assessments) that exhibited the highest agreement with Google Street View. One city with relatively high-quality walkable environments and one city with relatively low-quality walkable environments were examined, and three neighborhoods from each city were surveyed. Participants in each neighborhood used one of three approaches to assess the walkability of the environment: 15 local residents assessed the environment using a map, 15 participants made a field visit to assess the environment, and 15 participants used Google Street View to assess the environment, yielding a total of 90 valid samples for the two cities. Findings revealed that the three approaches to assessing neighborhood walkability were highly correlated for traffic safety, aesthetics, sidewalk quality, and physical barriers. Compared with assessments from participants making field visits, assessments by local residents were more highly correlated with Google Street View assessments. Google Street View provides a more convenient, low-cost, efficient, and safe approach to assess neighborhood walkability. The results of this study may facilitate future large-scale walkable environment surveys, effectively reduce expenses, and improve survey efficiency. PMID:28587186
Vanwolleghem, Griet; Van Dyck, Delfien; Ducheyne, Fabian; De Bourdeaudhuij, Ilse; Cardon, Greet
2014-06-10
Google Street View provides a valuable and efficient alternative to observe the physical environment compared to on-site fieldwork. However, studies on the use, reliability and validity of Google Street View in a cycling-to-school context are lacking. We aimed to study the intra-, inter-rater reliability and criterion validity of EGA-Cycling (Environmental Google Street View Based Audit - Cycling to school), a newly developed audit using Google Street View to assess the physical environment along cycling routes to school. Parents (n = 52) of 11-to-12-year old Flemish children, who mostly cycled to school, completed a questionnaire and identified their child's cycling route to school on a street map. Fifty cycling routes of 11-to-12-year olds were identified and physical environmental characteristics along the identified routes were rated with EGA-Cycling (5 subscales; 37 items), based on Google Street View. To assess reliability, two researchers performed the audit. Criterion validity of the audit was examined by comparing the ratings based on Google Street View with ratings through on-site assessments. Intra-rater reliability was high (kappa range 0.47-1.00). Large variations in the inter-rater reliability (kappa range -0.03-1.00) and criterion validity scores (kappa range -0.06-1.00) were reported, with acceptable inter-rater reliability values for 43% of all items and acceptable criterion validity for 54% of all items. EGA-Cycling can be used to assess physical environmental characteristics along cycling routes to school. However, to assess the micro-environment specifically related to cycling, on-site assessments have to be added.
Field Ground Truthing Data Collector - a Mobile Toolkit for Image Analysis and Processing
NASA Astrophysics Data System (ADS)
Meng, X.
2012-07-01
Field Ground Truthing Data Collector is one of the four key components of the NASA funded ICCaRS project, being developed in Southeast Michigan. The ICCaRS ground truthing toolkit entertains comprehensive functions: 1) Field functions, including determining locations through GPS, gathering and geo-referencing visual data, laying out ground control points for AEROKAT flights, measuring the flight distance and height, and entering observations of land cover (and use) and health conditions of ecosystems and environments in the vicinity of the flight field; 2) Server synchronization functions, such as, downloading study-area maps, aerial photos and satellite images, uploading and synchronizing field-collected data with the distributed databases, calling the geospatial web services on the server side to conduct spatial querying, image analysis and processing, and receiving the processed results in field for near-real-time validation; and 3) Social network communication functions for direct technical assistance and pedagogical support, e.g., having video-conference calls in field with the supporting educators, scientists, and technologists, participating in Webinars, or engaging discussions with other-learning portals. This customized software package is being built on Apple iPhone/iPad and Google Maps/Earth. The technical infrastructures, data models, coupling methods between distributed geospatial data processing and field data collector tools, remote communication interfaces, coding schema, and functional flow charts will be illustrated and explained at the presentation. A pilot case study will be also demonstrated.
Google Scholar is not enough to be used alone for systematic reviews.
Giustini, Dean; Boulos, Maged N Kamel
2013-01-01
Google Scholar (GS) has been noted for its ability to search broadly for important references in the literature. Gehanno et al. recently examined GS in their study: 'Is Google scholar enough to be used alone for systematic reviews?' In this paper, we revisit this important question, and some of Gehanno et al.'s other findings in evaluating the academic search engine. The authors searched for a recent systematic review (SR) of comparable size to run search tests similar to those in Gehanno et al. We selected Chou et al. (2013) contacting the authors for a list of publications they found in their SR on social media in health. We queried GS for each of those 506 titles (in quotes "), one by one. When GS failed to retrieve a paper, or produced too many results, we used the allintitle: command to find papers with the same title. Google Scholar produced records for ~95% of the papers cited by Chou et al. (n=476/506). A few of the 30 papers that were not in GS were later retrieved via PubMed and even regular Google Search. But due to its different structure, we could not run searches in GS that were originally performed by Chou et al. in PubMed, Web of Science, Scopus and PsycINFO®. Identifying 506 papers in GS was an inefficient process, especially for papers using similar search terms. Has Google Scholar improved enough to be used alone in searching for systematic reviews? No. GS' constantly-changing content, algorithms and database structure make it a poor choice for systematic reviews. Looking for papers when you know their titles is a far different issue from discovering them initially. Further research is needed to determine when and how (and for what purposes) GS can be used alone. Google should provide details about GS' database coverage and improve its interface (e.g., with semantic search filters, stored searching, etc.). Perhaps then it will be an appropriate choice for systematic reviews.
NASA Astrophysics Data System (ADS)
Zetterlind, V.; Pledgie, S.
2009-12-01
Low-cost, low-latency, robust geolocation and display of aerial video is a common need for a wide range of earth observing as well as emergency response and security applications. While hardware costs for aerial video collection systems, GPS, and inertial sensors have been decreasing, software costs for geolocation algorithms and reference imagery/DTED remain expensive and highly proprietary. As part of a Federal Small Business Innovative Research project, MosaicATM and EarthNC, Inc have developed a simple geolocation system based on the Google Earth API and Google's 'built-in' DTED and reference imagery libraries. This system geolocates aerial video based on platform and camera position, attitude, and field-of-view metadata using geometric photogrammetric principles of ray-intersection with DTED. Geolocated video can be directly rectified and viewed in the Google Earth API during processing. Work is underway to extend our geolocation code to NASA World Wind for additional flexibility and a fully open-source platform. In addition to our airborne remote sensing work, MosaicATM has developed the Surface Operations Data Analysis and Adaptation (SODAA) tool, funded by NASA Ames, which supports analysis of airport surface operations to optimize aircraft movements and reduce fuel burn and delays. As part of SODAA, MosaicATM and EarthNC, Inc have developed powerful tools to display national airspace data and time-animated 3D flight tracks in Google Earth for 4D analysis. The SODAA tool can convert raw format flight track data, FAA National Flight Data (NFD), and FAA 'Adaptation' airport surface data to a spatial database representation and then to Google Earth KML. The SODAA client provides users with a simple graphical interface through which to generate queries with a wide range of predefined and custom filters, plot results, and export for playback in Google Earth in conjunction with NFD and Adaptation overlays.
Barnhill, Rick; Heermann-Do, Kimberly A; Salzman, Keith L; Gimbel, Ronald W
2011-01-01
Objective To design, build, implement, and evaluate a personal health record (PHR), tethered to the Military Health System, that leverages Microsoft® HealthVault and Google® Health infrastructure based on user preference. Materials and methods A pilot project was conducted in 2008–2009 at Madigan Army Medical Center in Tacoma, Washington. Our PHR was architected to a flexible platform that incorporated standards-based models of Continuity of Document and Continuity of Care Record to map Department of Defense-sourced health data, via a secure Veterans Administration data broker, to Microsoft® HealthVault and Google® Health based on user preference. The project design and implementation were guided by provider and patient advisory panels with formal user evaluation. Results The pilot project included 250 beneficiary users. Approximately 73.2% of users were <65 years of age, and 38.4% were female. Of the users, 169 (67.6%) selected Microsoft® HealthVault, and 81 (32.4%) selected Google® Health as their PHR of preference. Sample evaluation of users reflected 100% (n=60) satisfied with convenience of record access and 91.7% (n=55) satisfied with overall functionality of PHR. Discussion Key lessons learned related to data-transfer decisions (push vs pull), purposeful delays in reporting sensitive information, understanding and mapping PHR use and clinical workflow, and decisions on information patients may choose to share with their provider. Conclusion Currently PHRs are being viewed as empowering tools for patient activation. Design and implementation issues (eg, technical, organizational, information security) are substantial and must be thoughtfully approached. Adopting standards into design can enhance the national goal of portability and interoperability. PMID:21292705
Usability evaluation of cloud-based mapping tools for the display of very large datasets
NASA Astrophysics Data System (ADS)
Stotz, Nicole Marie
The elasticity and on-demand nature of cloud services have made it easier to create web maps. Users only need access to a web browser and the Internet to utilize cloud based web maps, eliminating the need for specialized software. To encourage a wide variety of users, a map must be well designed; usability is a very important concept in designing a web map. Fusion Tables, a new product from Google, is one example of newer cloud-based distributed GIS services. It allows for easy spatial data manipulation and visualization, within the Google Maps framework. ESRI has also introduced a cloud based version of their software, called ArcGIS Online, built on Amazon's EC2 cloud. Utilizing a user-centered design framework, two prototype maps were created with data from the San Diego East County Economic Development Council. One map was built on Fusion Tables, and another on ESRI's ArcGIS Online. A usability analysis was conducted and used to compare both map prototypes in term so of design and functionality. Load tests were also ran, and performance metrics gathered on both map prototypes. The usability analysis was taken by 25 geography students, and consisted of time based tasks and questions on map design and functionality. Survey participants completed the time based tasks for the Fusion Tables map prototype quicker than those of the ArcGIS Online map prototype. While response was generally positive towards the design and functionality of both prototypes, overall the Fusion Tables map prototype was preferred. For the load tests, the data set was broken into 22 groups for a total of 44 tests. While the Fusion Tables map prototype performed more efficiently than the ArcGIS Online prototype, differences are almost unnoticeable. A SWOT analysis was conducted for each prototype. The results from this research point to the Fusion Tables map prototype. A redesign of this prototype would incorporate design suggestions from the usability survey, while some functionality would need to be dropped. This is a free product and would therefore be the best option if cost is an issue, but this map may not be supported in the future.
NASA Astrophysics Data System (ADS)
Cardille, J. A.; Crowley, M.; Fortin, J. A.; Lee, J.; Perez, E.; Sleeter, B. M.; Thau, D.
2016-12-01
With the opening of the Landsat archive, researchers have a vast new data source teeming with imagery and potential. Beyond Landsat, data from other sensors is newly available as well: these include ALOS/PALSAR, Sentinel-1 and -2, MERIS, and many more. Google Earth Engine, developed to organize and provide analysis tools for these immense data sets, is an ideal platform for researchers trying to sift through huge image stacks. It offers nearly unlimited processing power and storage with a straightforward programming interface. Yet labeling land-cover change through time remains challenging given the current state of the art for interpreting remote sensing image sequences. Moreover, combining data from very different image platforms remains quite difficult. To address these challenges, we developed the BULC algorithm (Bayesian Updating of Land Cover), designed for the continuous updating of land-cover classifications through time in large data sets. The algorithm ingests data from any of the wide variety of earth-resources sensors; it maintains a running estimate of land-cover probabilities and the most probable class at all time points along a sequence of events. Here we compare BULC results from two study sites that witnessed considerable forest change in the last 40 years: the Pacific Northwest of the United States and the Mato Grosso region of Brazil. In Brazil, we incorporated rough classifications from more than 100 images of varying quality, mixing imagery from more than 10 different sensors. In the Pacific Northwest, we used BULC to identify forest changes due to logging and urbanization from 1973 to the present. Both regions had classification sequences that were better than many of the component days, effectively ignoring clouds and other unwanted noise while fusing the information contained on several platforms. As we leave remote sensing's data-poor era and enter a period with multiple looks at Earth's surface from multiple sensors over a short period of time, the BULC algorithm can help to sift through images of varying quality in Google Earth Engine to extract the most useful information for mapping the state and history of Earth's land cover.
NASA Astrophysics Data System (ADS)
Cardille, J. A.
2015-12-01
With the opening of the Landsat archive, researchers have a vast new data source teeming with imagery and potential. Beyond Landsat, data from other sensors is newly available as well: these include ALOS/PALSAR, Sentinel-1 and -2, MERIS, and many more. Google Earth Engine, developed to organize and provide analysis tools for these immense data sets, is an ideal platform for researchers trying to sift through huge image stacks. It offers nearly unlimited processing power and storage with a straightforward programming interface. Yet labeling forest change through time remains challenging given the current state of the art for interpreting remote sensing image sequences. Moreover, combining data from very different image platforms remains quite difficult. To address these challenges, we developed the BULC algorithm (Bayesian Updating of Land Cover), designed for the continuous updating of land-cover classifications through time in large data sets. The algorithm ingests data from any of the wide variety of earth-resources sensors; it maintains a running estimate of land-cover probabilities and the most probable class at all time points along a sequence of events. Here we compare BULC results from two study sites that witnessed considerable forest change in the last 40 years: the Pacific Northwest of the United States and the Mato Grosso region of Brazil. In Brazil, we incorporated rough classifications from more than 100 images of varying quality, mixing imagery from more than 10 different sensors. In the Pacific Northwest, we used BULC to identify forest changes due to logging and urbanization from 1973 to the present. Both regions had classification sequences that were better than many of the component days, effectively ignoring clouds and other unwanted signal while fusing the information contained on several platforms. As we leave remote sensing's data-poor era and enter a period with multiple looks at Earth's surface from multiple sensors over a short period of time, this algorithm may help to sift through images of varying quality in Google Earth Engine to extract the most useful information for mapping.
DistMap: a toolkit for distributed short read mapping on a Hadoop cluster.
Pandey, Ram Vinay; Schlötterer, Christian
2013-01-01
With the rapid and steady increase of next generation sequencing data output, the mapping of short reads has become a major data analysis bottleneck. On a single computer, it can take several days to map the vast quantity of reads produced from a single Illumina HiSeq lane. In an attempt to ameliorate this bottleneck we present a new tool, DistMap - a modular, scalable and integrated workflow to map reads in the Hadoop distributed computing framework. DistMap is easy to use, currently supports nine different short read mapping tools and can be run on all Unix-based operating systems. It accepts reads in FASTQ format as input and provides mapped reads in a SAM/BAM format. DistMap supports both paired-end and single-end reads thereby allowing the mapping of read data produced by different sequencing platforms. DistMap is available from http://code.google.com/p/distmap/
DistMap: A Toolkit for Distributed Short Read Mapping on a Hadoop Cluster
Pandey, Ram Vinay; Schlötterer, Christian
2013-01-01
With the rapid and steady increase of next generation sequencing data output, the mapping of short reads has become a major data analysis bottleneck. On a single computer, it can take several days to map the vast quantity of reads produced from a single Illumina HiSeq lane. In an attempt to ameliorate this bottleneck we present a new tool, DistMap - a modular, scalable and integrated workflow to map reads in the Hadoop distributed computing framework. DistMap is easy to use, currently supports nine different short read mapping tools and can be run on all Unix-based operating systems. It accepts reads in FASTQ format as input and provides mapped reads in a SAM/BAM format. DistMap supports both paired-end and single-end reads thereby allowing the mapping of read data produced by different sequencing platforms. DistMap is available from http://code.google.com/p/distmap/ PMID:24009693
Tropical Cyclone Information System
NASA Technical Reports Server (NTRS)
Li, P. Peggy; Knosp, Brian W.; Vu, Quoc A.; Yi, Chao; Hristova-Veleva, Svetla M.
2009-01-01
The JPL Tropical Cyclone Infor ma tion System (TCIS) is a Web portal (http://tropicalcyclone.jpl.nasa.gov) that provides researchers with an extensive set of observed hurricane parameters together with large-scale and convection resolving model outputs. It provides a comprehensive set of high-resolution satellite (see figure), airborne, and in-situ observations in both image and data formats. Large-scale datasets depict the surrounding environmental parameters such as SST (Sea Surface Temperature) and aerosol loading. Model outputs and analysis tools are provided to evaluate model performance and compare observations from different platforms. The system pertains to the thermodynamic and microphysical structure of the storm, the air-sea interaction processes, and the larger-scale environment as depicted by ocean heat content and the aerosol loading of the environment. Currently, the TCIS is populated with satellite observations of all tropical cyclones observed globally during 2005. There is a plan to extend the database both forward in time till present as well as backward to 1998. The portal is powered by a MySQL database and an Apache/Tomcat Web server on a Linux system. The interactive graphic user interface is provided by Google Map.
EarthScope Plate Boundary Observatory Data in the College Classroom (Invited)
NASA Astrophysics Data System (ADS)
Eriksson, S. C.; Olds, S. E.
2009-12-01
The Plate Boundary Observatory (PBO) is the geodetic component of the EarthScope project, designed to study the 3-D strain field across the active boundary zone between the Pacific and North American tectonics plates in the western United States. All PBO data are freely available to scientific and educational communities and have been incorporated into a variety of activities for college and university classrooms. UNAVCO Education and Outreach program staff have worked closely with faculty users, scientific researchers, and facility staff to create materials that are scientifically and technically accurate as well as useful to the classroom user. Availability of processed GPS data is not new to the geoscience community. However, PBO data staff have worked with education staff to deliver data that are readily accessible to educators. The UNAVCO Data for Educators webpage, incorporating an embedded Google Map with PBO GPS locations and providing current GPS time series plots and downloadable data, extends and updates the datasets available to our community. Google Earth allows the visualization GPS data with other types of datasets, e.g. LiDAR, while maintaining the self-contained and easy-to-use interface of UNAVCO’s Jules Verne Voyager map tools, which have multiple sets of geological and geophysical data. Curricular materials provide scaffolds for using EarthScope data in a variety of forms for different learning goals. Simple visualization of earthquake epicenters and locations of volcanoes can be used with velocity vectors to make simple deductions of plate boundary behaviors. Readily available time series plots provide opportunities for additional science skills, and there are web and paper-based support materials for downloading data, manipulating tables, and using plotting programs for processed GPS data. Scientists have provided contextual materials to explore the importance of these data in interpreting the structure and dynamics of the Earth. These data and their scientific context are now incorporated into the Active Earth Display developed by IRIS. Formal and informal evaluations during the past five years have provided useful data for revision and on-line implementation.
Large Scale Crop Mapping in Ukraine Using Google Earth Engine
NASA Astrophysics Data System (ADS)
Shelestov, A.; Lavreniuk, M. S.; Kussul, N.
2016-12-01
There are no globally available high resolution satellite-derived crop specific maps at present. Only coarse-resolution imagery (> 250 m spatial resolution) has been utilized to derive global cropland extent. In 2016 we are going to carry out a country level demonstration of Sentinel-2 use for crop classification in Ukraine within the ESA Sen2-Agri project. But optical imagery can be contaminated by cloud cover that makes it difficult to acquire imagery in an optimal time range to discriminate certain crops. Due to the Copernicus program since 2015, a lot of Sentinel-1 SAR data at high spatial resolution is available for free for Ukraine. It allows us to use the time series of SAR data for crop classification. Our experiment for one administrative region in 2015 showed much higher crop classification accuracy with SAR data than with optical only time series [1, 2]. Therefore, in 2016 within the Google Earth Engine Research Award we use SAR data together with optical ones for large area crop mapping (entire territory of Ukraine) using cloud computing capabilities available at Google Earth Engine (GEE). This study compares different classification methods for crop mapping for the whole territory of Ukraine using data and algorithms from GEE. Classification performance assessed using overall classification accuracy, Kappa coefficients, and user's and producer's accuracies. Also, crop areas from derived classification maps compared to the official statistics [3]. S. Skakun et al., "Efficiency assessment of multitemporal C-band Radarsat-2 intensity and Landsat-8 surface reflectance satellite imagery for crop classification in Ukraine," IEEE Journal of Selected Topics in Applied Earth Observ. and Rem. Sens., 2015, DOI: 10.1109/JSTARS.2015.2454297. N. Kussul, S. Skakun, A. Shelestov, O. Kussul, "The use of satellite SAR imagery to crop classification in Ukraine within JECAM project," IEEE International Geoscience and Remote Sensing Symposium (IGARSS), pp.1497-1500, 13-18 July 2014, Quebec City, Canada. F.J. Gallego, N. Kussul, S. Skakun, O. Kravchenko, A. Shelestov, O. Kussul, "Efficiency assessment of using satellite data for crop area estimation in Ukraine," International Journal of Applied Earth Observation and Geoinformation vol. 29, pp. 22-30, 2014.
Results of Prospecting of Impact Craters in Morocco
NASA Astrophysics Data System (ADS)
Chaabout, S.; Chennaoui Aoudjehane, H.; Reimold, W. U.; Baratoux, D.
2014-09-01
This work is based to use satellite images of Google Earth and Yahoo-Maps scenes; we examined the surface of our country to be able to locate the structures that have a circular morphology such as impact craters, which potentially could be.
NASA Astrophysics Data System (ADS)
Davias, M. E.; Gilbride, J. L.
2011-12-01
Aerial photographs of Carolina bays taken in the 1930's sparked the initial research into their geomorphology. Satellite Imagery available today through the Google Earth Virtual Globe facility expands the regions available for interrogation, but reveal only part of their unique planforms. Digital Elevation Maps (DEMs), using Light Detection And Ranging (LiDAR) remote sensing data, accentuate the visual presentation of these aligned ovoid shallow basins by emphasizing their robust circumpheral rims. To support a geospatial survey of Carolina bay landforms in the continental USA, 400,000 km2 of hsv-shaded DEMs were created as KML-JPEG tile sets. A majority of these DEMs were generated with LiDAR-derived data. We demonstrate the tile generation process and their integration into Google Earth, where the DEMs augment available photographic imagery for the visualization of bay planforms. While the generic Carolina bay planform is considered oval, we document subtle regional variations. Using a small set of empirically derived planform shapes, we created corresponding Google Earth overlay templates. We demonstrate the analysis of an individual Carolina bay by placing an appropriate overlay onto the virtually globe, then orientating, sizing and rotating it by edit handles such that it satisfactorily represents the bay's rim. The resulting overlay data element is extracted from Google Earth's object directory and programmatically processed to generate metrics such as geographic location, elevation, major and minor axis and inferred orientation. Utilizing a virtual globe facility for data capture may result in higher quality data compared to methods that reference flat maps, where geospatial shape and orientation of the bays could be skewed and distorted in the orthographic projection process. Using the methodology described, we have measured over 25k distinct Carolina bays. We discuss the Google Fusion geospatial data repository facility, through which these data have been assembled and made web-accessible to other researchers. Preliminary findings from the survey are discussed, such as how bay surface area, eccentricity and orientation vary across ~800 1/4° × 1/4° grid elements. Future work includes measuring 25k additional bays, as well as interrogation of the orientation data to identify any possible systematic geospatial relationships.
Mapping of Sample Collection Data: GIS Tools for the Natural Product Researcher
Oberlies, Nicholas H.; Rineer, James I.; Alali, Feras Q.; Tawaha, Khaled; Falkinham, Joseph O.; Wheaton, William D.
2009-01-01
Scientists engaged in the research of natural products often either conduct field collections themselves or collaborate with partners who do, such as botanists, mycologists, or SCUBA divers. The information gleaned from such collecting trips (e.g. longitude/latitude coordinates, geography, elevation, and a multitude of other field observations) have provided valuable data to the scientific community (e.g., biodiversity), even if it is tangential to the direct aims of the natural products research, which are often focused on drug discovery and/or chemical ecology. Geographic Information Systems (GIS) have been used to display, manage, and analyze geographic data, including collection sites for natural products. However, to the uninitiated, these tools are often beyond the financial and/or computational means of the natural product scientist. With new, free, and easy-to-use geospatial visualization tools, such as Google Earth, mapping and geographic imaging of sampling data are now within the reach of natural products scientists. The goals of the present study were to develop simple tools that are tailored for the natural products setting, thereby presenting a means to map such information, particularly via open source software like Google Earth. PMID:20161345
Satellite Radar Detects Damage from Sept. 19, 2017 Raboso, Mexico, Quake
2017-09-20
The Advanced Rapid Imaging and Analysis (ARIA) team at NASA's Jet Propulsion Laboratory in Pasadena, California, and Caltech, also in Pasadena, created this Damage Proxy Map (DPM) depicting areas of Central Mexico, including Mexico City, that are likely damaged (shown by red and yellow pixels) from the magnitude 7.1 Raboso earthquake of Sept. 19, 2017 (local time). The map is derived from synthetic aperture radar (SAR) images from the Copernicus Sentinel-1A and Sentinel-1B satellites, operated by the European Space Agency (ESA). The images were taken before (Sept. 8, 2017) and after (Sept. 20, 2017) the earthquake. The map covers an area of 109 by 106 miles (175 by 170 kilometers). Each pixel measures about 33 yards (30 meters) across. The color variation from yellow to red indicates increasingly more significant ground and building surface change. Preliminary validation was done by comparing the DPM to a crowd-sourced Google Map (https://www.google.com/maps/d/u/0/viewer?mid=1_-V97lbdgLFHpx-CtqhLWlJAnYY&ll=19.41452166501326%2C-99.16498240436704&z=16). This damage proxy map should be used as guidance to identify damaged areas, and may be less reliable over vegetated areas. Sentinel-1 data were accessed through the Copernicus Open Access Hub. The image contains modified Copernicus Sentinel data (2017), processed by ESA and analyzed by the NASA-JPL/Caltech ARIA team. This research was carried out at JPL under contract with NASA. https://photojournal.jpl.nasa.gov/catalog/PIA21963
CERESVis: A QC Tool for CERES that Leverages Browser Technology for Data Validation
NASA Astrophysics Data System (ADS)
Chu, C.; Sun-Mack, S.; Heckert, E.; Chen, Y.; Doelling, D.
2015-12-01
In this poster, we are going to present three user interfaces that CERES team uses to validate pixel-level data. Besides our home grown tools, we will aslo present the browser technology that we use to provide interactive interfaces, such as jquery, HighCharts and Google Earth. We pass data to the users' browsers and use the browsers to do some simple computations. The three user interfaces are: Thumbnails -- it displays hundrends images to allow users to browse 24-hour data files in few seconds. Multiple-synchronized cursors -- it allows users to compare multiple images side by side. Bounding Boxes and Histograms -- it allows users to draw multiple bounding boxes on an image and the browser computes/display the histograms.
Cole, Justin; Beare, Richard; Phan, Thanh G; Srikanth, Velandai; MacIsaac, Andrew; Tan, Christianne; Tong, David; Yee, Susan; Ho, Jesslyn; Layland, Jamie
2017-01-01
Recent evidence suggests hospitals fail to meet guideline specified time to percutaneous coronary intervention (PCI) for a proportion of ST elevation myocardial infarction (STEMI) presentations. Implicit in achieving this time is the rapid assembly of crucial catheter laboratory staff. As a proof-of-concept, we set out to create regional maps that graphically show the impact of traffic congestion and distance to destination on staff recall travel times for STEMI, thereby producing a resource that could be used by staff to improve reperfusion time for STEMI. Travel times for staff recalled to one inner and one outer metropolitan hospital at midnight, 6 p.m., and 7 a.m. were estimated using Google Maps Application Programming Interface. Computer modeling predictions were overlaid on metropolitan maps showing color coded staff recall travel times for STEMI, occurring within non-peak and peak hour traffic congestion times. Inner metropolitan hospital staff recall travel times were more affected by traffic congestion compared with outer metropolitan times, and the latter was more affected by distance. The estimated mean travel times to hospital during peak hour were greater than midnight travel times by 13.4 min to the inner and 6.0 min to the outer metropolitan hospital at 6 p.m. ( p < 0.001). At 7 a.m., the mean difference was 9.5 min to the inner and 3.6 min to the outer metropolitan hospital ( p < 0.001). Only 45% of inner metropolitan staff were predicted to arrive within 30 min at 6 p.m. compared with 100% at midnight ( p < 0.001), and 56% of outer metropolitan staff at 6 p.m. ( p = 0.021). Our results show that integration of map software with traffic congestion data, distance to destination and travel time can predict optimal residence of staff when on-call for PCI.
Human guidance of mobile robots in complex 3D environments using smart glasses
NASA Astrophysics Data System (ADS)
Kopinsky, Ryan; Sharma, Aneesh; Gupta, Nikhil; Ordonez, Camilo; Collins, Emmanuel; Barber, Daniel
2016-05-01
In order for humans to safely work alongside robots in the field, the human-robot (HR) interface, which enables bi-directional communication between human and robot, should be able to quickly and concisely express the robot's intentions and needs. While the robot operates mostly in autonomous mode, the human should be able to intervene to effectively guide the robot in complex, risky and/or highly uncertain scenarios. Using smart glasses such as Google Glass∗, we seek to develop an HR interface that aids in reducing interaction time and distractions during interaction with the robot.
Measurable realistic image-based 3D mapping
NASA Astrophysics Data System (ADS)
Liu, W.; Wang, J.; Wang, J. J.; Ding, W.; Almagbile, A.
2011-12-01
Maps with 3D visual models are becoming a remarkable feature of 3D map services. High-resolution image data is obtained for the construction of 3D visualized models.The3D map not only provides the capabilities of 3D measurements and knowledge mining, but also provides the virtual experienceof places of interest, such as demonstrated in the Google Earth. Applications of 3D maps are expanding into the areas of architecture, property management, and urban environment monitoring. However, the reconstruction of high quality 3D models is time consuming, and requires robust hardware and powerful software to handle the enormous amount of data. This is especially for automatic implementation of 3D models and the representation of complicated surfacesthat still need improvements with in the visualisation techniques. The shortcoming of 3D model-based maps is the limitation of detailed coverage since a user can only view and measure objects that are already modelled in the virtual environment. This paper proposes and demonstrates a 3D map concept that is realistic and image-based, that enables geometric measurements and geo-location services. Additionally, image-based 3D maps provide more detailed information of the real world than 3D model-based maps. The image-based 3D maps use geo-referenced stereo images or panoramic images. The geometric relationships between objects in the images can be resolved from the geometric model of stereo images. The panoramic function makes 3D maps more interactive with users but also creates an interesting immersive circumstance. Actually, unmeasurable image-based 3D maps already exist, such as Google street view, but only provide virtual experiences in terms of photos. The topographic and terrain attributes, such as shapes and heights though are omitted. This paper also discusses the potential for using a low cost land Mobile Mapping System (MMS) to implement realistic image 3D mapping, and evaluates the positioning accuracy that a measureable realistic image-based (MRI) system can produce. The major contribution here is the implementation of measurable images on 3D maps to obtain various measurements from real scenes.
Mapping the Diagnosis Axis of an Interface Terminology to the NANDA International Taxonomy
Juvé Udina, Maria-Eulàlia; Gonzalez Samartino, Maribel; Matud Calvo, Cristina
2012-01-01
Background. Nursing terminologies are designed to support nursing practice but, as with any other clinical tool, they should be evaluated. Cross-mapping is a formal method for examining the validity of the existing controlled vocabularies. Objectives. The study aims to assess the inclusiveness and expressiveness of the nursing diagnosis axis of a newly implemented interface terminology by cross-mapping with the NANDA-I taxonomy. Design/Methods. The study applied a descriptive design, using a cross-sectional, bidirectional mapping strategy. The sample included 728 concepts from both vocabularies. Concept cross-mapping was carried out to identify one-to-one, negative, and hierarchical connections. The analysis was conducted using descriptive statistics. Results. Agreement of the raters' mapping achieved 97%. More than 60% of the nursing diagnosis concepts in the NANDA-I taxonomy were mapped to concepts in the diagnosis axis of the new interface terminology; 71.1% were reversely mapped. Conclusions. Main results for outcome measures suggest that the diagnosis axis of this interface terminology meets the validity criterion of cross-mapping when mapped from and to the NANDA-I taxonomy. PMID:22830046
Mapping the Diagnosis Axis of an Interface Terminology to the NANDA International Taxonomy.
Juvé Udina, Maria-Eulàlia; Gonzalez Samartino, Maribel; Matud Calvo, Cristina
2012-01-01
Background. Nursing terminologies are designed to support nursing practice but, as with any other clinical tool, they should be evaluated. Cross-mapping is a formal method for examining the validity of the existing controlled vocabularies. Objectives. The study aims to assess the inclusiveness and expressiveness of the nursing diagnosis axis of a newly implemented interface terminology by cross-mapping with the NANDA-I taxonomy. Design/Methods. The study applied a descriptive design, using a cross-sectional, bidirectional mapping strategy. The sample included 728 concepts from both vocabularies. Concept cross-mapping was carried out to identify one-to-one, negative, and hierarchical connections. The analysis was conducted using descriptive statistics. Results. Agreement of the raters' mapping achieved 97%. More than 60% of the nursing diagnosis concepts in the NANDA-I taxonomy were mapped to concepts in the diagnosis axis of the new interface terminology; 71.1% were reversely mapped. Conclusions. Main results for outcome measures suggest that the diagnosis axis of this interface terminology meets the validity criterion of cross-mapping when mapped from and to the NANDA-I taxonomy.
NASA Astrophysics Data System (ADS)
Igarashi, Masayasu; Murao, Osamu
In this paper, the authors develop a multiple regression model which estimates urban earthquake vulnerability (building collapse risk and conflagration risk) for different eras, and clarify the historical changes of urban risk in Marunouchi and Ginza Districts in Tokyo, Japan using old maps and contemporary geographic information data. Also, we compare the change of urban vulnerability of the districts with the significant historical events in Tokyo. Finally, the results are loaded onto Google Earth with timescale extension to consider the possibility of urban recovery digital archives in the era of the recent geoinformatic technologies.
NASA Technical Reports Server (NTRS)
Markert, Kel; Ashmall, William; Johnson, Gary; Saah, David; Mollicone, Danilo; Diaz, Alfonso Sanchez-Paus; Anderson, Eric; Flores, Africa; Griffin, Robert
2017-01-01
Collect Earth Online (CEO) is a free and open online implementation of the FAO Collect Earth system for collaboratively collecting environmental data through the visual interpretation of Earth observation imagery. The primary collection mechanism in CEO is human interpretation of land surface characteristics in imagery served via Web Map Services (WMS). However, interpreters may not have enough contextual information to classify samples by only viewing the imagery served via WMS, be they high resolution or otherwise. To assist in the interpretation and collection processes in CEO, SERVIR, a joint NASA-USAID initiative that brings Earth observations to improve environmental decision making in developing countries, developed the GeoDash system, an embedded and critical component of CEO. GeoDash leverages Google Earth Engine (GEE) by allowing users to set up custom browser-based widgets that pull from GEE's massive public data catalog. These widgets can be quick looks of other satellite imagery, time series graphs of environmental variables, and statistics panels of the same. Users can customize widgets with any of GEE's image collections, such as the historical Landsat collection with data available since the 1970s, select date ranges, image stretch parameters, graph characteristics, and create custom layouts, all on-the-fly to support plot interpretation in CEO. This presentation focuses on the implementation and potential applications, including the back-end links to GEE and the user interface with custom widget building. GeoDash takes large data volumes and condenses them into meaningful, relevant information for interpreters. While designed initially with national and global forest resource assessments in mind, the system will complement disaster assessments, agriculture management, project monitoring and evaluation, and more.
There's An App For That: Planning Ahead for the Solar Eclipse in August 2017
NASA Astrophysics Data System (ADS)
Chizek Frouard, Malynda R.; Lesniak, Michael V.; Bell, Steve
2017-01-01
With the total solar eclipse of 2017 August 21 over the continental United States approaching, the U.S. Naval Observatory (USNO) on-line Solar Eclipse Computer can now be accessed via an Android application, available on Google Play.Over the course of the eclipse, as viewed from a specific site, several events may be visible: the beginning and ending of the eclipse (first and fourth contacts), the beginning and ending of totality (second and third contacts), the moment of maximum eclipse, sunrise, or sunset. For each of these events, the USNO Solar Eclipse 2017 Android application reports the time, Sun's altitude and azimuth, and the event's position and vertex angles. The app also lists the duration of the total phase, the duration of the eclipse, the magnitude of the eclipse, and the percent of the Sun obscured for a particular eclipse site.All of the data available in the app comes from the flexible USNO Solar Eclipse Computer Application Programming Interface (API), which produces JavaScript Object Notation (JSON) that can be incorporated into third-party Web sites or custom applications. Additional information is available in the on-line documentation (http://aa.usno.navy.mil/data/docs/api.php).For those who prefer using a traditional data input form, the local circumstances can still be requested at http://aa.usno.navy.mil/data/docs/SolarEclipses.php.In addition the 2017 August 21 Solar Eclipse Resource page (http://aa.usno.navy.mil/data/docs/Eclipse2017.php) consolidates all of the USNO resources for this event, including a Google Map view of the eclipse track designed by Her Majesty's Nautical Almanac Office (HMNAO).Looking further ahead, a 2024 April 8 Solar Eclipse Resource page (http://aa.usno.navy.mil/data/docs/Eclipse2024.php) is also available.
NASA Astrophysics Data System (ADS)
Markert, K. N.; Ashmall, W.; Johnson, G.; Saah, D. S.; Anderson, E.; Flores Cordova, A. I.; Díaz, A. S. P.; Mollicone, D.; Griffin, R.
2017-12-01
Collect Earth Online (CEO) is a free and open online implementation of the FAO Collect Earth system for collaboratively collecting environmental data through the visual interpretation of Earth observation imagery. The primary collection mechanism in CEO is human interpretation of land surface characteristics in imagery served via Web Map Services (WMS). However, interpreters may not have enough contextual information to classify samples by only viewing the imagery served via WMS, be they high resolution or otherwise. To assist in the interpretation and collection processes in CEO, SERVIR, a joint NASA-USAID initiative that brings Earth observations to improve environmental decision making in developing countries, developed the GeoDash system, an embedded and critical component of CEO. GeoDash leverages Google Earth Engine (GEE) by allowing users to set up custom browser-based widgets that pull from GEE's massive public data catalog. These widgets can be quick looks of other satellite imagery, time series graphs of environmental variables, and statistics panels of the same. Users can customize widgets with any of GEE's image collections, such as the historical Landsat collection with data available since the 1970s, select date ranges, image stretch parameters, graph characteristics, and create custom layouts, all on-the-fly to support plot interpretation in CEO. This presentation focuses on the implementation and potential applications, including the back-end links to GEE and the user interface with custom widget building. GeoDash takes large data volumes and condenses them into meaningful, relevant information for interpreters. While designed initially with national and global forest resource assessments in mind, the system will complement disaster assessments, agriculture management, project monitoring and evaluation, and more.
Building a SuAVE browse interface to R2R's Linked Data
NASA Astrophysics Data System (ADS)
Clark, D.; Stocks, K. I.; Arko, R. A.; Zaslavsky, I.; Whitenack, T.
2017-12-01
The Rolling Deck to Repository program (R2R) is creating and evaluating a new browse portal based on the SuAVE platform and the R2R linked data graph. R2R manages the underway sensor data collected by the fleet of US academic research vessels, and provides a discovery and access point to those data at its website, www.rvdata.us. R2R has a database-driven search interface, but seeks a more capable and extensible browse interface that could be built off of the substantial R2R linked data resources. R2R's Linked Data graph organizes its data holdings around key concepts (e.g. cruise, vessel, device type, operator, award, organization, publication), anchored by persistent identifiers where feasible. The "Survey Analysis via Visual Exploration" or SuAVE platform (suave.sdsc.edu) is a system for online publication, sharing, and analysis of images and metadata. It has been implemented as an interface to diverse data collections, but has not been driven off of linked data in the past. SuAVE supports several features of interest to R2R, including faceted searching, collaborative annotations, efficient subsetting, Google maps-like navigation over an image gallery, and several types of data analysis. Our initial SuAVE-based implementation was through a CSV export from the R2R PostGIS-enabled PostgreSQL database. This served to demonstrate the utility of SuAVE but was static and required reloading as R2R data holdings grew. We are now working to implement a SPARQL-based ("RDF Query Language") service that directly leverages the R2R Linked Data graph and offers the ability to subset and/or customize output.We will show examples of SuAVE faceted searches on R2R linked data concepts, and discuss our experience to date with this work in progress.
JournalMap: Geo-semantic searching for relevant knowledge
USDA-ARS?s Scientific Manuscript database
Ecologists struggling to understand rapidly changing environments and evolving ecosystem threats need quick access to relevant research and documentation of natural systems. The advent of semantic and aggregation searching (e.g., Google Scholar, Web of Science) has made it easier to find useful lite...
Gault, Lora V.; Shultz, Mary; Davies, Kathy J.
2002-01-01
Objectives: This study compared the mapping of natural language patron terms to the Medical Subject Headings (MeSH) across six MeSH interfaces for the MEDLINE database. Methods: Test data were obtained from search requests submitted by patrons to the Library of the Health Sciences, University of Illinois at Chicago, over a nine-month period. Search request statements were parsed into separate terms or phrases. Using print sources from the National Library of Medicine, Each parsed patron term was assigned corresponding MeSH terms. Each patron term was entered into each of the selected interfaces to determine how effectively they mapped to MeSH. Data were collected for mapping success, accessibility of MeSH term within mapped list, and total number of MeSH choices within each list. Results: The selected MEDLINE interfaces do not map the same patron term in the same way, nor do they consistently lead to what is considered the appropriate MeSH term. Conclusions: If searchers utilize the MEDLINE database to its fullest potential by mapping to MeSH, the results of the mapping will vary between interfaces. This variance may ultimately impact the search results. These differences should be considered when choosing a MEDLINE interface and when instructing end users. PMID:11999175
Gault, Lora V; Shultz, Mary; Davies, Kathy J
2002-04-01
This study compared the mapping of natural language patron terms to the Medical Subject Headings (MeSH) across six MeSH interfaces for the MEDLINE database. Test data were obtained from search requests submitted by patrons to the Library of the Health Sciences, University of Illinois at Chicago, over a nine-month period. Search request statements were parsed into separate terms or phrases. Using print sources from the National Library of Medicine, Each parsed patron term was assigned corresponding MeSH terms. Each patron term was entered into each of the selected interfaces to determine how effectively they mapped to MeSH. Data were collected for mapping success, accessibility of MeSH term within mapped list, and total number of MeSH choices within each list. The selected MEDLINE interfaces do not map the same patron term in the same way, nor do they consistently lead to what is considered the appropriate MeSH term. If searchers utilize the MEDLINE database to its fullest potential by mapping to MeSH, the results of the mapping will vary between interfaces. This variance may ultimately impact the search results. These differences should be considered when choosing a MEDLINE interface and when instructing end users.
A Mobile, Map-Based Tasking Interface for Human-Robot Interaction
2010-12-01
A MOBILE, MAP-BASED TASKING INTERFACE FOR HUMAN-ROBOT INTERACTION By Eli R. Hooten Thesis Submitted to the Faculty of the Graduate School of...SUBTITLE A Mobile, Map-Based Tasking Interface for Human-Robot Interaction 5a. CONTRACT NUMBER 5b. GRANT NUMBER 5c. PROGRAM ELEMENT NUMBER 6...3 II.1 Interactive Modalities and Multi-Touch . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3 II.2
New Delhi Metallo-beta-lactamase around the world: an eReview using Google Maps.
Berrazeg, M; Diene, Sm; Medjahed, L; Parola, P; Drissi, M; Raoult, D; Rolain, Jm
2014-05-22
Gram-negative carbapenem-resistant bacteria, in particular those producing New Delhi Metallo-betalactamase-1 (NDM-1), are a major global health problem. To inform the scientific and medical community in real time about worldwide dissemination of isolates of NDM-1-producing bacteria, we used the PubMed database to review all available publications from the first description in 2009 up to 31 December 2012, and created a regularly updated worldwide dissemination map using a web-based mapping application. We retrieved 33 reviews, and 136 case reports describing 950 isolates of NDM-1-producing bacteria. Klebsiella pneumoniae (n= 359) and Escherichia coli (n=268) were the most commonly reported bacteria producing NDM-1 enzyme. Several case reports of infections due to imported NDM-1 producing bacteria have been reported in a number of countries, including the United Kingdom, Italy, and Oman. In most cases (132/153, 86.3%), patients had connections with the Indian subcontinent or Balkan countries. Those infected were originally from these areas, had either spent time and/or been hospitalised there, or were potentially linked to other patients who had been hospitalised in these regions. By using Google Maps, we were able to trace spread of NDM-1-producing bacteria. We strongly encourage epidemiologists to use these types of interactive tools for surveillance purposes and use the information to prevent the spread and outbreaks of such bacteria.
Researchermap: a tool for visualizing author locations using Google maps.
Rastegar-Mojarad, Majid; Bales, Michael E; Yu, Hong
2013-01-01
We hereby present ResearcherMap, a tool to visualize locations of authors of scholarly papers. In response to a query, the system returns a map of author locations. To develop the system we first populated a database of author locations, geocoding institution locations for all available institutional affiliation data in our database. The database includes all authors of Medline papers from 1990 to 2012. We conducted a formative heuristic usability evaluation of the system and measured the system's accuracy and performance. The accuracy of finding the accurate address is 97.5% in our system.
Zhang, Huaizhong; Fan, Jun; Perkins, Simon; Pisconti, Addolorata; Simpson, Deborah M.; Bessant, Conrad; Hubbard, Simon; Jones, Andrew R.
2015-01-01
The mzQuantML standard has been developed by the Proteomics Standards Initiative for capturing, archiving and exchanging quantitative proteomic data, derived from mass spectrometry. It is a rich XML‐based format, capable of representing data about two‐dimensional features from LC‐MS data, and peptides, proteins or groups of proteins that have been quantified from multiple samples. In this article we report the development of an open source Java‐based library of routines for mzQuantML, called the mzqLibrary, and associated software for visualising data called the mzqViewer. The mzqLibrary contains routines for mapping (peptide) identifications on quantified features, inference of protein (group)‐level quantification values from peptide‐level values, normalisation and basic statistics for differential expression. These routines can be accessed via the command line, via a Java programming interface access or a basic graphical user interface. The mzqLibrary also contains several file format converters, including import converters (to mzQuantML) from OpenMS, Progenesis LC‐MS and MaxQuant, and exporters (from mzQuantML) to other standards or useful formats (mzTab, HTML, csv). The mzqViewer contains in‐built routines for viewing the tables of data (about features, peptides or proteins), and connects to the R statistical library for more advanced plotting options. The mzqLibrary and mzqViewer packages are available from https://code.google.com/p/mzq‐lib/. PMID:26037908
Qi, Da; Zhang, Huaizhong; Fan, Jun; Perkins, Simon; Pisconti, Addolorata; Simpson, Deborah M; Bessant, Conrad; Hubbard, Simon; Jones, Andrew R
2015-09-01
The mzQuantML standard has been developed by the Proteomics Standards Initiative for capturing, archiving and exchanging quantitative proteomic data, derived from mass spectrometry. It is a rich XML-based format, capable of representing data about two-dimensional features from LC-MS data, and peptides, proteins or groups of proteins that have been quantified from multiple samples. In this article we report the development of an open source Java-based library of routines for mzQuantML, called the mzqLibrary, and associated software for visualising data called the mzqViewer. The mzqLibrary contains routines for mapping (peptide) identifications on quantified features, inference of protein (group)-level quantification values from peptide-level values, normalisation and basic statistics for differential expression. These routines can be accessed via the command line, via a Java programming interface access or a basic graphical user interface. The mzqLibrary also contains several file format converters, including import converters (to mzQuantML) from OpenMS, Progenesis LC-MS and MaxQuant, and exporters (from mzQuantML) to other standards or useful formats (mzTab, HTML, csv). The mzqViewer contains in-built routines for viewing the tables of data (about features, peptides or proteins), and connects to the R statistical library for more advanced plotting options. The mzqLibrary and mzqViewer packages are available from https://code.google.com/p/mzq-lib/. © 2015 The Authors. PROTEOMICS Published by Wiley-VCH Verlag GmbH & Co. KGaA, Weinheim.
Kushniruk, Andre W; Kan, Min-Yem; McKeown, Kathleen; Klavans, Judith; Jordan, Desmond; LaFlamme, Mark; Patel, Vimia L
2002-01-01
This paper describes the comparative evaluation of an experimental automated text summarization system, Centrifuser and three conventional search engines - Google, Yahoo and About.com. Centrifuser provides information to patients and families relevant to their questions about specific health conditions. It then produces a multidocument summary of articles retrieved by a standard search engine, tailored to the user's question. Subjects, consisting of friends or family of hospitalized patients, were asked to "think aloud" as they interacted with the four systems. The evaluation involved audio- and video recording of subject interactions with the interfaces in situ at a hospital. Results of the evaluation show that subjects found Centrifuser's summarization capability useful and easy to understand. In comparing Centrifuser to the three search engines, subjects' ratings varied; however, specific interface features were deemed useful across interfaces. We conclude with a discussion of the implications for engineering Web-based retrieval systems.
NASA Astrophysics Data System (ADS)
Vásquez-Ramírez, Raquel; Alor-Hernández, Giner; Sánchez-Ramírez, Cuauhtémoc; Guzmán-Luna, Jaime; Zatarain-Cabada, Ramón; Barrón-Estrada, María-Lucía
2014-07-01
Education has become a key component of any society since it is the means by which humanity functions and governs itself. It allows individuals to appropriately integrate into a given community. For this reason, new ways of interaction between students and educational contents are emerging in order to improve the quality of education. In this context, devices such as computers, smartphones, or electronic tablets represent new ways of accessing educational resources which do not limit students to their usage merely inside the classroom since these devices are available anywhere. Nowadays, television has become one of these technological tools able to support the teaching-learning process through documentary films or movies, among others. However, two main issues appear. First, some of these educational contents are not those needed by a professor since information is restricted, and second, the development of TV-based applications requires an integrative approach involving the support of several specialists in education who provide the guidelines needed to build high-quality contents, as well as application designers and developers who are able to deliver the educational applications demanded by students. This work presents a system called AthenaTV to generate android-based educational applications for TV. AthenaTV takes into account the 10-foot design scheme used by Google to develop interfaces based on interface design patterns established in Google TV, and it is based on the android development guidelines and HTML5 standard.
Signalling maps in cancer research: construction and data analysis
Kondratova, Maria; Sompairac, Nicolas; Barillot, Emmanuel; Zinovyev, Andrei
2018-01-01
Abstract Generation and usage of high-quality molecular signalling network maps can be augmented by standardizing notations, establishing curation workflows and application of computational biology methods to exploit the knowledge contained in the maps. In this manuscript, we summarize the major aims and challenges of assembling information in the form of comprehensive maps of molecular interactions. Mainly, we share our experience gained while creating the Atlas of Cancer Signalling Network. In the step-by-step procedure, we describe the map construction process and suggest solutions for map complexity management by introducing a hierarchical modular map structure. In addition, we describe the NaviCell platform, a computational technology using Google Maps API to explore comprehensive molecular maps similar to geographical maps and explain the advantages of semantic zooming principles for map navigation. We also provide the outline to prepare signalling network maps for navigation using the NaviCell platform. Finally, several examples of cancer high-throughput data analysis and visualization in the context of comprehensive signalling maps are presented. PMID:29688383
The Self-Paced Graz Brain-Computer Interface: Methods and Applications
Scherer, Reinhold; Schloegl, Alois; Lee, Felix; Bischof, Horst; Janša, Janez; Pfurtscheller, Gert
2007-01-01
We present the self-paced 3-class Graz brain-computer interface (BCI) which is based on the detection of sensorimotor electroencephalogram (EEG) rhythms induced by motor imagery. Self-paced operation means that the BCI is able to determine whether the ongoing brain activity is intended as control signal (intentional control) or not (non-control state). The presented system is able to automatically reduce electrooculogram (EOG) artifacts, to detect electromyographic (EMG) activity, and uses only three bipolar EEG channels. Two applications are presented: the freeSpace virtual environment (VE) and the Brainloop interface. The freeSpace is a computer-game-like application where subjects have to navigate through the environment and collect coins by autonomously selecting navigation commands. Three subjects participated in these feedback experiments and each learned to navigate through the VE and collect coins. Two out of the three succeeded in collecting all three coins. The Brainloop interface provides an interface between the Graz-BCI and Google Earth. PMID:18350133
Providing Web Interfaces to the NSF EarthScope USArray Transportable Array
NASA Astrophysics Data System (ADS)
Vernon, Frank; Newman, Robert; Lindquist, Kent
2010-05-01
Since April 2004 the EarthScope USArray seismic network has grown to over 850 broadband stations that stream multi-channel data in near real-time to the Array Network Facility in San Diego. Providing secure, yet open, access to real-time and archived data for a broad range of audiences is best served by a series of platform agnostic low-latency web-based applications. We present a framework of tools that mediate between the world wide web and Boulder Real Time Technologies Antelope Environmental Monitoring System data acquisition and archival software. These tools provide comprehensive information to audiences ranging from network operators and geoscience researchers, to funding agencies and the general public. This ranges from network-wide to station-specific metadata, state-of-health metrics, event detection rates, archival data and dynamic report generation over a station's two year life span. Leveraging open source web-site development frameworks for both the server side (Perl, Python and PHP) and client-side (Flickr, Google Maps/Earth and jQuery) facilitates the development of a robust extensible architecture that can be tailored on a per-user basis, with rapid prototyping and development that adheres to web-standards. Typical seismic data warehouses allow online users to query and download data collected from regional networks, without the scientist directly visually assessing data coverage and/or quality. Using a suite of web-based protocols, we have recently developed an online seismic waveform interface that directly queries and displays data from a relational database through a web-browser. Using the Python interface to Datascope and the Python-based Twisted network package on the server side, and the jQuery Javascript framework on the client side to send and receive asynchronous waveform queries, we display broadband seismic data using the HTML Canvas element that is globally accessible by anyone using a modern web-browser. We are currently creating additional interface tools to create a rich-client interface for accessing and displaying seismic data that can be deployed to any system running the Antelope Real Time System. The software is freely available from the Antelope contributed code Git repository (http://www.antelopeusersgroup.org).
Brunger, Fern; Welch, Vivian; Asghari, Shabnam; Kaposy, Chris
2018-01-01
Background This paper focuses on the collision of three factors: a growing emphasis on sharing research through open access publication, an increasing awareness of big data and its potential uses, and an engaged public interested in the privacy and confidentiality of their personal health information. One conceptual space where this collision is brought into sharp relief is with the open availability of patient medical photographs from peer-reviewed journal articles in the search results of online image databases such as Google Images. Objective The aim of this study was to assess the availability of patient medical photographs from published journal articles in Google Images search results and the factors impacting this availability. Methods We conducted a cross-sectional study using data from an evidence map of research with transgender, gender non-binary, and other gender diverse (trans) participants. For the original evidence map, a comprehensive search of 15 academic databases was developed in collaboration with a health sciences librarian. Initial search results produced 25,230 references after duplicates were removed. Eligibility criteria were established to include empirical research of any design that included trans participants or their personal information and that was published in English in peer-reviewed journals. We identified all articles published between 2008 and 2015 with medical photographs of trans participants. For each reference, images were individually numbered in order to track the total number of medical photographs. We used odds ratios (OR) to assess the association between availability of the clinical photograph on Google Images and the following factors: whether the article was openly available online (open access, Researchgate.net, or Academia.edu), whether the article included genital images, if the photographs were published in color, and whether the photographs were located on the journal article landing page. Results We identified 94 articles with medical photographs of trans participants, including a total of 605 photographs. Of the 94 publications, 35 (37%) included at least one medical photograph that was found on Google Images. The ability to locate the article freely online contributes to the availability of at least one image from the article on Google Images (OR 2.99, 95% CI 1.20-7.45). Conclusions This is the first study to document the existence of medical photographs from peer-reviewed journals appearing in Google Images search results. Almost all of the images we searched for included sensitive photographs of patient genitals, chests, or breasts. Given that it is unlikely that patients consented to sharing their personal health information in these ways, this constitutes a risk to patient privacy. Based on the impact of current practices, revisions to informed consent policies and guidelines are required. PMID:29483069
Projection Mapping User Interface for Disabled People
Simutis, Rimvydas; Maskeliūnas, Rytis
2018-01-01
Difficulty in communicating is one of the key challenges for people suffering from severe motor and speech disabilities. Often such person can communicate and interact with the environment only using assistive technologies. This paper presents a multifunctional user interface designed to improve communication efficiency and person independence. The main component of this interface is a projection mapping technique used to highlight objects in the environment. Projection mapping makes it possible to create a natural augmented reality information presentation method. The user interface combines a depth sensor and a projector to create camera-projector system. We provide a detailed description of camera-projector system calibration procedure. The described system performs tabletop object detection and automatic projection mapping. Multiple user input modalities have been integrated into the multifunctional user interface. Such system can be adapted to the needs of people with various disabilities. PMID:29686827
Projection Mapping User Interface for Disabled People.
Gelšvartas, Julius; Simutis, Rimvydas; Maskeliūnas, Rytis
2018-01-01
Difficulty in communicating is one of the key challenges for people suffering from severe motor and speech disabilities. Often such person can communicate and interact with the environment only using assistive technologies. This paper presents a multifunctional user interface designed to improve communication efficiency and person independence. The main component of this interface is a projection mapping technique used to highlight objects in the environment. Projection mapping makes it possible to create a natural augmented reality information presentation method. The user interface combines a depth sensor and a projector to create camera-projector system. We provide a detailed description of camera-projector system calibration procedure. The described system performs tabletop object detection and automatic projection mapping. Multiple user input modalities have been integrated into the multifunctional user interface. Such system can be adapted to the needs of people with various disabilities.
GLIMPSE: Google Glass interface for sensory feedback in myoelectric hand prostheses
NASA Astrophysics Data System (ADS)
Markovic, Marko; Karnal, Hemanth; Graimann, Bernhard; Farina, Dario; Dosen, Strahinja
2017-06-01
Objective. Providing sensory feedback to the user of the prosthesis is an important challenge. The common approach is to use tactile stimulation, which is easy to implement but requires training and has limited information bandwidth. In this study, we propose an alternative approach based on augmented reality. Approach. We have developed the GLIMPSE, a Google Glass application which connects to the prosthesis via a Bluetooth interface and renders the prosthesis states (EMG signals, aperture, force and contact) using augmented reality (see-through display) and sound (bone conduction transducer). The interface was tested in healthy subjects that used the prosthesis with (FB group) and without (NFB group) feedback during a modified clothespins test that allowed us to vary the difficulty of the task. The outcome measures were the number of unsuccessful trials, the time to accomplish the task, and the subjective ratings of the relevance of the feedback. Main results. There was no difference in performance between FB and NFB groups in the case of a simple task (basic, same-color clothespins test), but the feedback significantly improved the performance in a more complex task (pins of different resistances). Importantly, the GLIMPSE feedback did not increase the time to accomplish the task. Therefore, the supplemental feedback might be useful in the tasks which are more demanding, and thereby less likely to benefit from learning and feedforward control. The subjects integrated the supplemental feedback with the intrinsic sources (vision and muscle proprioception), developing their own idiosyncratic strategies to accomplish the task. Significance. The present study demonstrates a novel self-contained, ready-to-deploy, wearable feedback interface. The interface was successfully tested and was proven to be feasible and functionally beneficial. The GLIMPSE can be used as a practical solution but also as a general and flexible instrument to investigate closed-loop prosthesis control.
A web-based tool for groundwater mapping and drought analysis
NASA Astrophysics Data System (ADS)
Christensen, S.; Burns, M.; Jones, N.; Strassberg, G.
2012-12-01
In 2011-2012, the state of Texas saw the worst one-year drought on record. Fluctuations in gravity measured by GRACE satellites indicate that as much as 100 cubic kilometers of water was lost during this period. Much of this came from reservoirs and shallow soil moisture, but a significant amount came from aquifers. In response to this crisis, a Texas Drought Technology Steering Committee (TDTSC) consisting of academics and water managers was formed to develop new tools and strategies to assist the state in monitoring, predicting, and responding to drought events. In this presentation, we describe one of the tools that was developed as part of this effort. When analyzing the impact of drought on groundwater levels, it is fairly common to examine time series data at selected monitoring wells. However, accurately assessing impacts and trends requires both spatial and temporal analysis involving the development of detailed water level maps at various scales. Creating such maps in a flexible and rapid fashion is critical for effective drought analysis, but can be challenging due to the massive amounts of data involved and the processing required to generate such maps. Furthermore, wells are typically not sampled at the same points in time, and so developing a water table map for a particular date requires both spatial and temporal interpolation of water elevations. To address this challenge, a Cloud-based water level mapping system was developed for the state of Texas. The system is based on the Texas Water Development Board (TWDB) groundwater database, but can be adapted to use other databases as well. The system involves a set of ArcGIS workflows running on a server with a web-based front end and a Google Earth plug-in. A temporal interpolation geoprocessing tool was developed to estimate the piezometric heads for all wells in a given region at a specific date using a regression analysis. This interpolation tool is coupled with other geoprocessing tools to filter data and interpolate point elevations spatially to produce water level, drawdown, and depth to groundwater maps. The web interface allows for users to generate these maps at locations and times of interest. A sequence of maps can be generated over a period of time and animated to visualize how water levels are changing. The time series regression analysis can also be used to do short-term predictions of future water levels.
Transport Statistics - Transport - UNECE
Statistics and Data Online Infocards Database SDG Papers E-Road Census Traffic Census Map Traffic Census 2015 available. Two new datasets have been added to the transport statistics database: bus and coach statistics Database Evaluations Follow UNECE Facebook Rss Twitter You tube Contact us Instagram Flickr Google+ Â
An Automated Approach to Extracting River Bank Locations from Aerial Imagery Using Image Texture
2013-01-01
Atchafalaya River, LA. Map Data: Google, United States Department of Agriculture Farm Ser- vice Agency, Europa Technologies AUTOMATED RIVER BANK...traverse morphologically smooth landscapes including rivers in sand or ice . Within these limitations, we hold that this technique rep- resents a valuable
Children Creating Multimodal Stories about a Familiar Environment
ERIC Educational Resources Information Center
Kervin, Lisa; Mantei, Jessica
2017-01-01
Storytelling is a practice that enables children to apply their literacy skills. This article shares a collaborative literacy strategy devised to enable children to create multimodal stories about their familiar school environment. The strategy uses resources, including the children's own drawings, images from Google Maps, and the Puppet Pals…
Spatial data software integration - Merging CAD/CAM/mapping with GIS and image processing
NASA Technical Reports Server (NTRS)
Logan, Thomas L.; Bryant, Nevin A.
1987-01-01
The integration of CAD/CAM/mapping with image processing using geographic information systems (GISs) as the interface is examined. Particular emphasis is given to the development of software interfaces between JPL's Video Image Communication and Retrieval (VICAR)/Imaged Based Information System (IBIS) raster-based GIS and the CAD/CAM/mapping system. The design and functions of the VICAR and IBIS are described. Vector data capture and editing are studied. Various software programs for interfacing between the VICAR/IBIS and CAD/CAM/mapping are presented and analyzed.
NASA Astrophysics Data System (ADS)
Khakpour, Mohammad; Paulik, Christoph; Hahn, Sebastian
2016-04-01
Communication about remote sensing data quality between data providers and users as well as between the users is often difficult. The users have a hard time figuring out if a product has known problems over their region of interest and data providers have to spend a lot of effort to make this information available, if it exists. Scientific publications are one tool for communicating with the users base but they are static and mostly one way. As a data provider it is also often difficult to make feedback, received from users, available to the complete user base. The Geo Issue Tracking System (GeoITS) is an Open Source Web Application which has been developed to mitigate these problems. GeoITS combines a mapping interface (Google Maps) with a simple wiki platform. It allows users to give region specific feedback on a remote sensing product by drawing a polygon on the map and describing the problems they had using the remote sensing product in this area. These geolocated wiki entries are then viewable by other users as well as the data providers which can modify and extend the entries. In this way the conversations between the users and the data provider are no longer hidden in e.g. emails but open for all users of the dataset. This new kind of communication platform can enable better cooperation between users and data providers. It will also provide data providers with the ability to track problems their dataset might have in certain areas and resolve them with new product releases. The source code is available via http://github.com/TUW-GEO/geoits_dev A running instance can be tried at https://geoits.herokuapp.com/
Interactive 3D Mars Visualization
NASA Technical Reports Server (NTRS)
Powell, Mark W.
2012-01-01
The Interactive 3D Mars Visualization system provides high-performance, immersive visualization of satellite and surface vehicle imagery of Mars. The software can be used in mission operations to provide the most accurate position information for the Mars rovers to date. When integrated into the mission data pipeline, this system allows mission planners to view the location of the rover on Mars to 0.01-meter accuracy with respect to satellite imagery, with dynamic updates to incorporate the latest position information. Given this information so early in the planning process, rover drivers are able to plan more accurate drive activities for the rover than ever before, increasing the execution of science activities significantly. Scientifically, this 3D mapping information puts all of the science analyses to date into geologic context on a daily basis instead of weeks or months, as was the norm prior to this contribution. This allows the science planners to judge the efficacy of their previously executed science observations much more efficiently, and achieve greater science return as a result. The Interactive 3D Mars surface view is a Mars terrain browsing software interface that encompasses the entire region of exploration for a Mars surface exploration mission. The view is interactive, allowing the user to pan in any direction by clicking and dragging, or to zoom in or out by scrolling the mouse or touchpad. This set currently includes tools for selecting a point of interest, and a ruler tool for displaying the distance between and positions of two points of interest. The mapping information can be harvested and shared through ubiquitous online mapping tools like Google Mars, NASA WorldWind, and Worldwide Telescope.
Using open-source programs to create a web-based portal for hydrologic information
NASA Astrophysics Data System (ADS)
Kim, H.
2013-12-01
Some hydrologic data sets, such as basin climatology, precipitation, and terrestrial water storage, are not easily obtainable and distributable due to their size and complexity. We present a Hydrologic Information Portal (HIP) that has been implemented at the University of California for Hydrologic Modeling (UCCHM) and that has been organized around the large river basins of North America. This portal can be easily accessed through a modern web browser that enables easy access and visualization of such hydrologic data sets. Some of the main features of our HIP include a set of data visualization features so that users can search, retrieve, analyze, integrate, organize, and map data within large river basins. Recent information technologies such as Google Maps, Tornado (Python asynchronous web server), NumPy/SciPy (Scientific Library for Python) and d3.js (Visualization library for JavaScript) were incorporated into the HIP to create ease in navigating large data sets. With such open source libraries, HIP can give public users a way to combine and explore various data sets by generating multiple chart types (Line, Bar, Pie, Scatter plot) directly from the Google Maps viewport. Every rendered object such as a basin shape on the viewport is clickable, and this is the first step to access the visualization of data sets.
First Prototype of a Web Map Interface for ESA's Planetary Science Archive (PSA)
NASA Astrophysics Data System (ADS)
Manaud, N.; Gonzalez, J.
2014-04-01
We present a first prototype of a Web Map Interface that will serve as a proof of concept and design for ESA's future fully web-based Planetary Science Archive (PSA) User Interface. The PSA is ESA's planetary science archiving authority and central repository for all scientific and engineering data returned by ESA's Solar System missions [1]. All data are compliant with NASA's Planetary Data System (PDS) Standards and are accessible through several interfaces [2]: in addition to serving all public data via FTP and the Planetary Data Access Protocol (PDAP), a Java-based User Interface provides advanced search, preview, download, notification and delivery-basket functionality. It allows the user to query and visualise instrument observations footprints using a map-based interface (currently only available for Mars Express HRSC and OMEGA instruments). During the last decade, the planetary mapping science community has increasingly been adopting Geographic Information System (GIS) tools and standards, originally developed for and used in Earth science. There is an ongoing effort to produce and share cartographic products through Open Geospatial Consortium (OGC) Web Services, or as standalone data sets, so that they can be readily used in existing GIS applications [3,4,5]. Previous studies conducted at ESAC [6,7] have helped identify the needs of Planetary GIS users, and define key areas of improvement for the future Web PSA User Interface. Its web map interface shall will provide access to the full geospatial content of the PSA, including (1) observation geometry footprints of all remote sensing instruments, and (2) all georeferenced cartographic products, such as HRSC map-projected data or OMEGA global maps from Mars Express. It shall aim to provide a rich user experience for search and visualisation of this content using modern and interactive web mapping technology. A comprehensive set of built-in context maps from external sources, such as MOLA topography, TES infrared maps or planetary surface nomenclature, provided in both simple cylindrical and polar stereographic projections, shall enhance this user experience. In addition, users should be able to import and export data in commonly used open- GIS formats. It is also intended to serve all PSA geospatial data through OGC-compliant Web Services so that they can be captured, visualised and analysed directly from GIS software, along with data from other sources. The following figure illustrates how the PSA web map interface and services shall fit in a typical Planetary GIS user working environment.
Understanding Urban Watersheds through Digital Interactive Maps, San Francisco Bay Area, California
NASA Astrophysics Data System (ADS)
Sowers, J. M.; Ticci, M. G.; Mulvey, P.
2014-12-01
Dense urbanization has resulted in the "disappearance" of many local creeks in urbanized areas surrounding the San Francisco Bay. Long reaches of creeks now flow in underground pipes. Municipalities and water agencies trying to reduce non-point-source pollution are faced with a public that cannot see and therefore does not understand the interconnected nature of the drainage system or its ultimate discharge to the bay. Since 1993, we have collaborated with the Oakland Museum, the San Francisco Estuary Institute, public agencies, and municipalities to create creek and watershed maps to address the need for public understanding of watershed concepts. Fifteen paper maps are now published (www.museumca.org/creeks), which have become a standard reference for educators and anyone working on local creek-related issues. We now present digital interactive creek and watershed maps in Google Earth. Four maps are completed covering urbanized areas of Santa Clara and Alameda Counties. The maps provide a 3D visualization of the watersheds, with cartography draped over the landscape in transparent colors. Each mapped area includes both Present and Past (circa 1800s) layers which can be clicked on or off by the user. The Present layers include the modern drainage network, watershed boundaries, and reservoirs. The Past layers include the 1800s-era creek systems, tidal marshes, lagoons, and other habitats. All data are developed in ArcGIS software and converted to Google Earth format. To ensure the maps are interesting and engaging, clickable icons pop-up provide information on places to visit, restoration projects, history, plants, and animals. Maps of Santa Clara Valley are available at http://www.valleywater.org/WOW.aspx. Maps of western Alameda County will soon be available at http://acfloodcontrol.org/. Digital interactive maps provide several advantages over paper maps. They are seamless within each map area, and the user can zoom in or out, and tilt, and fly over to explore any area of interest. They can be easily customized, for example, adding placemarks or notes. Enrichment information can be added, using clickable icons, without cluttering the map. Best, the maps are fun to use. Digital interactive maps will be another effective tool for enhancing public understanding of urban creeks & watersheds.
Helioviewer: A Web 2.0 Tool for Visualizing Heterogeneous Heliophysics Data
NASA Astrophysics Data System (ADS)
Hughitt, V. K.; Ireland, J.; Lynch, M. J.; Schmeidel, P.; Dimitoglou, G.; Müeller, D.; Fleck, B.
2008-12-01
Solar physics datasets are becoming larger, richer, more numerous and more distributed. Feature/event catalogs (describing objects of interest in the original data) are becoming important tools in navigating these data. In the wake of this increasing influx of data and catalogs there has been a growing need for highly sophisticated tools for accessing and visualizing this wealth of information. Helioviewer is a novel tool for integrating and visualizing disparate sources of solar and Heliophysics data. Taking advantage of the newly available power of modern web application frameworks, Helioviewer merges image and feature catalog data, and provides for Heliophysics data a familiar interface not unlike Google Maps or MapQuest. In addition to streamlining the process of combining heterogeneous Heliophysics datatypes such as full-disk images and coronagraphs, the inclusion of visual representations of automated and human-annotated features provides the user with an integrated and intuitive view of how different factors may be interacting on the Sun. Currently, Helioviewer offers images from The Extreme ultraviolet Imaging Telescope (EIT), The Large Angle and Spectrometric COronagraph experiment (LASCO) and the Michelson Doppler Imager (MDI) instruments onboard The Solar and Heliospheric Observatory (SOHO), as well as The Transition Region and Coronal Explorer (TRACE). Helioviewer also incorporates feature/event information from the LASCO CME List, NOAA Active Regions, CACTus CME and Type II Radio Bursts feature/event catalogs. The project is undergoing continuous development with many more data sources and additional functionality planned for the near future.
NASA Astrophysics Data System (ADS)
Piman, T.; Schellekens, J.; Haag, A.; Donchyts, G.; Apirumanekul, C.; Hlaing, K. T.
2017-12-01
River morphology changes is one of the key issues in Ayeyarwady River in Myanmar which cause impacts on navigation, riverine habitats, agriculture lands, communities and livelihoods near the bank of the river. This study is aimed to track the changes in river morphology in the middle reach of Ayeyarwady River over last 30 years from 1984-2014 to improve understanding of riverbank dynamic, erosion and deposition procress. Earth observations including LandSat-7, LandSat-8, Digital Elevation Model from SRTM Plus and, ASTER-2 GoogleMap and Open Street Map were obtained for the study. GIS and remote sensing tools were used to analyze changes in river morphology while surface water mapping tool was applied to determine how the dynamic behaviour of the surface river and effect of river morphology changes. The tool consists of two components: (1) a Google Earth Engine (GEE) javascript or python application that performs image analysis and (2) a user-friendly site/app using Google's appspot.com that exposes the application to the users. The results of this study shown that the fluvial morphology in the middle reach of Ayeyarwady River is continuously changing under the influence of high water flows in particularly from extreme flood events and land use change from mining and deforestation. It was observed that some meandering sections of the riverbank were straightened, which results in the movement of sediment downstream and created new sections of meandering riverbank. Several large islands have formed due to the stabilization by vegetation and is enforced by sedimentation while many small bars were formed and migrated dynamically due to changes in water levels and flow velocity in the wet and dry seasons. The main channel was changed to secondary channel in some sections of the river. This results a constant shift of the navigation route. We also found that some villages were facing riverbank erosion which can force villagers to relocate. The study results demonstrated that the products from earth observations and the surface water mapping tool could detect dynamic changes of river morphology in the Ayeyarwady River. This information is useful to support navigation and riverbank protection planning and formulating mitigation measures for local communities that are affecting by riverbank erosion.
Smart "geomorphological" map browsing - a tale about geomorphological maps and the internet
NASA Astrophysics Data System (ADS)
Geilhausen, M.; Otto, J.-C.
2012-04-01
With the digital production of geomorphological maps, the dissemination of research outputs now extends beyond simple paper products. Internet technologies can contribute to both, the dissemination of geomorphological maps and access to geomorphologic data and help to make geomorphological knowledge available to a greater public. Indeed, many national geological surveys employ end-to-end digital workflows from data capture in the field to final map production and dissemination. This paper deals with the potential of web mapping applications and interactive, portable georeferenced PDF maps for the distribution of geomorphological information. Web mapping applications such as Google Maps have become very popular and widespread and increased the interest and access to mapping. They link the Internet with GIS technology and are a common way of presenting dynamic maps online. The GIS processing is performed online and maps are visualised in interactive web viewers characterised by different capabilities such as zooming, panning or adding further thematic layers, with the map refreshed after each task. Depending on the system architecture and the components used, advanced symbology, map overlays from different applications and sources and their integration into a Desktop GIS are possible. This interoperability is achieved through the use of international open standards that include mechanisms for the integration and visualisation of information from multiple sources. The portable document format (PDF) is commonly used for printing and is a standard format that can be processed by many graphic software and printers without loss of information. A GeoPDF enables the sharing of geospatial maps and data in PDF documents. Multiple, independent map frames with individual spatial reference systems are possible within a GeoPDF, for example, for map overlays or insets. Geospatial functionality of a GeoPDF includes scalable map display, layer visibility control, access to attribute data, coordinate queries and spatial measurements. The full functionality of GeoPDFs requires free and user-friendly plug-ins for PDF readers and GIS software. A GeoPDF enables fundamental GIS functionality turning the formerly static PDF map into an interactive, portable georeferenced PDF map. GeoPDFs are easy to create and provide an interesting and valuable way to disseminate geomorphological maps. Our motivation to engage with the online distribution of geomorphological maps originates in the increasing number of web mapping applications available today indicating that the Internet has become a medium for displaying geographical information in rich forms and user-friendly interfaces. So, why not use the Internet to distribute geomorphological maps and enhance their practical application? Web mapping and dynamic PDF maps can play a key role in the movement towards a global dissemination of geomorphological information. This will be exemplified by live demonstrations of i.) existing geomorphological WebGIS applications, ii.) data merging from various sources using web map services, and iii.) free to download GeoPDF maps during the presentations.
NASA Astrophysics Data System (ADS)
De Paor, D. G.; Bailey, J. E.; Whitmeyer, S. J.
2012-12-01
Our TUES research centers on the role of digital data, visualizations, animations, and simulations in undergraduate geoscience education. Digital hardware (smartphones, tablets, GPSs, GigaPan robotic camera mounts, etc.) are revolutionizing field data collection. Software products (GIS, 3-D scanning and modeling programs, virtual globes, etc.) have truly transformed the way geoscientists teach, learn, and do research. Whilst Google-Earth-style visualizations are famously user-friend for the person browsing, they can be notoriously unfriendly for the content creator. Therefore, we developed tools to help educators create and share visualizations as easily as if posting on Facebook. Anyone whoIf you wish to display geological cross sections on Google Earth, go to digitalplanet.org, upload image files, position them on a line of section, and share with the world through our KMZ hosting service. Other tools facilitate screen overlay and 3-D map symbol generation. We advocate use of such technology to enable undergraduate students to 'publish' their first mapping efforts even while they are working in the field. A second outcome of our TUES projects merges Second-Life-style interaction with Google Earth. We created games in which students act as first responders for natural hazard mitigation, prospectors for natural resource explorations, and structural geologist for map-making. Students are represented by avatars and collaborate by exchange of text messages - the natural mode of communication for the current generation. Teachers view logs showing student movements as well as transcripts of text messages and can scaffold student learning and geofence students to prevent wandering. Early results of in-class testing show positive learning outcomes. The third aspect of our program emphasizes dissemination. Experience shows that great effort is required to overcome activation energy and ensure adoption of new technology into the curriculum. We organized a GSA Penrose Conference, a GSA Pardee Keynote Symposium, and AGU Townhall Meeting, and numerous workshops at annual and regional meetings, and set up a web site dedicated to dissemination of program products. Future plans include development of augmented reality teaching resources, hosting of community mapping services, and creation of a truly 4-D virtual globe.;
Improving Land Cover Mapping: a Mobile Application Based on ESA Sentinel 2 Imagery
NASA Astrophysics Data System (ADS)
Melis, M. T.; Dessì, F.; Loddo, P.; La Mantia, C.; Da Pelo, S.; Deflorio, A. M.; Ghiglieri, G.; Hailu, B. T.; Kalegele, K.; Mwasi, B. N.
2018-04-01
The increasing availability of satellite data is a real value for the enhancement of environmental knowledge and land management. Possibilities to integrate different source of geo-data are growing and methodologies to create thematic database are becoming very sophisticated. Moreover, the access to internet services and, in particular, to web mapping services is well developed and spread either between expert users than the citizens. Web map services, like Google Maps or Open Street Maps, give the access to updated optical imagery or topographic maps but information on land cover/use - are not still provided. Therefore, there are many failings in the general utilization -non-specialized users- and access to those maps. This issue is particularly felt where the digital (web) maps could form the basis for land use management as they are more economic and accessible than the paper maps. These conditions are well known in many African countries where, while the internet access is becoming open to all, the local map agencies and their products are not widespread.
A global map of rainfed cropland areas (GMRCA) at the end of last millennium using remote sensing
Biradar, C.M.; Thenkabail, P.S.; Noojipady, P.; Li, Y.; Dheeravath, V.; Turral, H.; Velpuri, M.; Gumma, M.K.; Gangalakunta, O.R.P.; Cai, X.L.; Xiao, X.; Schull, M.A.; Alankara, R.D.; Gunasinghe, S.; Mohideen, S.
2009-01-01
The overarching goal of this study was to produce a global map of rainfed cropland areas (GMRCA) and calculate country-by-country rainfed area statistics using remote sensing data. A suite of spatial datasets, methods and protocols for mapping GMRCA were described. These consist of: (a) data fusion and composition of multi-resolution time-series mega-file data-cube (MFDC), (b) image segmentation based on precipitation, temperature, and elevation zones, (c) spectral correlation similarity (SCS), (d) protocols for class identification and labeling through uses of SCS R2-values, bi-spectral plots, space-time spiral curves (ST-SCs), rich source of field-plot data, and zoom-in-views of Google Earth (GE), and (e) techniques for resolving mixed classes by decision tree algorithms, and spatial modeling. The outcome was a 9-class GMRCA from which country-by-country rainfed area statistics were computed for the end of the last millennium. The global rainfed cropland area estimate from the GMRCA 9-class map was 1.13 billion hectares (Bha). The total global cropland areas (rainfed plus irrigated) was 1.53 Bha which was close to national statistics compiled by FAOSTAT (1.51 Bha). The accuracies and errors of GMRCA were assessed using field-plot and Google Earth data points. The accuracy varied between 92 and 98% with kappa value of about 0.76, errors of omission of 2-8%, and the errors of commission of 19-36%. ?? 2008 Elsevier B.V.
Planetary-Scale Geospatial Data Analysis Techniques in Google's Earth Engine Platform (Invited)
NASA Astrophysics Data System (ADS)
Hancher, M.
2013-12-01
Geoscientists have more and more access to new tools for large-scale computing. With any tool, some tasks are easy and other tasks hard. It is natural to look to new computing platforms to increase the scale and efficiency of existing techniques, but there is a more exiting opportunity to discover and develop a new vocabulary of fundamental analysis idioms that are made easy and effective by these new tools. Google's Earth Engine platform is a cloud computing environment for earth data analysis that combines a public data catalog with a large-scale computational facility optimized for parallel processing of geospatial data. The data catalog includes a nearly complete archive of scenes from Landsat 4, 5, 7, and 8 that have been processed by the USGS, as well as a wide variety of other remotely-sensed and ancillary data products. Earth Engine supports a just-in-time computation model that enables real-time preview during algorithm development and debugging as well as during experimental data analysis and open-ended data exploration. Data processing operations are performed in parallel across many computers in Google's datacenters. The platform automatically handles many traditionally-onerous data management tasks, such as data format conversion, reprojection, resampling, and associating image metadata with pixel data. Early applications of Earth Engine have included the development of Google's global cloud-free fifteen-meter base map and global multi-decadal time-lapse animations, as well as numerous large and small experimental analyses by scientists from a range of academic, government, and non-governmental institutions, working in a wide variety of application areas including forestry, agriculture, urban mapping, and species habitat modeling. Patterns in the successes and failures of these early efforts have begun to emerge, sketching the outlines of a new set of simple and effective approaches to geospatial data analysis.
Wildland-urban interface maps vary with purpose and context
Susan I. Stewart; Bo Wilmer; Roger B. Hammer; Gregory H. Aplet; Todd J. Hawbaker; Carol Miller; Volker C. Radeloff
2009-01-01
Maps of the wildland-urban interface (WUI) are both policy tools and powerful visual images. Although the growing number of WUI maps serve similar purposes, this article indicates that WUI maps derived from the same data sets can differ in important ways related to their original intended application. We discuss the use of ancillary data in modifying census data to...
UNAVCO Software and Services for Visualization and Exploration of Geoscience Data
NASA Astrophysics Data System (ADS)
Meertens, C.; Wier, S.
2007-12-01
UNAVCO has been involved in visualization of geoscience data to support education and research for several years. An early and ongoing service is the Jules Verne Voyager, a web browser applet built on the GMT that displays any area on Earth, with many data set choices, including maps, satellite images, topography, geoid heights, sea-floor ages, strain rates, political boundaries, rivers and lakes, earthquake and volcano locations, focal mechanisms, stress axes, and observed and modeled plate motion and deformation velocity vectors from geodetic measurements around the world. As part of the GEON project, UNAVCO has developed the GEON IDV, a research-level, 4D (earth location, depth and/or altitude, and time), Java application for interactive display and analysis of geoscience data. The GEON IDV is designed to meet the challenge of investigating complex, multi-variate, time-varying, three-dimensional geoscience data anywhere on earth. The GEON IDV supports simultaneous displays of data sets from differing sources, with complete control over colors, time animation, map projection, map area, point of view, and vertical scale. The GEON IDV displays gridded and point data, images, GIS shape files, and several other types of data. The GEON IDV has symbols and displays for GPS velocity vectors, seismic tomography, earthquake focal mechanisms, earthquake locations with magnitude or depth, seismic ray paths in 3D, seismic anisotropy, convection model visualization, earth strain axes and strain field imagery, and high-resolution 3D topographic relief maps. Multiple data sources and display types may appear in one view. As an example of GEON IDV utility, it can display hypocenters under a volcano, a surface geology map of the volcano draped over 3D topographic relief, town locations and political boundaries, and real-time 3D weather radar clouds of volcanic ash in the atmosphere, with time animation. The GEON IDV can drive a GeoWall or other 3D stereo system. IDV output includes imagery, movies, and KML files for Google Earth use of IDV static images, where Google Earth can handle the display. The IDV can be scripted to create display images on user request or automatically on data arrival, offering the use of the IDV as a back end to support a data web site. We plan to extend the power of the IDV by accepting new data types and data services, such as GeoSciML. An active program of online and video training in GEON IDV use is planned. UNAVCO will support users who need assistance converting their data to the standard formats used by the GEON IDV. The UNAVCO Facility provides web-accessible support for Google Earth and Google Maps display of any of more than 9500 GPS stations and survey points, including metadata for each installation. UNAVCO provides corresponding Open Geospatial Consortium (OGC) web services with the same data. UNAVCO's goal is to facilitate data access, interoperability, and efficient searches, exploration, and use of data by promoting web services, standards for GEON IDV data formats and metadata, and software able to simultaneously read and display multiple data sources, formats, and map locations or projections. Retention and propagation of semantics and metadata with observational and experimental values is essential for interoperability and understanding diverse data sources.
Near real-time qualitative monitoring of lake water chlorophyll globally using GoogleEarth Engine
NASA Astrophysics Data System (ADS)
Zlinszky, András; Supan, Peter; Koma, Zsófia
2017-04-01
Monitoring ocean chlorophyll and suspended sediment has been made possible using optical satellite imaging, and has contributed immensely to our understanding of the Earth and its climate. However, lake water quality monitoring has limitations due to the optical complexity of shallow, sediment- and organic matter-laden waters. Meanwhile, timely and detailed information on basic lake water quality parameters would be essential for sustainable management of inland waters. Satellite-based remote sensing can deliver area-covering, high resolution maps of basic lake water quality parameters, but scientific application of these datasets for lake monitoring has been hindered by limitations to calibration and accuracy evaluation, and therefore access to such data has been the privilege of scientific users. Nevertheless, since for many inland waters satellite imaging is the only source of monitoring data, we believe it is urgent to make map products of chlorophyll and suspended sediment concentrations available to a wide range of users. Even if absolute accuracy can not be validated, patterns, processes and qualitative information delivered by such datasets in near-real time can act as an early warning system, raise awareness to water quality processes and serve education, in addition to complementing local monitoring activities. By making these datasets openly available on the internet through an easy to use framework, dialogue between stakeholders, management and governance authorities can be facilitated. We use GoogleEarthEngine to access and process archive and current satellite data. GoogleEarth Engine is a development and visualization framework that provides access to satellite datasets and processing capacity for analysis at the Petabyte scale. Based on earlier investigations, we chose the fluorescence line height index to represent water chlorophyll concentration. This index relies on the chlorophyll fluorescence peak at 680 nm, and has been tested for open ocean but also inland lake situations for MODIS and MERIS satellite sensor data. In addition to being relatively robust and less sensitive to atmospheric influence, this algorithm is also very simple, being based on the height of the 680 nm peak above the linear interpolation of the two neighbouring bands. However, not all satellite datasets suitable for FLH are catalogued for GoogleEarth Engine. In the current testing phase, Landsat 7, Landsat 8 (30 m resolution), and Sentinel 2 (20 m) are being tested. Landsat 7 has suitable band configuration, but has a strip error due to a sensor problem. Landsat 8 and Sentinel 2 lack a single spectral optimal for FLH. Sentinel 3 would be an optimal data source and has shown good performace during small-scale initial tests, but is not distributed globally for GoogleEarth Engine. In addition to FLH data from these satellites, our system delivers cloud and ice masking, qualitative suspended sediment data (based on the band closest to 600 nm) and true colour images, all within an easy-to-use Google Maps background. This allows on-demand understanding and interpretation of water quality patterns and processes in near real time. While the system is still under development, we believe it could significantly contribute to lake water quality management and monitoring worldwide.
The Dimensions of the Solar System
ERIC Educational Resources Information Center
Schneider, Stephen E.; Davis, Kathleen S.
2007-01-01
A few new wrinkles have been added to the popular activity of building a scale model of the solar system. Students can learn about maps and scaling using easily accessible online resources that include satellite images. This is accomplished by taking advantage of some of the special features of Google Earth. This activity gives students a much…
Geospatial Services in Special Libraries: A Needs Assessment Perspective
ERIC Educational Resources Information Center
Barnes, Ilana
2013-01-01
Once limited to geographers and mapmakers, Geographic Information Systems (GIS) has taken a growing central role in information management and visualization. Geospatial services run a gamut of different products and services from Google maps to ArcGIS servers to Mobile development. Geospatial services are not new. Libraries have been writing about…
Map Scale, Proportion, and Google[TM] Earth
ERIC Educational Resources Information Center
Roberge, Martin C.; Cooper, Linda L.
2010-01-01
Aerial imagery has a great capacity to engage and maintain student interest while providing a contextual setting to strengthen their ability to reason proportionally. Free, on-demand, high-resolution, large-scale aerial photography provides both a bird's eye view of the world and a new perspective on one's own community. This article presents an…
Midekisa, Alemayehu; Holl, Felix; Savory, David J; Andrade-Pacheco, Ricardo; Gething, Peter W; Bennett, Adam; Sturrock, Hugh J W
2017-01-01
Quantifying and monitoring the spatial and temporal dynamics of the global land cover is critical for better understanding many of the Earth's land surface processes. However, the lack of regularly updated, continental-scale, and high spatial resolution (30 m) land cover data limit our ability to better understand the spatial extent and the temporal dynamics of land surface changes. Despite the free availability of high spatial resolution Landsat satellite data, continental-scale land cover mapping using high resolution Landsat satellite data was not feasible until now due to the need for high-performance computing to store, process, and analyze this large volume of high resolution satellite data. In this study, we present an approach to quantify continental land cover and impervious surface changes over a long period of time (15 years) using high resolution Landsat satellite observations and Google Earth Engine cloud computing platform. The approach applied here to overcome the computational challenges of handling big earth observation data by using cloud computing can help scientists and practitioners who lack high-performance computational resources.
Detecting Potential Water Quality Issues by Mapping Trophic Status Using Google Earth Engine
NASA Astrophysics Data System (ADS)
Nguy-Robertson, A. L.; Harvey, K.; Huening, V.; Robinson, H.
2017-12-01
The identification, timing, and spatial distribution of recurrent algal blooms and aquatic vegetation can help water managers and policy makers make better water resource decisions. In many parts of the world there is little monitoring or reporting of water quality due to the required costs and effort to collect and process water samples. We propose to use Google Earth Engine to quickly identify the recurrence of trophic states in global inland water systems. Utilizing Landsat and Sentinel multispectral imagery, inland water quality parameters (i.e. chlorophyll a concentration) can be estimated and waters can be classified by trophic state; oligotrophic, mesotrophic, eutrophic, and hypereutrophic. The recurrence of eutrophic and hypereutrophic observations can highlight potentially problematic locations where algal blooms or aquatic vegetation occur routinely. Eutrophic and hypereutrophic waters commonly include many harmful algal blooms and waters prone to fish die-offs from hypoxia. While these maps may be limited by the accuracy of the algorithms utilized to estimate chlorophyll a; relative comparisons at a local scale can help water managers to focus limited resources.
Holl, Felix; Savory, David J.; Andrade-Pacheco, Ricardo; Gething, Peter W.; Bennett, Adam; Sturrock, Hugh J. W.
2017-01-01
Quantifying and monitoring the spatial and temporal dynamics of the global land cover is critical for better understanding many of the Earth’s land surface processes. However, the lack of regularly updated, continental-scale, and high spatial resolution (30 m) land cover data limit our ability to better understand the spatial extent and the temporal dynamics of land surface changes. Despite the free availability of high spatial resolution Landsat satellite data, continental-scale land cover mapping using high resolution Landsat satellite data was not feasible until now due to the need for high-performance computing to store, process, and analyze this large volume of high resolution satellite data. In this study, we present an approach to quantify continental land cover and impervious surface changes over a long period of time (15 years) using high resolution Landsat satellite observations and Google Earth Engine cloud computing platform. The approach applied here to overcome the computational challenges of handling big earth observation data by using cloud computing can help scientists and practitioners who lack high-performance computational resources. PMID:28953943
Shultz, Mary
2006-01-01
Introduction: Given the common use of acronyms and initialisms in the health sciences, searchers may be entering these abbreviated terms rather than full phrases when searching online systems. The purpose of this study is to evaluate how various MEDLINE Medical Subject Headings (MeSH) interfaces map acronyms and initialisms to the MeSH vocabulary. Methods: The interfaces used in this study were: the PubMed MeSH database, the PubMed Automatic Term Mapping feature, the NLM Gateway Term Finder, and Ovid MEDLINE. Acronyms and initialisms were randomly selected from 2 print sources. The test data set included 415 randomly selected acronyms and initialisms whose related meanings were found to be MeSH terms. Each acronym and initialism was entered into each MEDLINE MeSH interface to determine if it mapped to the corresponding MeSH term. Separately, 46 commonly used acronyms and initialisms were tested. Results: While performance differed widely, the success rates were low across all interfaces for the randomly selected terms. The common acronyms and initialisms tested at higher success rates across the interfaces, but the differences between the interfaces remained. Conclusion: Online interfaces do not always map medical acronyms and initialisms to their corresponding MeSH phrases. This may lead to inaccurate results and missed information if acronyms and initialisms are used in search strategies. PMID:17082832
A landslide susceptibility map of Africa
NASA Astrophysics Data System (ADS)
Broeckx, Jente; Vanmaercke, Matthias; Duchateau, Rica; Poesen, Jean
2017-04-01
Studies on landslide risks and fatalities indicate that landslides are a global threat to humans, infrastructure and the environment, certainly in Africa. Nonetheless our understanding of the spatial patterns of landslides and rockfalls on this continent is very limited. Also in global landslide susceptibility maps, Africa is mostly underrepresented in the inventories used to construct these maps. As a result, predicted landslide susceptibilities remain subject to very large uncertainties. This research aims to produce a first continent-wide landslide susceptibility map for Africa, calibrated with a well-distributed landslide dataset. As a first step, we compiled all available landslide inventories for Africa. This data was supplemented by additional landslide mapping with Google Earth in underrepresented regions. This way, we compiled 60 landslide inventories from the literature (ca. 11000 landslides) and an additional 6500 landslides through mapping in Google Earth (including 1500 rockfalls). Various environmental variables such as slope, lithology, soil characteristics, land use, precipitation and seismic activity, were investigated for their significance in explaining the observed spatial patterns of landslides. To account for potential mapping biases in our dataset, we used Monte Carlo simulations that selected different subsets of mapped landslides, tested the significance of the considered environmental variables and evaluated the performance of the fitted multiple logistic regression model against another subset of mapped landslides. Based on these analyses, we constructed two landslide susceptibility maps for Africa: one for all landslide types and one excluding rockfalls. In both maps, topography, lithology and seismic activity were the most significant variables. The latter factor may be surprising, given the overall limited degree of seismicity in Africa. However, its significance indicates that frequent seismic events may serve as in important preparatory factor for landslides. This finding concurs with several other recent studies. Rainfall explains a significant, but limited part of the observed landslide pattern and becomes insignificant when also rockfalls are considered. This may be explained by the fact that a significant fraction of the mapped rockfalls occurred in the Sahara desert. Overall, both maps perform well in predicting intra-continental patterns of mass movements in Africa and explain about 80% of the observed variance in landslide occurrence. As a result, these maps may be a valuable tool for planning and risk reduction strategies.
Factors influencing delivered mean airway pressure during nasal CPAP with the RAM cannula.
Gerdes, Jeffrey S; Sivieri, Emidio M; Abbasi, Soraya
2016-01-01
To measure mean airway pressure (MAP) delivered through the RAM Cannula® when used with a ventilator in CPAP mode as a function of percent nares occlusion in a simulated nasal interface/test lung model and to compare the results to MAPs using a nasal continuous positive airway pressure (NCPAP) interface with nares fully occluded. An artificial airway model was connected to a spontaneous breathing lung model in which MAP was measured at set NCPAP levels between 4 and 8 cmH2 O provided by a Dräger Evita XL® ventilator and delivered through three sizes of RAM cannulae. Measurements were performed with varying leakage at the nasal interface by decreasing occlusion from 100% to 29%, half-way prong insertion, and simulated mouth leakage. Comparison measurements were made using the Dräger BabyFlow® NCPAP interface with a full nasal seal. With simulated mouth closed, the Dräger interface delivered MAPs within 0.5 cmH2 O of set CPAP levels. For the RAM cannula, with 60-80% nares occlusion, overall delivered MAPs were 60 ± 17% less than set CPAP levels (P < 0.001). Further, MAP decreased progressively with decreasing percent nares occlusion. The simulated open mouth condition resulted in significantly lower MAPs to <1.7 cmH2 O. The one-half prong insertion depth condition, with closed mouth, yielded MAPs approximately 35 ± 9% less than full insertion pressures (P < 0.001). In our bench tests, the RAM interface connected to a ventilator in NCPAP mode failed to deliver set CPAP levels when applied using the manufacturer recommended 60-80% nares occlusion, even with closed mouth and full nasal prong insertion conditions. © 2015 Wiley Periodicals, Inc.
NASA Astrophysics Data System (ADS)
Guion, A., Jr.; Hodgkins, H.
2015-12-01
The Center of Excellence in Remote Sensing Education and Research (CERSER) has implemented three research projects during the summer Research Experience for Undergraduates (REU) program gathering water quality data for local waterways. The data has been compiled manually utilizing pen and paper and then entered into a spreadsheet. With the spread of electronic devices capable of interacting with databases, the development of an electronic method of entering and manipulating the water quality data was pursued during this project. This project focused on the development of an interactive database to gather, display, and analyze data collected from local waterways. The database and entry form was built in MySQL on a PHP server allowing participants to enter data from anywhere Internet access is available. This project then researched applying this data to the Google Maps site to provide labeling and information to users. The NIA server at http://nia.ecsu.edu is used to host the application for download and for storage of the databases. Water Quality Database Team members included the authors plus Derek Morris Jr., Kathryne Burton and Mr. Jeff Wood as mentor.
Environmental Information Management For Data Discovery and Access System
NASA Astrophysics Data System (ADS)
Giriprakash, P.
2011-01-01
Mercury is a federated metadata harvesting, search and retrieval tool based on both open source software and software developed at Oak Ridge National Laboratory. It was originally developed for NASA, and the Mercury development consortium now includes funding from NASA, USGS, and DOE. A major new version of Mercury was developed during 2007 and released in early 2008. This new version provides orders of magnitude improvements in search speed, support for additional metadata formats, integration with Google Maps for spatial queries, support for RSS delivery of search results, and ready customization to meet the needs of the multiple projects which use Mercury. For the end users, Mercury provides a single portal to very quickly search for data and information contained in disparate data management systems. It collects metadata and key data from contributing project servers distributed around the world and builds a centralized index. The Mercury search interfaces then allow ! the users to perform simple, fielded, spatial and temporal searches across these metadata sources. This centralized repository of metadata with distributed data sources provides extremely fast search results to the user, while allowing data providers to advertise the availability of their data and maintain complete control and ownership of that data.
OntoCAT -- simple ontology search and integration in Java, R and REST/JavaScript
2011-01-01
Background Ontologies have become an essential asset in the bioinformatics toolbox and a number of ontology access resources are now available, for example, the EBI Ontology Lookup Service (OLS) and the NCBO BioPortal. However, these resources differ substantially in mode, ease of access, and ontology content. This makes it relatively difficult to access each ontology source separately, map their contents to research data, and much of this effort is being replicated across different research groups. Results OntoCAT provides a seamless programming interface to query heterogeneous ontology resources including OLS and BioPortal, as well as user-specified local OWL and OBO files. Each resource is wrapped behind easy to learn Java, Bioconductor/R and REST web service commands enabling reuse and integration of ontology software efforts despite variation in technologies. It is also available as a stand-alone MOLGENIS database and a Google App Engine application. Conclusions OntoCAT provides a robust, configurable solution for accessing ontology terms specified locally and from remote services, is available as a stand-alone tool and has been tested thoroughly in the ArrayExpress, MOLGENIS, EFO and Gen2Phen phenotype use cases. Availability http://www.ontocat.org PMID:21619703
OntoCAT--simple ontology search and integration in Java, R and REST/JavaScript.
Adamusiak, Tomasz; Burdett, Tony; Kurbatova, Natalja; Joeri van der Velde, K; Abeygunawardena, Niran; Antonakaki, Despoina; Kapushesky, Misha; Parkinson, Helen; Swertz, Morris A
2011-05-29
Ontologies have become an essential asset in the bioinformatics toolbox and a number of ontology access resources are now available, for example, the EBI Ontology Lookup Service (OLS) and the NCBO BioPortal. However, these resources differ substantially in mode, ease of access, and ontology content. This makes it relatively difficult to access each ontology source separately, map their contents to research data, and much of this effort is being replicated across different research groups. OntoCAT provides a seamless programming interface to query heterogeneous ontology resources including OLS and BioPortal, as well as user-specified local OWL and OBO files. Each resource is wrapped behind easy to learn Java, Bioconductor/R and REST web service commands enabling reuse and integration of ontology software efforts despite variation in technologies. It is also available as a stand-alone MOLGENIS database and a Google App Engine application. OntoCAT provides a robust, configurable solution for accessing ontology terms specified locally and from remote services, is available as a stand-alone tool and has been tested thoroughly in the ArrayExpress, MOLGENIS, EFO and Gen2Phen phenotype use cases. http://www.ontocat.org.
Snake River Plain Geothermal Play Fairway Analysis - Phase 1 KMZ files
John Shervais
2015-10-10
This dataset contain raw data files in kmz files (Google Earth georeference format). These files include volcanic vent locations and age, the distribution of fine-grained lacustrine sediments (which act as both a seal and an insulating layer for hydrothermal fluids), and post-Miocene faults compiled from the Idaho Geological Survey, the USGS Quaternary Fault database, and unpublished mapping. It also contains the Composite Common Risk Segment Map created during Phase 1 studies, as well as a file with locations of select deep wells used to interrogate the subsurface.
Combating Conflict Related Sexual Violence: More Than a Stability Concern
2014-06-13
violence can cause serious bodily harm or mental harm to members of the group (International Criminal Court 2002, 3; Ellis 2007). Under crimes against...population was subjugated to Japanese 46 rule and were provided with horrific visual, physical, and emotional reminders of the futility of any...maps.google.com/ maps /ms?msid=214870171076954118166.0004b9bcb533b0ee2c1f8&msa=0&ie= UTF8&ll=12.848235,58.136902&spn=43.135476,135.258178&t=m&output=em bed
NASA Astrophysics Data System (ADS)
Li, P.; Turk, J.; Vu, Q.; Knosp, B.; Hristova-Veleva, S. M.; Lambrigtsen, B.; Poulsen, W. L.; Licata, S.
2009-12-01
NASA is planning a new field experiment, the Genesis and Rapid Intensification Processes (GRIP), in the summer of 2010 to better understand how tropical storms form and develop into major hurricanes. The DC-8 aircraft and the Global Hawk Unmanned Airborne System (UAS) will be deployed loaded with instruments for measurements including lightning, temperature, 3D wind, precipitation, liquid and ice water contents, aerosol and cloud profiles. During the field campaign, both the spaceborne and the airborne observations will be collected in real-time and integrated with the hurricane forecast models. This observation-model integration will help the campaign achieve its science goals by allowing team members to effectively plan the mission with current forecasts. To support the GRIP experiment, JPL developed a website for interactive visualization of all related remote-sensing observations in the GRIP’s geographical domain using the new Google Earth API. All the observations are collected in near real-time (NRT) with 2 to 5 hour latency. The observations include a 1KM blended Sea Surface Temperature (SST) map from GHRSST L2P products; 6-hour composite images of GOES IR; stability indices, temperature and vapor profiles from AIRS and AMSU-B; microwave brightness temperature and rain index maps from AMSR-E, SSMI and TRMM-TMI; ocean surface wind vectors, vorticity and divergence of the wind from QuikSCAT; the 3D precipitation structure from TRMM-PR and vertical profiles of cloud and precipitation from CloudSAT. All the NRT observations are collected from the data centers and science facilities at NASA and NOAA, subsetted, re-projected, and composited into hourly or daily data products depending on the frequency of the observation. The data products are then displayed on the 3D Google Earth plug-in at the JPL Tropical Cyclone Information System (TCIS) website. The data products offered by the TCIS in the Google Earth display include image overlays, wind vectors, clickable placemarks with vertical profiles for temperature and water vapors and curtain plots along the satellite tracks. Multiple products can be overlaid with individual adjustable opacity control. The time sequence visualization is supported by calendar and Google Earth time animation. The work described here was performed at the Jet Propulsion Laboratory, California Institute of Technology, under contract with the National Aeronautics and Space Administration.
Accuracy comparison in mapping water bodies using Landsat images and Google Earth Images
NASA Astrophysics Data System (ADS)
Zhou, Z.; Zhou, X.
2016-12-01
A lot of research has been done for the extraction of water bodies with multiple satellite images. The Water Indexes with the use of multi-spectral images are the mostly used methods for the water bodies' extraction. In order to extract area of water bodies from satellite images, accuracy may depend on the spatial resolution of images and relative size of the water bodies. To quantify the impact of spatial resolution and size (major and minor lengths) of the water bodies on the accuracy of water area extraction, we use Georgetown Lake, Montana and coalbed methane (CBM) water retention ponds in the Montana Powder River Basin as test sites to evaluate the impact of spatial resolution and the size of water bodies on water area extraction. Data sources used include Landsat images and Google Earth images covering both large water bodies and small ponds. Firstly we used water indices to extract water coverage from Landsat images for both large lake and small ponds. Secondly we used a newly developed visible-index method to extract water coverage from Google Earth images covering both large lake and small ponds. Thirdly, we used the image fusion method in which the Google Earth Images are fused with multi-spectral Landsat images to obtain multi-spectral images of the same high spatial resolution as the Google earth images. The actual area of the lake and ponds are measured using GPS surveys. Results will be compared and the optimal method will be selected for water body extraction.
Marshall, Zack; Brunger, Fern; Welch, Vivian; Asghari, Shabnam; Kaposy, Chris
2018-02-26
This paper focuses on the collision of three factors: a growing emphasis on sharing research through open access publication, an increasing awareness of big data and its potential uses, and an engaged public interested in the privacy and confidentiality of their personal health information. One conceptual space where this collision is brought into sharp relief is with the open availability of patient medical photographs from peer-reviewed journal articles in the search results of online image databases such as Google Images. The aim of this study was to assess the availability of patient medical photographs from published journal articles in Google Images search results and the factors impacting this availability. We conducted a cross-sectional study using data from an evidence map of research with transgender, gender non-binary, and other gender diverse (trans) participants. For the original evidence map, a comprehensive search of 15 academic databases was developed in collaboration with a health sciences librarian. Initial search results produced 25,230 references after duplicates were removed. Eligibility criteria were established to include empirical research of any design that included trans participants or their personal information and that was published in English in peer-reviewed journals. We identified all articles published between 2008 and 2015 with medical photographs of trans participants. For each reference, images were individually numbered in order to track the total number of medical photographs. We used odds ratios (OR) to assess the association between availability of the clinical photograph on Google Images and the following factors: whether the article was openly available online (open access, Researchgate.net, or Academia.edu), whether the article included genital images, if the photographs were published in color, and whether the photographs were located on the journal article landing page. We identified 94 articles with medical photographs of trans participants, including a total of 605 photographs. Of the 94 publications, 35 (37%) included at least one medical photograph that was found on Google Images. The ability to locate the article freely online contributes to the availability of at least one image from the article on Google Images (OR 2.99, 95% CI 1.20-7.45). This is the first study to document the existence of medical photographs from peer-reviewed journals appearing in Google Images search results. Almost all of the images we searched for included sensitive photographs of patient genitals, chests, or breasts. Given that it is unlikely that patients consented to sharing their personal health information in these ways, this constitutes a risk to patient privacy. Based on the impact of current practices, revisions to informed consent policies and guidelines are required. ©Zack Marshall, Fern Brunger, Vivian Welch, Shabnam Asghari, Chris Kaposy. Originally published in the Journal of Medical Internet Research (http://www.jmir.org), 26.02.2018.
Sedimentation and erosion in Lake Diefenbaker, Canada: solutions for shoreline retreat monitoring.
Sadeghian, Amir; de Boer, Dirk; Lindenschmidt, Karl-Erich
2017-09-15
This study looks into sedimentation and erosion rates in Lake Diefenbaker, a prairie reservoir, in Saskatchewan, Canada, which has been in operation since 1968. First, we looked at the historical data in all different formats over the last 70 years, which includes data from more than 20 years before the formation of the lake. The field observations indicate high rates of shoreline erosion, especially in the upstream portion as a potential region for shoreline retreat. Because of the great importance of this waterbody to the province, monitoring sedimentation and erosion rates is necessary for maintaining the quality of water especially after severe floods which are more common due to climate change effects. Second, we used Google Maps Elevation API, a new tool from Google that provides elevation data for cross sections drawn between two points, by drawing 24 cross sections in the upstream area extending 250 m from each bank. This feature from Google can be used as an easy and fast monitoring tool, is free of charge, and provides excellent control capabilities for monitoring changes in cross-sectional profiles.
2016-02-27
Sam Choi and Naiara Pinto observe Google Earth overlaid with in almost real time what the synthetic aperture radar is mapping from the C-20A aircraft. Researchers were in the sky and on the ground to take measurements of plant mass, distribution of trees, shrubs and ground cover and the diversity of plants and how much carbon is absorbed by them.
Applying Modern Stage Theory to Mauritania: A Prescription to Encourage Entrepreneurship
2014-12-01
entrepreneurship, stage theory, development, Africa , factor-driven, trade freedom, business freedom 15. NUMBER OF PAGES 77 16. PRICE CODE 17...SOUTH ASIA, SUB-SAHARAN AFRICA ) from the NAVAL POSTGRADUATE SCHOOL December 2014 Author: Jennifer M. Warren Approved by: Robert E...Notes, Coins) .......................................................................... 4 Figure 2. Satellite map of West Africa (from Google Earth
Re-Purposing Google Maps Visualisation for Teaching Logistics Systems
ERIC Educational Resources Information Center
Cheong, France; Cheong, Christopher; Jie, Ferry
2012-01-01
Routing is the process of selecting appropriate paths and ordering waypoints in a network. It plays an important part in logistics and supply chain management as choosing the optimal route can minimise distribution costs. Routing optimisation, however, is a difficult problem to solve and computer software is often used to determine the best route.…
On-line Geoscience Data Resources for Today's Undergraduates
NASA Astrophysics Data System (ADS)
Goodwillie, A. M.; Ryan, W.; Carbotte, S.; Melkonian, A.; Coplan, J.; Arko, R.; O'Hara, S.; Ferrini, V.; Leung, A.; Bonckzowski, J.
2008-12-01
Broadening the experience of undergraduates can be achieved by enabling free, unrestricted and convenient access to real scientific data. With funding from the U.S. National Science Foundation, the Marine Geoscience Data System (MGDS) (http://www.marine-geo.org/) serves as the integrated data portal for various NSF-funded projects and provides free public access and preservation to a wide variety of marine and terrestrial data including rock, fluid, biology and sediment samples information, underway geophysical data and multibeam bathymetry, water column and multi-channel seismics data. Users can easily view the locations of cruise tracks, sample and station locations against a backdrop of a multi-resolution global digital elevation model. A Search For Data web page rapidly extracts data holdings from the database and can be filtered on data and device type, field program ID, investigator name, geographical and date bounds. The data access experience is boosted by the MGDS use of standardised OGC-compliant Web Services to support uniform programmatic interfaces. GeoMapApp (http://www.geomapapp.org/), a free MGDS data visualization tool, supports map-based dynamic exploration of a broad suite of geosciences data. Built-in land and marine data sets include tectonic plate boundary compilations, DSDP/ODP core logs, earthquake events, seafloor photos, and submersible dive tracks. Seamless links take users to data held by external partner repositories including PetDB, UNAVCO, IRIS and NGDC. Users can generate custom maps and grids and import their own data sets and grids. A set of short, video-style on-line tutorials familiarises users step- by-step with GeoMapApp functionality (http://www.geomapapp.org/tutorials/). Virtual Ocean (http://www.virtualocean.org/) combines the functionality of GeoMapApp with a 3-D earth browser built using the NASA WorldWind API for a powerful new data resource. MGDS education involvement (http://www.marine-geo.org/, go to Education tab) includes the searchable Media Bank of images and video; KML files for viewing several MGDS data sets in Google Earth (tm); support in developing undergraduate- level teaching modules using NSF-MARGINS data. Examples of many of these data sets will be shown.
Fostering learners' interaction with content: A learner-centered mobile device interface
NASA Astrophysics Data System (ADS)
Abdous, M.
2015-12-01
With the ever-increasing omnipresence of mobile devices in student life, leveraging smart devices to foster students' interaction with course content is critical. Following a learner-centered design iterative approach, we designed a mobile interface that may enable learners to access and interact with online course content efficiently and intuitively. Our design process leveraged recent technologies, such as bootstrap, Google's Material Design, HTML5, and JavaScript to design an intuitive, efficient, and portable mobile interface with a variety of built-in features, including context sensitive bookmarking, searching, progress tracking, captioning, and transcript display. The mobile interface also offers students the ability to ask context-related questions and to complete self-checks as they watch audio/video presentations. Our design process involved ongoing iterative feedback from learners, allowing us to refine and tweak the interface to provide learners with a unified experience across platforms and devices. The innovative combination of technologies built around well-structured and well-designed content seems to provide an effective learning experience to mobile learners. Early feedback indicates a high level of satisfaction with the interface's efficiency, intuitiveness, and robustness from both students and faculty.
Common Ground: An Interactive Visual Exploration and Discovery for Complex Health Data
2014-04-01
annotate other ontologies for the visual interface client. Finally, we are actively working on software development of both a backend server and the...the following infrastructure and resources. For the development and management of the ontologies, we installed a framework consisting of a server...that is being developed by Google. Using these 9 technologies, we developed an HTML5 client that runs on Windows, Mac OSX, Linux and mobile systems
How Would You Move Mount Fuji - And Why Would You Want To?
NASA Astrophysics Data System (ADS)
de Paor, D. G.
2008-12-01
According to author William Poundstone, "How Would You Move Mt Fuji?" typifies the kind of question that corporations such as Microsoft are wont to ask job applicants in order to test their lateral thinking skills. One answer (albeit not one that would necessarily secure a job at Microsoft) is: "With Google Earth and a Macintosh or PC." The answer to the more profound follow-up question "Why Would You Want To?" is hinted at by one of the great quotations of earth science, namely Charles Lyell's proposition that "The Present Is Key to the Past." Google Earth is a phenomenally powerful tool for visualizing today's earth, ocean, and atmosphere. With the aid of Google SketchUp, that visualization can be extended to reconstruct the past using relocated samples of present-day landscapes and environments as models of paleo-DEM and paleogeography. Volcanoes are particularly useful models because their self similar growth can be simulated by changing KML altitude tags within a timespan, but numerous other landforms and geologic structures serve as useful keys to the past. Examples range in scale from glaciers and fault scarps to island arcs and mountain ranges. The ability to generate a paleo-terrain model in Google Earth brings us one step closer to a truly four- dimensional, interactive geological map of the world throughout time.
Seismicity map tools for earthquake studies
NASA Astrophysics Data System (ADS)
Boucouvalas, Anthony; Kaskebes, Athanasios; Tselikas, Nikos
2014-05-01
We report on the development of new and online set of tools for use within Google Maps, for earthquake research. We demonstrate this server based and online platform (developped with PHP, Javascript, MySQL) with the new tools using a database system with earthquake data. The platform allows us to carry out statistical and deterministic analysis on earthquake data use of Google Maps and plot various seismicity graphs. The tool box has been extended to draw on the map line segments, multiple straight lines horizontally and vertically as well as multiple circles, including geodesic lines. The application is demonstrated using localized seismic data from the geographic region of Greece as well as other global earthquake data. The application also offers regional segmentation (NxN) which allows the studying earthquake clustering, and earthquake cluster shift within the segments in space. The platform offers many filters such for plotting selected magnitude ranges or time periods. The plotting facility allows statistically based plots such as cumulative earthquake magnitude plots and earthquake magnitude histograms, calculation of 'b' etc. What is novel for the platform is the additional deterministic tools. Using the newly developed horizontal and vertical line and circle tools we have studied the spatial distribution trends of many earthquakes and we here show for the first time the link between Fibonacci Numbers and spatiotemporal location of some earthquakes. The new tools are valuable for examining visualizing trends in earthquake research as it allows calculation of statistics as well as deterministic precursors. We plan to show many new results based on our newly developed platform.
NASA Astrophysics Data System (ADS)
Drosos, Vasileios C.; Liampas, Sarantis-Aggelos G.; Doukas, Aristotelis-Kosmas G.
2014-08-01
In our time, the Geographic Information Systems (GIS) have become important tools, not only in the geosciences and environmental sciences, as well as virtually for all researches that require monitoring, planning or land management. The purpose of this paper was to develop a planning tool and decision making tool using AutoCAD Map software, ArcGIS and Google Earth with emphasis on the investigation of the suitability of forest roads' mapping and the range of its implementation in Greece in prefecture level. Integrating spatial information into a database makes data available throughout the organization; improving quality, productivity, and data management. Also working in such an environment, you can: Access and edit information, integrate and analyze data and communicate effectively. To select desirable information such as forest road network in a very early stage in the planning of silviculture operations, for example before the planning of the harvest is carried out. The software programs that were used were AutoCAD Map for the export in shape files for the GPS data, and ArcGIS in shape files (ArcGlobe), while Google Earth with KML files (Keyhole Markup Language) in order to better visualize and evaluate existing conditions, design in a real-world context and exchange information with government agencies, utilities, and contractors in both CAD and GIS data formats. The automation of the updating procedure and transfer of any files between agencies-departments is one of the main tasks of the integrated GIS-tool among the others should be addressed.
Thompson, Jolinda L; Obrig, Kathe S; Abate, Laura E
2013-01-01
Funds made available at the close of the 2010-11 fiscal year allowed purchase of the EBSCO Discovery Service (EDS) for a year-long trial. The appeal of this web-scale discovery product that offers a Google-like interface to library resources was counter-balanced by concerns about quality of search results in an academic health science setting and the challenge of configuring an interface that serves the needs of a diverse group of library users. After initial configuration, usability testing with library users revealed the need for further work before general release. Of greatest concern were continuing issues with the relevance of items retrieved, appropriateness of system-supplied facet terms, and user difficulties with navigating the interface. EBSCO has worked with the library to better understand and identify problems and solutions. External roll-out to users occurred in June 2012.
Google Earth-Based Grand Tours of the World's Ocean Basins and Marine Sediments
NASA Astrophysics Data System (ADS)
St John, K. K.; De Paor, D. G.; Suranovic, B.; Robinson, C.; Firth, J. V.; Rand, C.
2016-12-01
The GEODE project has produced a collection of Google Earth-based marine geology teaching resources that offer grand tours of the world's ocean basins and marine sediments. We use a map of oceanic crustal ages from Müller et al (2008; doi:10.1029/2007GC001743), and a set of emergent COLLADA models of IODP drill core data as a basis for a Google Earth tour introducing students to the world's ocean basins. Most students are familiar with basic seafloor spreading patterns but teaching experience suggests that few students have an appreciation of the number of abandoned ocean basins on Earth. Students also lack a valid visualization of the west Pacific where the oldest crust forms an isolated triangular patch and the ocean floor becomes younger towards the subduction zones. Our tour links geographic locations to mechanical models of rifting, seafloor spreading, subduction, and transform faulting. Google Earth's built-in earthquake and volcano data are related to ocean floor patterns. Marine sediments are explored in a Google Earth tour that draws on exemplary IODP core samples of a range of sediment types (e.g., turbidites, diatom ooze). Information and links are used to connect location to sediment type. This tour compliments a physical core kit of core catcher sections that can be employed for classroom instruction (geode.net/marine-core-kit/). At a larger scale, we use data from IMLGS to explore the distribution of the marine sediments types in the modern global ocean. More than 2,500 sites are plotted with access to original data. Students are guided to compare modern "type sections" of primary marine sediment lithologies, as well as examine site transects to address questions of bathymetric setting, ocean circulation, chemistry (e.g., CCD), and bioproductivity as influences on modern seafloor sedimentation. KMZ files, student exercises, and tips for instructors are available at geode.net/exploring-marine-sediments-using-google-earth.
NASA Astrophysics Data System (ADS)
Fritz, S.; Nordling, J.; See, L. M.; McCallum, I.; Perger, C.; Becker-Reshef, I.; Mucher, S.; Bydekerke, L.; Havlik, P.; Kraxner, F.; Obersteiner, M.
2014-12-01
The International Institute for Applied Systems Analysis (IIASA) has developed a global cropland extent map, which supports the monitoring and assessment activities of GEOGLAM (Group on Earth Observations Global Agricultural Monitoring Initiative). Through the European-funded SIGMA (Stimulating Innovation for Global Monitoring of Agriculture and its Impact on the Environment in support of GEOGLAM) project, IIASA is continuing to support GEOGLAM by providing cropland projections in the future and modelling environmental impacts on agriculture under various scenarios. In addition, IIASA is focusing on two specific elements within SIGMA: the development of a global field size and irrigation map; and mobile app development for in-situ data collection and validation of remotely-sensed products. Cropland field size is a very useful indicator for agricultural monitoring yet the information we have at a global scale is currently very limited. IIASA has already created a global map of field size at a 1 km resolution using crowdsourced data from Geo-Wiki as a first approximation. Using automatic classification of Landsat imagery and algorithms contained within Google Earth Engine, initial experimentation has shown that circular fields and landscape structures can easily be extracted. Not only will this contribute to improving the global map of field size, it can also be used to create a global map that contains a large proportion of the world's irrigated areas, which will be another useful contribution to GEOGLAM. The field size map will also be used to stratify and develop a global crop map in SIGMA. Mobile app development in support of in-situ data collection is another area where IIASA is currently working. An Android app has been built using the Open Data Toolkit (ODK) and extended further with spatial mapping capabilities called GeoODK. The app allows users to collect data on different crop types and delineate fields on the ground, which can be used to validate the field size map. The app can also cache map data so that high resolution satellite imagery and reference data from the users can be viewed in the field without the need for an internet connection. This app will be used for calibration and validation of the data products in SIGMA, e.g. data collection at JECAM (Joint Experiment of Crop Assessment and Monitoring) sites.
NASA Astrophysics Data System (ADS)
Chen, Bangqian; Xiao, Xiangming; Li, Xiangping; Pan, Lianghao; Doughty, Russell; Ma, Jun; Dong, Jinwei; Qin, Yuanwei; Zhao, Bin; Wu, Zhixiang; Sun, Rui; Lan, Guoyu; Xie, Guishui; Clinton, Nicholas; Giri, Chandra
2017-09-01
Due to rapid losses of mangrove forests caused by anthropogenic disturbances and climate change, accurate and contemporary maps of mangrove forests are needed to understand how mangrove ecosystems are changing and establish plans for sustainable management. In this study, a new classification algorithm was developed using the biophysical characteristics of mangrove forests in China. More specifically, these forests were mapped by identifying: (1) greenness, canopy coverage, and tidal inundation from time series Landsat data, and (2) elevation, slope, and intersection-with-sea criterion. The annual mean Normalized Difference Vegetation Index (NDVI) was found to be a key variable in determining the classification thresholds of greenness, canopy coverage, and tidal inundation of mangrove forests, which are greatly affected by tide dynamics. In addition, the integration of Sentinel-1A VH band and modified Normalized Difference Water Index (mNDWI) shows great potential in identifying yearlong tidal and fresh water bodies, which is related to mangrove forests. This algorithm was developed using 6 typical Regions of Interest (ROIs) as algorithm training and was run on the Google Earth Engine (GEE) cloud computing platform to process 1941 Landsat images (25 Path/Row) and 586 Sentinel-1A images circa 2015. The resultant mangrove forest map of China at 30 m spatial resolution has an overall/users/producer's accuracy greater than 95% when validated with ground reference data. In 2015, China's mangrove forests had a total area of 20,303 ha, about 92% of which was in the Guangxi Zhuang Autonomous Region, Guangdong, and Hainan Provinces. This study has demonstrated the potential of using the GEE platform, time series Landsat and Sentine-1A SAR images to identify and map mangrove forests along the coastal zones. The resultant mangrove forest maps are likely to be useful for the sustainable management and ecological assessments of mangrove forests in China.
Dong, Jinwei; Xiao, Xiangming; Menarguez, Michael A.; Zhang, Geli; Qin, Yuanwei; Thau, David; Biradar, Chandrashekhar; Moore, Berrien
2016-01-01
Area and spatial distribution information of paddy rice are important for understanding of food security, water use, greenhouse gas emission, and disease transmission. Due to climatic warming and increasing food demand, paddy rice has been expanding rapidly in high latitude areas in the last decade, particularly in northeastern (NE) Asia. Current knowledge about paddy rice fields in these cold regions is limited. The phenology- and pixel-based paddy rice mapping (PPPM) algorithm, which identifies the flooding signals in the rice transplanting phase, has been effectively applied in tropical areas, but has not been tested at large scale of cold regions yet. Despite the effects from more snow/ice, paddy rice mapping in high latitude areas is assumed to be more encouraging due to less clouds, lower cropping intensity, and more observations from Landsat sidelaps. Moreover, the enhanced temporal and geographic coverage from Landsat 8 provides an opportunity to acquire phenology information and map paddy rice. This study evaluated the potential of Landsat 8 images on annual paddy rice mapping in NE Asia which was dominated by single cropping system, including Japan, North Korea, South Korea, and NE China. The cloud computing approach was used to process all the available Landsat 8 imagery in 2014 (143 path/rows, ~3290 scenes) with the Google Earth Engine (GEE) platform. The results indicated that the Landsat 8, GEE, and improved PPPM algorithm can effectively support the yearly mapping of paddy rice in NE Asia. The resultant paddy rice map has a high accuracy with the producer (user) accuracy of 73% (92%), based on the validation using very high resolution images and intensive field photos. Geographic characteristics of paddy rice distribution were analyzed from aspects of country, elevation, latitude, and climate. The resultant 30-m paddy rice map is expected to provide unprecedented details about the area, spatial distribution, and landscape pattern of paddy rice fields in NE Asia, which will contribute to food security assessment, water resource management, estimation of greenhouse gas emissions, and disease control. PMID:28025586
Dong, Jinwei; Xiao, Xiangming; Menarguez, Michael A; Zhang, Geli; Qin, Yuanwei; Thau, David; Biradar, Chandrashekhar; Moore, Berrien
2016-11-01
Area and spatial distribution information of paddy rice are important for understanding of food security, water use, greenhouse gas emission, and disease transmission. Due to climatic warming and increasing food demand, paddy rice has been expanding rapidly in high latitude areas in the last decade, particularly in northeastern (NE) Asia. Current knowledge about paddy rice fields in these cold regions is limited. The phenology- and pixel-based paddy rice mapping (PPPM) algorithm, which identifies the flooding signals in the rice transplanting phase, has been effectively applied in tropical areas, but has not been tested at large scale of cold regions yet. Despite the effects from more snow/ice, paddy rice mapping in high latitude areas is assumed to be more encouraging due to less clouds, lower cropping intensity, and more observations from Landsat sidelaps. Moreover, the enhanced temporal and geographic coverage from Landsat 8 provides an opportunity to acquire phenology information and map paddy rice. This study evaluated the potential of Landsat 8 images on annual paddy rice mapping in NE Asia which was dominated by single cropping system, including Japan, North Korea, South Korea, and NE China. The cloud computing approach was used to process all the available Landsat 8 imagery in 2014 (143 path/rows, ~3290 scenes) with the Google Earth Engine (GEE) platform. The results indicated that the Landsat 8, GEE, and improved PPPM algorithm can effectively support the yearly mapping of paddy rice in NE Asia. The resultant paddy rice map has a high accuracy with the producer (user) accuracy of 73% (92%), based on the validation using very high resolution images and intensive field photos. Geographic characteristics of paddy rice distribution were analyzed from aspects of country, elevation, latitude, and climate. The resultant 30-m paddy rice map is expected to provide unprecedented details about the area, spatial distribution, and landscape pattern of paddy rice fields in NE Asia, which will contribute to food security assessment, water resource management, estimation of greenhouse gas emissions, and disease control.
NASA Astrophysics Data System (ADS)
Thompson, C. K.; Bingham, A. W.; Hall, J. R.; Alarcon, C.; Plesea, L.; Henderson, M. L.; Levoe, S.
2011-12-01
The State of the Oceans (SOTO) web tool was developed at NASA's Physical Oceanography Distributed Active Archive Center (PO.DAAC) at the Jet Propulsion Laboratory (JPL) as an interactive means for users to visually explore and assess ocean-based geophysical parameters extracted from the latest archived data products. The SOTO system consists of four extensible modules, a data polling tool, a preparation and imaging package, image server software, and the graphical user interface. Together, these components support multi-resolution visualization of swath (Level 2) and gridded Level 3/4) data products as either raster- or vector- based KML layers on Google Earth. These layers are automatically updated periodically throughout the day. Current parameters available include sea surface temperature, chlorophyll concentration, ocean winds, sea surface height anomaly, and sea surface temperature anomaly. SOTO also supports mash-ups, allowing KML feeds from other sources to be overlaid directly onto Google Earth such as hurricane tracks and buoy data. A version of the SOTO software has also been installed at Goddard Space Flight Center (GSFC) to support the Land Atmosphere Near real-time Capability for EOS (LANCE). The State of the Earth (SOTE) has similar functionality to SOTO but supports different data sets, among them the MODIS 250m data product.
Do Interactive Globes and Games Help Students Learn Planetary Science?
NASA Astrophysics Data System (ADS)
Coba, Filis; Burgin, Stephen; De Paor, Declan; Georgen, Jennifer
2016-01-01
The popularity of animations and interactive visualizations in undergraduate science education might lead one to assume that these teaching aids enhance student learning. We tested this assumption for the case of the Google Earth virtual globe with a comparison of control and treatment student groups in a general education class of over 370 students at a large public university. Earth and Planetary Science course content was developed in two formats: using Keyhole Markup Language (KML) to create interactive tours in Google Earth (the treatment group) and Portable Document Format (PDF) for on-screen reading (the control group). The PDF documents contained identical text and images to the placemark balloons or "tour stops" in the Google Earth version. Some significant differences were noted between the two groups based on the immediate post-questionnaire with the KML students out-performing the PDF students, but not on the delayed measure. In a separate but related project, we undertake preliminary investigations into methods of teaching basic concepts in planetary mantle convection using numerical simulations. The goal of this project is to develop an interface with a two-dimensional finite element model that will allow students to vary parameters such as the temperatures assigned to the boundaries of the model domain, to help them actively explore important variables that control convection.
TerraceM: A Matlab® tool to analyze marine terraces from high-resolution topography
NASA Astrophysics Data System (ADS)
Jara-Muñoz, Julius; Melnick, Daniel; Strecker, Manfred
2015-04-01
To date, Light detection and ranging (LiDAR), high- resolution topographic data sets enable remote identification of submeter-scale geomorphic features bringing valuable information of the landscape and geomorphic markers of tectonic deformation such as fault-scarp offsets, fluvial and marine terraces. Recent studies of marine terraces using LiDAR data have demonstrated that these landforms can be readily isolated from other landforms in the landscape, using slope and roughness parameters that allow for unambiguously mapping regional extents of terrace sequences. Marine terrace elevation has been used since decades as geodetic benchmarks of Quaternary deformation. Uplift rates may be estimated by locating the shoreline angle, a geomorphic feature correlated with the high-stand position of past sea levels. Indeed, precise identification of the shoreline-angle position is an important requirement to obtain reliable tectonic rates and coherent spatial correlation. To improve our ability to rapidly assess and map different shoreline angles at a regional scale we have developed the TerraceM application. TerraceM is a Matlab® tool that allows estimating the shoreline angle and its associated error using high-resolution topography. For convenience, TerraceM includes a graphical user interface (GUI) linked with Google Maps® API. The analysis starts by defining swath profiles from a shapefile created on a GIS platform orientated orthogonally to the terrace riser. TerraceM functions are included to extract and analyze the swath profiles. Two types of coastal landscapes may be analyzed using different methodologies: staircase sequences of multiple terraces and rough, rocky coasts. The former are measured by outlining the paleo-cliffs and paleo-platforms, whereas the latter are assessed by picking the elevation of sea-stack tops. By calculating the intersection between first-order interpolations of the maximum topography of swath profiles we define the shoreline angle in staircase terraces. For rocky coasts, the maximum stack peaks for a defined search ratio as well as a defined inflection point on the adjacent main cliff are interpolated to calculate the shoreline angle at the intersection with the cliff. Error estimates are based on the standard deviation of the linear regressions. The geomorphic age of terraces (Kt) can be also calculated by the linear diffusion equation (Hanks et al., 1989), with a best-fitting model found by minimizing the RMS. TerraceM has the ability to efficiently process several profiles in batch-mode run. Results may be exported in various formats, including Google Earth and ArcGis, basic statistics are automatically computed. Test runs have been made at Santa Cruz, California, using various topographic data sets and comparing results with published field measurements (Anderson and Menking, 1994). Repeatability was evaluated using multiple test runs made by students in a classroom setting.
NASA Astrophysics Data System (ADS)
Listyorini, Tri; Muzid, Syafiul
2017-06-01
The promotion team of Muria Kudus University (UMK) has done annual promotion visit to several senior high schools in Indonesia. The visits were done to numbers of schools in Kudus, Jepara, Demak, Rembang and Purwodadi. To simplify the visit, each visit round is limited to 15 (fifteen) schools. However, the team frequently faces some obstacles during the visit, particularly in determining the route that they should take toward the targeted school. It is due to the long distance or the difficult route to reach the targeted school that leads to elongated travel duration and inefficient fuel cost. To solve these problems, the development of a certain application using heuristic genetic algorithm method based on the dynamic of population size or Population Resizing on Fitness lmprovement Genetic Algorithm (PRoFIGA), was done. This android-based application was developed to make the visit easier and to determine a shorter route for the team, hence, the visiting period will be effective and efficient. The result of this research was an android-based application to determine the shortest route by combining heuristic method and Google Maps Application Programming lnterface (API) that display the route options for the team.
KML Super Overlay to WMS Translator
NASA Technical Reports Server (NTRS)
Plesea, Lucian
2007-01-01
This translator is a server-based application that automatically generates KML super overlay configuration files required by Google Earth for map data access via the Open Geospatial Consortium WMS (Web Map Service) standard. The translator uses a set of URL parameters that mirror the WMS parameters as much as possible, and it also can generate a super overlay subdivision of any given area that is only loaded when needed, enabling very large areas of coverage at very high resolutions. It can make almost any dataset available as a WMS service visible and usable in any KML application, without the need to reformat the data.
NASA Astrophysics Data System (ADS)
Lew, Roger; Dobre, Mariana; Elliot, William; Robichaud, Pete; Brooks, Erin; Frankenberger, Jim
2017-04-01
There is an increased interest in the United States to use soil burn severity maps in watershed-scale hydrologic models to estimate post-fire sediment erosion from burned areas. This information is needed by stakeholders in order to concentrate their pre- or post-fire management efforts in ecologically sensitive areas to decrease the probability of post-fire sediment delivery. But these tools traditionally have been time consuming and difficult to use by managers because input datasets must be obtained and correctly processed for valid results. The Water Erosion Prediction Project (WEPP) has previously been developed as an online and easy-to-use interface to help land managers with running simulations without any knowledge of computer programming or hydrologic modeling. The interface automates the acquisition of DEM, climate, soils, and landcover data, and also automates channel and hillslope delineation for the users. The backend is built with Mapserver, GDAL, PHP, C++, Python while the front end uses OpenLayers, and, of course, JavaScript. The existing WEPP online interface was enhanced to provide better usability to stakeholders in United States (Forest Service, BLM, USDA) as well as to provide enhanced functionality for managing both pre-fire and post-fire treatments. Previously, only site administrators could add burn severity maps. The interface now allows users to create accounts to upload and share FlamMap prediction maps, differenced Normalized Burned Ratio (dNBR), or Burned Area Reflectance Classification (BARC) maps. All maps are loaded into a sortable catalog so users can quickly find their area of interest. Once loaded, the interface has been modified to support running comparisons between baseline condition with "no burn" and with a burn severity classification map. The interface has also been enhanced to allow users to conduct single storm analyses to examine, for example, how much soil loss would result after a 100-year storm. An OpenLayers map allows users to overlay the watershed hillslopes and channels, burn severity, and erosion. The interface provides flowpath results for each hillslope and at the outlet, as well as return period and frequency analysis reports. Once problematic areas have been identified, the interface allows users to export the watershed in a format that can be used by the Erosion Risk Management Tool (ERMiT) and Disturbed WEPP (post-disturbance modeling) for more detailed hillslope-level analyses. Numerous other changes were made to improve the overall usability of the interface: allow simulations in both SI and English units, added immovable pop-up dialogs to guide the users, and removed extraneous information from the interface. In upcoming months, a workshop will be conducted to demonstrate these new capabilities to stakeholders. Efforts are underway to use site-specific SSURGO soils to that are modified based on burn severity rather than using generic soil classes.
The Lunar Mapping and Modeling Project
NASA Technical Reports Server (NTRS)
Nall, M.; French, R.; Noble, S.; Muery, K.
2010-01-01
The Lunar Mapping and Modeling Project (LMMP) is managing a suite of lunar mapping and modeling tools and data products that support lunar exploration activities, including the planning, de-sign, development, test, and operations associated with crewed and/or robotic operations on the lunar surface. Although the project was initiated primarily to serve the needs of the Constellation program, it is equally suited for supporting landing site selection and planning for a variety of robotic missions, including NASA science and/or human precursor missions and commercial missions such as those planned by the Google Lunar X-Prize participants. In addition, LMMP should prove to be a convenient and useful tool for scientific analysis and for education and public out-reach (E/PO) activities.
NASA Astrophysics Data System (ADS)
Webley, P.; Dehn, J.; Dean, K. G.; Macfarlane, S.
2010-12-01
Volcanic eruptions are a global hazard, affecting local infrastructure, impacting airports and hindering the aviation community, as seen in Europe during Spring 2010 from the Eyjafjallajokull eruption in Iceland. Here, we show how remote sensing data is used through web-based interfaces for monitoring volcanic activity, both ground based thermal signals and airborne ash clouds. These ‘web tools’, http://avo.images.alaska.edu/, provide timely availability of polar orbiting and geostationary data from US National Aeronautics and Space Administration, National Oceanic and Atmosphere Administration and Japanese Meteorological Agency satellites for the North Pacific (NOPAC) region. This data is used operationally by the Alaska Volcano Observatory (AVO) for monitoring volcanic activity, especially at remote volcanoes and generates ‘alarms’ of any detected volcanic activity and ash clouds. The webtools allow the remote sensing team of AVO to easily perform their twice daily monitoring shifts. The web tools also assist the National Weather Service, Alaska and Kamchatkan Volcanic Emergency Response Team, Russia in their operational duties. Users are able to detect ash clouds, measure the distance from the source, area and signal strength. Within the web tools, there are 40 x 40 km datasets centered on each volcano and a searchable database of all acquired data from 1993 until present with the ability to produce time series data per volcano. Additionally, a data center illustrates the acquired data across the NOPAC within the last 48 hours, http://avo.images.alaska.edu/tools/datacenter/. We will illustrate new visualization tools allowing users to display the satellite imagery within Google Earth/Maps, and ArcGIS Explorer both as static maps and time-animated imagery. We will show these tools in real-time as well as examples of past large volcanic eruptions. In the future, we will develop the tools to produce real-time ash retrievals, run volcanic ash dispersion models from detected ash clouds and develop the browser interfaces to display other remote sensing datasets, such as volcanic sulfur dioxide detection.
Spurgeon, Steven R; Du, Yingge; Chambers, Scott A
2017-06-01
With the development of affordable aberration correctors, analytical scanning transmission electron microscopy (STEM) studies of complex interfaces can now be conducted at high spatial resolution at laboratories worldwide. Energy-dispersive X-ray spectroscopy (EDS) in particular has grown in popularity, as it enables elemental mapping over a wide range of ionization energies. However, the interpretation of atomically resolved data is greatly complicated by beam-sample interactions that are often overlooked by novice users. Here we describe the practical factors-namely, sample thickness and the choice of ionization edge-that affect the quantification of a model perovskite oxide interface. Our measurements of the same sample, in regions of different thickness, indicate that interface profiles can vary by as much as 2-5 unit cells, depending on the spectral feature. This finding is supported by multislice simulations, which reveal that on-axis maps of even perfectly abrupt interfaces exhibit significant delocalization. Quantification of thicker samples is further complicated by channeling to heavier sites across the interface, as well as an increased signal background. We show that extreme care must be taken to prepare samples to minimize channeling effects and argue that it may not be possible to extract atomically resolved information from many chemical maps.
Using Webgis and Cloud Tools to Promote Cultural Heritage Dissemination: the Historic up Project
NASA Astrophysics Data System (ADS)
Tommasi, A.; Cefalo, R.; Zardini, F.; Nicolaucig, M.
2017-05-01
On the occasion of the First World War centennial, GeoSNav Lab (Geodesy and Satellite Navigation Laboratory), Department of Engineering and Architecture, University of Trieste, Italy, in coooperation with Radici&Futuro Association, Trieste, Italy, carried out an educational Project named "Historic Up" involving a group of students from "F. Petrarca" High School of Trieste, Italy. The main goal of the project is to make available to students of Middle and High Schools a set of historical and cultural contents in a simple and immediate way, through the production of a virtual and interactive tour following the event that caused the burst of the First World War: the assassination of Franz Ferdinand and his wife Sofia in Sarajevo occurred on June 28, 1914. A set of Google Apps was used, including Google Earth, Maps, Tour Builder, Street View, Gmail, Drive, and Docs. The Authors instructed the students about software and team-working and supported them along the research. After being checked, all the historical and geographic data have been uploaded on a Google Tour Builder to create a sequence of historical checkpoints. Each checkpoint has texts, pictures and videos that connect the tour-users to 1914. Moreover, GeoSNaV Lab researchers produced a KML (Keyhole Markup Language) file, formed by several polylines and points, representing the itinerary of the funeral procession that has been superimposed on ad-hoc georeferenced historical maps. This tour, freely available online, starts with the arrival of the royals, on June 28th 1914, and follows the couple along the events, from the assassination to the burial in Arstetten (Austria), including their passages through Trieste (Italy), Ljubljana (Slovenia), Graz and Wien (Austria).
TumorMap: Exploring the Molecular Similarities of Cancer Samples in an Interactive Portal.
Newton, Yulia; Novak, Adam M; Swatloski, Teresa; McColl, Duncan C; Chopra, Sahil; Graim, Kiley; Weinstein, Alana S; Baertsch, Robert; Salama, Sofie R; Ellrott, Kyle; Chopra, Manu; Goldstein, Theodore C; Haussler, David; Morozova, Olena; Stuart, Joshua M
2017-11-01
Vast amounts of molecular data are being collected on tumor samples, which provide unique opportunities for discovering trends within and between cancer subtypes. Such cross-cancer analyses require computational methods that enable intuitive and interactive browsing of thousands of samples based on their molecular similarity. We created a portal called TumorMap to assist in exploration and statistical interrogation of high-dimensional complex "omics" data in an interactive and easily interpretable way. In the TumorMap, samples are arranged on a hexagonal grid based on their similarity to one another in the original genomic space and are rendered with Google's Map technology. While the important feature of this public portal is the ability for the users to build maps from their own data, we pre-built genomic maps from several previously published projects. We demonstrate the utility of this portal by presenting results obtained from The Cancer Genome Atlas project data. Cancer Res; 77(21); e111-4. ©2017 AACR . ©2017 American Association for Cancer Research.
Map based multimedia tool on Pacific theatre in World War II
NASA Astrophysics Data System (ADS)
Pakala Venkata, Devi Prasada Reddy
Maps have been used for depicting data of all kinds in the educational community for many years. A standout amongst the rapidly changing methods of teaching is through the development of interactive and dynamic maps. The emphasis of the thesis is to develop an intuitive map based multimedia tool, which provides a timeline of battles and events in the Pacific theatre of World War II. The tool contains summaries of major battles and commanders and has multimedia content embedded in it. The primary advantage of this Map tool is that one can quickly know about all the battles and campaigns of the Pacific Theatre by accessing Timeline of Battles in each region or Individual Battles in each region or Summary of each Battle in an interactive way. This tool can be accessed via any standard web browser and motivate the user to know more about the battles involved in the Pacific Theatre. It was made responsive using Google maps API, JavaScript, HTML5 and CSS.
NASA Astrophysics Data System (ADS)
Moore, R. T.; Hansen, M. C.
2011-12-01
Google Earth Engine is a new technology platform that enables monitoring and measurement of changes in the earth's environment, at planetary scale, on a large catalog of earth observation data. The platform offers intrinsically-parallel computational access to thousands of computers in Google's data centers. Initial efforts have focused primarily on global forest monitoring and measurement, in support of REDD+ activities in the developing world. The intent is to put this platform into the hands of scientists and developing world nations, in order to advance the broader operational deployment of existing scientific methods, and strengthen the ability for public institutions and civil society to better understand, manage and report on the state of their natural resources. Earth Engine currently hosts online nearly the complete historical Landsat archive of L5 and L7 data collected over more than twenty-five years. Newly-collected Landsat imagery is downloaded from USGS EROS Center into Earth Engine on a daily basis. Earth Engine also includes a set of historical and current MODIS data products. The platform supports generation, on-demand, of spatial and temporal mosaics, "best-pixel" composites (for example to remove clouds and gaps in satellite imagery), as well as a variety of spectral indices. Supervised learning methods are available over the Landsat data catalog. The platform also includes a new application programming framework, or "API", that allows scientists access to these computational and data resources, to scale their current algorithms or develop new ones. Under the covers of the Google Earth Engine API is an intrinsically-parallel image-processing system. Several forest monitoring applications powered by this API are currently in development and expected to be operational in 2011. Combining science with massive data and technology resources in a cloud-computing framework can offer advantages of computational speed, ease-of-use and collaboration, as well as transparency in data and methods. Methods developed for global processing of MODIS data to map land cover are being adopted for use with Landsat data. Specifically, the MODIS Vegetation Continuous Field product methodology has been applied for mapping forest extent and change at national scales using Landsat time-series data sets. Scaling this method to continental and global scales is enabled by Google Earth Engine computing capabilities. By combining the supervised learning VCF approach with the Landsat archive and cloud computing, unprecedented monitoring of land cover dynamics is enabled.
Urban topography for flood modeling by fusion of OpenStreetMap, SRTM and local knowledge
NASA Astrophysics Data System (ADS)
Winsemius, Hessel; Donchyts, Gennadii; Eilander, Dirk; Chen, Jorik; Leskens, Anne; Coughlan, Erin; Mawanda, Shaban; Ward, Philip; Diaz Loaiza, Andres; Luo, Tianyi; Iceland, Charles
2016-04-01
Topography data is essential for understanding and modeling of urban flood hazard. Within urban areas, much of the topography is defined by highly localized man-made features such as roads, channels, ditches, culverts and buildings. This results in the requirement that urban flood models require high resolution topography, and water conveying connections within the topography are considered. In recent years, more and more topography information is collected through LIDAR surveys however there are still many cities in the world where high resolution topography data is not available. Furthermore, information on connectivity is required for flood modelling, even when LIDAR data are used. In this contribution, we demonstrate how high resolution terrain data can be synthesized using a fusion between features in OpenStreetMap (OSM) data (including roads, culverts, channels and buildings) and existing low resolution and noisy SRTM elevation data using the Google Earth Engine platform. Our method uses typical existing OSM properties to estimate heights and topology associated with the features, and uses these to correct noise and burn features on top of the existing low resolution SRTM elevation data. The method has been setup in the Google Earth Engine platform so that local stakeholders and mapping teams can on-the-fly propose, include and visualize the effect of additional features and properties of features, which are deemed important for topography and water conveyance. These features can be included in a workshop environment. We pilot our tool over Dar Es Salaam.
Machine Learning for Flood Prediction in Google Earth Engine
NASA Astrophysics Data System (ADS)
Kuhn, C.; Tellman, B.; Max, S. A.; Schwarz, B.
2015-12-01
With the increasing availability of high-resolution satellite imagery, dynamic flood mapping in near real time is becoming a reachable goal for decision-makers. This talk describes a newly developed framework for predicting biophysical flood vulnerability using public data, cloud computing and machine learning. Our objective is to define an approach to flood inundation modeling using statistical learning methods deployed in a cloud-based computing platform. Traditionally, static flood extent maps grounded in physically based hydrologic models can require hours of human expertise to construct at significant financial cost. In addition, desktop modeling software and limited local server storage can impose restraints on the size and resolution of input datasets. Data-driven, cloud-based processing holds promise for predictive watershed modeling at a wide range of spatio-temporal scales. However, these benefits come with constraints. In particular, parallel computing limits a modeler's ability to simulate the flow of water across a landscape, rendering traditional routing algorithms unusable in this platform. Our project pushes these limits by testing the performance of two machine learning algorithms, Support Vector Machine (SVM) and Random Forests, at predicting flood extent. Constructed in Google Earth Engine, the model mines a suite of publicly available satellite imagery layers to use as algorithm inputs. Results are cross-validated using MODIS-based flood maps created using the Dartmouth Flood Observatory detection algorithm. Model uncertainty highlights the difficulty of deploying unbalanced training data sets based on rare extreme events.
DOE Office of Scientific and Technical Information (OSTI.GOV)
White, E. R., E-mail: ewhite@physics.ucla.edu; Kerelsky, Alexander; Hubbard, William A.
2015-11-30
Heterostructure devices with specific and extraordinary properties can be fabricated by stacking two-dimensional crystals. Cleanliness at the inter-crystal interfaces within a heterostructure is crucial for maximizing device performance. However, because these interfaces are buried, characterizing their impact on device function is challenging. Here, we show that electron-beam induced current (EBIC) mapping can be used to image interfacial contamination and to characterize the quality of buried heterostructure interfaces with nanometer-scale spatial resolution. We applied EBIC and photocurrent imaging to map photo-sensitive graphene-MoS{sub 2} heterostructures. The EBIC maps, together with concurrently acquired scanning transmission electron microscopy images, reveal how a device's photocurrentmore » collection efficiency is adversely affected by nanoscale debris invisible to optical-resolution photocurrent mapping.« less
Caching strategies for improving performance of web-based Geographic applications
NASA Astrophysics Data System (ADS)
Liu, M.; Brodzik, M.; Collins, J. A.; Lewis, S.; Oldenburg, J.
2012-12-01
The NASA Operation IceBridge mission collects airborne remote sensing measurements to bridge the gap between NASA's Ice, Cloud and Land Elevation Satellite (ICESat) mission and the upcoming ICESat-2 mission. The IceBridge Data Portal from the National Snow and Ice Data Center provides an intuitive web interface for accessing IceBridge mission observations and measurements. Scientists and users usually do not have knowledge about the individual campaigns but are interested in data collected in a specific place. We have developed a high-performance map interface to allow users to quickly zoom to an area of interest and see any Operation IceBridge overflights. The map interface consists of two layers: the user can pan and zoom on the base map layer; the flight line layer that overlays the base layer provides all the campaign missions that intersect with the current map view. The user can click on the flight campaigns and download the data as needed. The OpenGIS® Web Map Service Interface Standard (WMS) provides a simple HTTP interface for requesting geo-registered map images from one or more distributed geospatial databases. Web Feature Service (WFS) provides an interface allowing requests for geographical features across the web using platform-independent calls. OpenLayers provides vector support (points, polylines and polygons) to build a WMS/WFS client for displaying both layers on the screen. Map Server, an open source development environment for building spatially enabled internet applications, is serving the WMS and WFS spatial data to OpenLayers. Early releases of the portal displayed unacceptably poor load time performance for flight lines and the base map tiles. This issue was caused by long response times from the map server in generating all map tiles and flight line vectors. We resolved the issue by implementing various caching strategies on top of the WMS and WFS services, including the use of Squid (www.squid-cache.org) to cache frequently-used content. Our presentation includes the architectural design of the application, and how we use OpenLayers, WMS and WFS with Squid to build a responsive web application capable of efficiently displaying geospatial data to allow the user to quickly interact with the displayed information. We describe the design, implementation and performance improvement of our caching strategies, and the tools and techniques developed to assist our data caching strategies.
BrainIACS: a system for web-based medical image processing
NASA Astrophysics Data System (ADS)
Kishore, Bhaskar; Bazin, Pierre-Louis; Pham, Dzung L.
2009-02-01
We describe BrainIACS, a web-based medical image processing system that permits and facilitates algorithm developers to quickly create extensible user interfaces for their algorithms. Designed to address the challenges faced by algorithm developers in providing user-friendly graphical interfaces, BrainIACS is completely implemented using freely available, open-source software. The system, which is based on a client-server architecture, utilizes an AJAX front-end written using the Google Web Toolkit (GWT) and Java Servlets running on Apache Tomcat as its back-end. To enable developers to quickly and simply create user interfaces for configuring their algorithms, the interfaces are described using XML and are parsed by our system to create the corresponding user interface elements. Most of the commonly found elements such as check boxes, drop down lists, input boxes, radio buttons, tab panels and group boxes are supported. Some elements such as the input box support input validation. Changes to the user interface such as addition and deletion of elements are performed by editing the XML file or by using the system's user interface creator. In addition to user interface generation, the system also provides its own interfaces for data transfer, previewing of input and output files, and algorithm queuing. As the system is programmed using Java (and finally Java-script after compilation of the front-end code), it is platform independent with the only requirements being that a Servlet implementation be available and that the processing algorithms can execute on the server platform.
iAnn: an event sharing platform for the life sciences.
Jimenez, Rafael C; Albar, Juan P; Bhak, Jong; Blatter, Marie-Claude; Blicher, Thomas; Brazas, Michelle D; Brooksbank, Cath; Budd, Aidan; De Las Rivas, Javier; Dreyer, Jacqueline; van Driel, Marc A; Dunn, Michael J; Fernandes, Pedro L; van Gelder, Celia W G; Hermjakob, Henning; Ioannidis, Vassilios; Judge, David P; Kahlem, Pascal; Korpelainen, Eija; Kraus, Hans-Joachim; Loveland, Jane; Mayer, Christine; McDowall, Jennifer; Moran, Federico; Mulder, Nicola; Nyronen, Tommi; Rother, Kristian; Salazar, Gustavo A; Schneider, Reinhard; Via, Allegra; Villaveces, Jose M; Yu, Ping; Schneider, Maria V; Attwood, Teresa K; Corpas, Manuel
2013-08-01
We present iAnn, an open source community-driven platform for dissemination of life science events, such as courses, conferences and workshops. iAnn allows automatic visualisation and integration of customised event reports. A central repository lies at the core of the platform: curators add submitted events, and these are subsequently accessed via web services. Thus, once an iAnn widget is incorporated into a website, it permanently shows timely relevant information as if it were native to the remote site. At the same time, announcements submitted to the repository are automatically disseminated to all portals that query the system. To facilitate the visualization of announcements, iAnn provides powerful filtering options and views, integrated in Google Maps and Google Calendar. All iAnn widgets are freely available. http://iann.pro/iannviewer manuel.corpas@tgac.ac.uk.
QuakeSim: a Web Service Environment for Productive Investigations with Earth Surface Sensor Data
NASA Astrophysics Data System (ADS)
Parker, J. W.; Donnellan, A.; Granat, R. A.; Lyzenga, G. A.; Glasscoe, M. T.; McLeod, D.; Al-Ghanmi, R.; Pierce, M.; Fox, G.; Grant Ludwig, L.; Rundle, J. B.
2011-12-01
The QuakeSim science gateway environment includes a visually rich portal interface, web service access to data and data processing operations, and the QuakeTables ontology-based database of fault models and sensor data. The integrated tools and services are designed to assist investigators by covering the entire earthquake cycle of strain accumulation and release. The Web interface now includes Drupal-based access to diverse and changing content, with new ability to access data and data processing directly from the public page, as well as the traditional project management areas that require password access. The system is designed to make initial browsing of fault models and deformation data particularly engaging for new users. Popular data and data processing include GPS time series with data mining techniques to find anomalies in time and space, experimental forecasting methods based on catalogue seismicity, faulted deformation models (both half-space and finite element), and model-based inversion of sensor data. The fault models include the CGS and UCERF 2.0 faults of California and are easily augmented with self-consistent fault models from other regions. The QuakeTables deformation data include the comprehensive set of UAVSAR interferograms as well as a growing collection of satellite InSAR data.. Fault interaction simulations are also being incorporated in the web environment based on Virtual California. A sample usage scenario is presented which follows an investigation of UAVSAR data from viewing as an overlay in Google Maps, to selection of an area of interest via a polygon tool, to fast extraction of the relevant correlation and phase information from large data files, to a model inversion of fault slip followed by calculation and display of a synthetic model interferogram.
Alabama Public Scoping Meeting | NOAA Gulf Spill Restoration
: Mobile, AL Start Time: 6:30 p.m. Central Time Description: As part of the public scoping process, the co open at 6:30 p.m. and the meeting will begin at 7:30 p.m. Location: The Battle House Renaissance Mobile Hotel & Spa 26 North Royal Street Mobile, AL 36602 (google map of location) Gulf Spill Restoration
Tactical Level Commander and Staff Toolkit
2010-01-01
Sites Geodata.gov (for maps) http://gos2.geodata.gov Google Earth for .mil (United States Army Corps of Engineers (USACE) site) https...the eyes, ears, head, hands, back, and feet. When appropriate, personnel should wear protective lenses, goggles, or face shields . Leaders should...Typical hurricanes are about 300 miles wide, although they can vary considerably. Size is not necessarily an indication of hurricane intensity. The
Tan, Tan-Hsu; Gochoo, Munkhjargal; Chen, Yung-Fu; Hu, Jin-Jia; Chiang, John Y.; Chang, Ching-Su; Lee, Ming-Huei; Hsu, Yung-Nian; Hsu, Jiin-Chyr
2017-01-01
This study presents a new ubiquitous emergency medical service system (UEMS) that consists of a ubiquitous tele-diagnosis interface and a traffic guiding subsystem. The UEMS addresses unresolved issues of emergency medical services by managing the sensor wires for eliminating inconvenience for both patients and paramedics in an ambulance, providing ubiquitous accessibility of patients’ biosignals in remote areas where the ambulance cannot arrive directly, and offering availability of real-time traffic information which can make the ambulance reach the destination within the shortest time. In the proposed system, patient’s biosignals and real-time video, acquired by wireless biosensors and a webcam, can be simultaneously transmitted to an emergency room for pre-hospital treatment via WiMax/3.5 G networks. Performances of WiMax and 3.5 G, in terms of initialization time, data rate, and average end-to-end delay are evaluated and compared. A driver can choose the route of the shortest time among the suggested routes by Google Maps after inspecting the current traffic conditions based on real-time CCTV camera streams and traffic information. The destination address can be inputted vocally for easiness and safety in driving. A series of field test results validates the feasibility of the proposed system for application in real-life scenarios. PMID:28117724
Tan, Tan-Hsu; Gochoo, Munkhjargal; Chen, Yung-Fu; Hu, Jin-Jia; Chiang, John Y; Chang, Ching-Su; Lee, Ming-Huei; Hsu, Yung-Nian; Hsu, Jiin-Chyr
2017-01-21
This study presents a new ubiquitous emergency medical service system (UEMS) that consists of a ubiquitous tele-diagnosis interface and a traffic guiding subsystem. The UEMS addresses unresolved issues of emergency medical services by managing the sensor wires for eliminating inconvenience for both patients and paramedics in an ambulance, providing ubiquitous accessibility of patients' biosignals in remote areas where the ambulance cannot arrive directly, and offering availability of real-time traffic information which can make the ambulance reach the destination within the shortest time. In the proposed system, patient's biosignals and real-time video, acquired by wireless biosensors and a webcam, can be simultaneously transmitted to an emergency room for pre-hospital treatment via WiMax/3.5 G networks. Performances of WiMax and 3.5 G, in terms of initialization time, data rate, and average end-to-end delay are evaluated and compared. A driver can choose the route of the shortest time among the suggested routes by Google Maps after inspecting the current traffic conditions based on real-time CCTV camera streams and traffic information. The destination address can be inputted vocally for easiness and safety in driving. A series of field test results validates the feasibility of the proposed system for application in real-life scenarios.
Design of Deformation Monitoring System for Volcano Mitigation
NASA Astrophysics Data System (ADS)
Islamy, M. R. F.; Salam, R. A.; Munir, M. M.; Irsyam, M.; Khairurrijal
2016-08-01
Indonesia has many active volcanoes that are potentially disastrous. It needs good mitigation systems to prevent victims and to reduce casualties from potential disaster caused by volcanoes eruption. Therefore, the system to monitor the deformation of volcano was built. This system employed telemetry with the combination of Radio Frequency (RF) communications of XBEE and General Packet Radio Service (GPRS) communication of SIM900. There are two types of modules in this system, first is the coordinator as a parent and second is the node as a child. Each node was connected to coordinator forming a Wireless Sensor Network (WSN) with a star topology and it has an inclinometer based sensor, a Global Positioning System (GPS), and an XBEE module. The coordinator collects data to each node, one a time, to prevent collision data between nodes, save data to SD Card and transmit data to web server via GPRS. Inclinometer was calibrated with self-built in calibrator and tested in high temperature environment to check the durability. The GPS was tested by displaying its position in web server via Google Map Application Protocol Interface (API v.3). It was shown that the coordinator can receive and transmit data from every node to web server very well and the system works well in a high temperature environment.
The Use of LANCE Imagery Products to Investigate Hazards and Disasters
NASA Astrophysics Data System (ADS)
Schmaltz, J. E.; Teague, M.; Conover, H.; Regner, K.; Masuoka, E.; Vollmer, B. E.; Durbin, P.; Murphy, K. J.; Boller, R. A.; Davies, D.; Ilavajhala, S.; Thompson, C. K.; Bingham, A.; Rao, S.
2011-12-01
The NASA/GSFC Land Atmospheres Near-real time Capability for EOS (LANCE) has endeavored to integrate a variety of products from the Terra, Aqua, and Aura missions to assist in meeting the needs of the applications user community. This community has a need for imagery products to support the investigation of a wide variety of phenomena including hazards and disasters. The Evjafjallajokull eruption, the tsunamis/flood in Japan, and the Gulf of Mexico oil spill are recent examples of applications benefiting from the timely and synoptic view afforded by LANCE data. Working with the instrument science teams and the applications community, LANCE has identified 14 applications categories and the LANCE products that will support their investigation. The categories are: Smoke Plumes, Ash Plumes, Dust Storms, Pollution, Severe Storms, Shipping hazards, Fishery hazards, Land Transportation, Fires, Floods, Drought, Vegetation, Agriculture, and Oil Spills. Forty products from AMSR-E, MODIS, AIRS, and OMI have been identified to support analyses and investigations of these phenomena. In each case multiple products from two or more instruments are available which gives a more complete picture of the evolving hazard or disaster. All Level 2 (L2) products are available within 2.5 hours of observation at the spacecraft and the daily L3 products are updated incrementally as new data become available. LANCE provides user access to imagery using two systems: a Web Mapping Service (WMS) and a Google Earth-based interface known as the State of the Earth (SOTE). The latter has resulted from a partnership between LANCE and the Physical Oceanography Distributed Active Archive Center (PO DAAC). When the user selects one of the 14 categories, the relevant products are established within the WMS (http://lance2.modaps.eosdis.nasa.gov/wms/). For each application, population density data are available for densities in excess of 100 people/sqkm with user-defined opacity. These data are provided by the EOSDIS Socio-Economic Data and Applications Center (SEDAC). Certain users may not want to be constrained by the pre-defined categories and related products and all 40 products may be added as potential overlays. The most recent 10 days of near-real time data are available through the WMS. The SOTE provides an interface to the products grouped in the same fashion as the WMS. The SOTE servers stream imagery and data in the OGC KML format and these feeds can be visualized through the Google Earth browser plug-in. SOTE provides visualization through a virtual globe environment by allowing users to interact with the globe via zooming, rotating, and tilting.
The UK Soil Observatory (UKSO) and mySoil app: crowdsourcing and disseminating soil information.
NASA Astrophysics Data System (ADS)
Robinson, David; Bell, Patrick; Emmett, Bridget; Panagos, Panos; Lawley, Russell; Shelley, Wayne
2017-04-01
Digital technologies in terms of web based data portals and mobiles apps offer a new way to provide both information to the public, and to engage the public in becoming involved in contributing to the effort of collecting data through crowdsourcing. We are part of the Landpotential.org consortium which is a global partnership committed to developing and supporting the adoption of freely available technology and tools for sustainable land use management, monitoring, and connecting people across the globe. The mySoil app was launched in 2012 and is an example of a free mobile application downloadable from iTunes and Google Play. It serves as a gateway tool to raise interest in, and awareness of, soils. It currently has over 50,000 dedicated users and has crowd sourced more than 4000 data records. Recent developments have expanded the coverage of mySoil from the United Kingdom to Europe, introduced a new user interface and provided language capability, while the UKSO displays the crowd-sourced records from across the globe. We are now trying to identify which industry, education and citizen sectors are using these platforms and how they can be improved. Please help us by providing feedback or taking the survey on the UKSO website. www.UKSO.org The UKSO is a collaboration between major UK soil-data holders to provide maps, spatial data and real-time temporal data from observing platforms such as the UK soil moisture network. Both UKSO and mySoil have crowdsourcing capability and are slowly building global citizen science maps of soil properties such as pH and texture. Whilst these data can't replace professional monitoring data, the information they provide both stimulates public interest and can act as 'soft data' that can help support the interpretation of monitoring data, or guide future monitoring, identifying areas that don't correspond with current analysis. In addition, soft data can be used to map soils with machine learning approaches, such as SoilGrids.
Jung, Younbo; Park, Namkee; Lee, Kwan Min
2015-12-01
This study investigated the effects of trait-level hostility, interface types, and character identification on aggressive thoughts and overall game experience after playing a violent video game. Results showed that the mapping interface made participants with high trait-level hostility more readily accessible to aggressive contracts, yet it did not have any significant impact for participants with low trait-level hostility. Participants with low trait-level hostility reported more positive game experience in the mapping interface condition, while participants with high trait-level hostility in the same condition reported more negative game experience. Results also indicated that character identification has moderating effects on activating aggressive thoughts and mediating effects on overall game experience. Implications regarding possible ways of reducing potentially negative outcomes from violent games are discussed.
Experimental setup for evaluating an adaptive user interface for teleoperation control
NASA Astrophysics Data System (ADS)
Wijayasinghe, Indika B.; Peetha, Srikanth; Abubakar, Shamsudeen; Saadatzi, Mohammad Nasser; Cremer, Sven; Popa, Dan O.
2017-05-01
A vital part of human interactions with a machine is the control interface, which single-handedly could define the user satisfaction and the efficiency of performing a task. This paper elaborates the implementation of an experimental setup to study an adaptive algorithm that can help the user better tele-operate the robot. The formulation of the adaptive interface and associate learning algorithms are general enough to apply when the mapping between the user controls and the robot actuators is complex and/or ambiguous. The method uses a genetic algorithm to find the optimal parameters that produce the input-output mapping for teleoperation control. In this paper, we describe the experimental setup and associated results that was used to validate the adaptive interface to a differential drive robot from two different input devices; a joystick, and a Myo gesture control armband. Results show that after the learning phase, the interface converges to an intuitive mapping that can help even inexperienced users drive the system to a goal location.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Spurgeon, Steven R.; Du, Yingge; Chambers, Scott A.
Abstract With the development of affordable aberration correctors, analytical scanning transmission electron microscopy (STEM) studies of complex interfaces can now be conducted at high spatial resolution at laboratories worldwide. Energy-dispersive X-ray spectroscopy (EDS) in particular has grown in popularity, as it enables elemental mapping over a wide range of ionization energies. However, the interpretation of atomically resolved data is greatly complicated by beam–sample interactions that are often overlooked by novice users. Here we describe the practical factors—namely, sample thickness and the choice of ionization edge—that affect the quantification of a model perovskite oxide interface. Our measurements of the same sample,more » in regions of different thickness, indicate that interface profiles can vary by as much as 2–5 unit cells, depending on the spectral feature. This finding is supported by multislice simulations, which reveal that on-axis maps of even perfectly abrupt interfaces exhibit significant delocalization. Quantification of thicker samples is further complicated by channeling to heavier sites across the interface, as well as an increased signal background. We show that extreme care must be taken to prepare samples to minimize channeling effects and argue that it may not be possible to extract atomically resolved information from many chemical maps.« less
Spurgeon, Steven R.; Du, Yingge; Chambers, Scott A.
2017-04-05
Abstract With the development of affordable aberration correctors, analytical scanning transmission electron microscopy (STEM) studies of complex interfaces can now be conducted at high spatial resolution at laboratories worldwide. Energy-dispersive X-ray spectroscopy (EDS) in particular has grown in popularity, as it enables elemental mapping over a wide range of ionization energies. However, the interpretation of atomically resolved data is greatly complicated by beam–sample interactions that are often overlooked by novice users. Here we describe the practical factors—namely, sample thickness and the choice of ionization edge—that affect the quantification of a model perovskite oxide interface. Our measurements of the same sample,more » in regions of different thickness, indicate that interface profiles can vary by as much as 2–5 unit cells, depending on the spectral feature. This finding is supported by multislice simulations, which reveal that on-axis maps of even perfectly abrupt interfaces exhibit significant delocalization. Quantification of thicker samples is further complicated by channeling to heavier sites across the interface, as well as an increased signal background. We show that extreme care must be taken to prepare samples to minimize channeling effects and argue that it may not be possible to extract atomically resolved information from many chemical maps.« less
IntegratedMap: a Web interface for integrating genetic map data.
Yang, Hongyu; Wang, Hongyu; Gingle, Alan R
2005-05-01
IntegratedMap is a Web application and database schema for storing and interactively displaying genetic map data. Its Web interface includes a menu for direct chromosome/linkage group selection, a search form for selection based on mapped object location and linkage group displays. An overview display provides convenient access to the full range of mapped and anchored object types with genetic locus details, such as numbers, types and names of mapped/anchored objects displayed in a compact scrollable list box that automatically updates based on selected map location and object type. Also, multilinkage group and localized map views are available along with links that can be configured for integration with other Web resources. IntegratedMap is implemented in C#/ASP.NET and the package, including a MySQL schema creation script, is available from http://cggc.agtec.uga.edu/Data/download.asp
Wildland-urban interface maps vary with purpose and context
Stewart, S.I.; Wilmer, B.; Hammer, R.B.; Aplet, G.H.; Hawbaker, T.J.; Miller, C.; Radeloff, V.C.
2009-01-01
Maps of the wildland-urban interface (WUI) are both policy tools and powerful visual images. Although the growing number of WUI maps serve similar purposes, this article indicates that WUI maps derived from the same data sets can differ in important ways related to their original intended application. We discuss the use of ancillary data in modifying census data to improve WUI maps and offer a cautionary note about this practice. A comparison of two WUI mapping approaches suggests that no single map is "best" because users' needs vary. The analysts who create maps are responsible for ensuring that users understand their purpose, data, and methods; map users are responsible for paying attention to these features and using each map accordingly. These considerations should apply to any analysis but are especially important to analyses of the WUI on which policy decisions will be made.
Utility of Mobile phones to support In-situ data collection for Land Cover Mapping
NASA Astrophysics Data System (ADS)
Oduor, P.; Omondi, S.; Wahome, A.; Mugo, R. M.; Flores, A.
2017-12-01
With the compelling need to create better monitoring tools for our landscapes to enhance better decision making processes, it becomes imperative to do so in much more sophisticated yet simple ways. Making it possible to leverage untapped potential of our "lay men" at the same time enabling us to respond to the complexity of the information we have to get out. SERVIR Eastern and Southern Africa has developed a mobile app that can be utilized with very little prior knowledge or no knowledge at all to collect spatial information on land cover. This set of in-situ data can be collected by masses because the tools is very simple to use, and have this information fed in classification algorithms than can then be used to map out our ever changing landscape. The LULC Mapper is a subset of JiMap system and is able to pull the google earth imagery and open street maps to enable user familiarize with their location. It uses phone GPS, phone network information to map location coordinates and at the same time gives the user sample picture of what to categorize their landscape. The system is able to work offline and when user gets access to internet they can push the information into an amazon database as bulk data. The location details including geotagged photos allows the data to be used in development of a lot of spatial information including land cover data. The app is currently available in Google Play Store and will soon be uploaded on Appstore for utilization by a wider community. We foresee a lot of potential in this tool in terms of making data collection cheaper and affordable. Taking advantage of the advances made in phone technology. We envisage to do a data collection campaign where we can have the tool used for crowdsourcing.
Using Public Input to Create a Better Online Flood Mapping Framework
NASA Astrophysics Data System (ADS)
Eubanks, K. E.; Jackson, C.; Carlberg, B.; Cohen, S.
2017-12-01
One topic of consistent relevance in flooding research is how best to provide information and communicate risk from scientists and researchers to the general public. Additionally, communicators face challenges on how to fully convey the dangers flooding poses in a manner that the public comprehends and will apply to reactions to flooding. Many of the inundation and hazard maps currently in use are highly technical, making it difficult for the average person, without formal education in flooding, to glean valuable information and insight from the intended tools. Working with the public, a set of three surveys were administered via social media to gain insight into public understanding of floods and flooding risk. The surveys indicated that the general population does not have a firm understanding of basic flooding terms or how to navigate current, technical flood inundation maps. The surveys also suggested that those surveyed desire a simpler interface for flood maps that also relates a sense of varying risk. Using the feedback from each survey, a conceptual framework was produced for a set of inundation maps, including more relatable terms and educational components within a user-friendly web interface. Goals for the website, shaped by survey feedback, included simple, readable map layers that convey a sense of uncertainty, a clear and detailed legend, the ability show or hide components of the map, and the option to learn more about flood terminology on the site or via links to outside resources. The public indicated that the final map interface was more concise and simplified than the current inundation map platforms they navigated as part of the first survey, and that the proposed interface was overall more likely to be used. Using public input is one way to bridge the gap between scientific data and predictions to the general public, who need this information. It is vital to provide accurate data in a form that is relatable, and therefore helpful, to the members of the community trying to make educated decisions. The findings on gearing inundation map web interfaces to the public are being used to create tools that are more usable, therefore hopefully saving lives by better informing those in danger of their risk.
Using the Browser for Science: A Collaborative Toolkit for Astronomy
NASA Astrophysics Data System (ADS)
Connolly, A. J.; Smith, I.; Krughoff, K. S.; Gibson, R.
2011-07-01
Astronomical surveys have yielded hundreds of terabytes of catalogs and images that span many decades of the electromagnetic spectrum. Even when observatories provide user-friendly web interfaces, exploring these data resources remains a complex and daunting task. In contrast, gadgets and widgets have become popular in social networking (e.g. iGoogle, Facebook). They provide a simple way to make complex data easily accessible that can be customized based on the interest of the user. With ASCOT (an AStronomical COllaborative Toolkit) we expand on these concepts to provide a customizable and extensible gadget framework for use in science. Unlike iGoogle, where all of the gadgets are independent, the gadgets we develop communicate and share information, enabling users to visualize and interact with data through multiple, simultaneous views. With this approach, web-based applications for accessing and visualizing data can be generated easily and, by linking these tools together, integrated and powerful data analysis and discovery tools can be constructed.
FwWebViewPlus: integration of web technologies into WinCC OA based Human-Machine Interfaces at CERN
NASA Astrophysics Data System (ADS)
Golonka, Piotr; Fabian, Wojciech; Gonzalez-Berges, Manuel; Jasiun, Piotr; Varela-Rodriguez, Fernando
2014-06-01
The rapid growth in popularity of web applications gives rise to a plethora of reusable graphical components, such as Google Chart Tools and JQuery Sparklines, implemented in JavaScript and run inside a web browser. In the paper we describe the tool that allows for seamless integration of web-based widgets into WinCC Open Architecture, the SCADA system used commonly at CERN to build complex Human-Machine Interfaces. Reuse of widely available widget libraries and pushing the development efforts to a higher abstraction layer based on a scripting language allow for significant reduction in maintenance of the code in multi-platform environments compared to those currently used in C++ visualization plugins. Adequately designed interfaces allow for rapid integration of new web widgets into WinCC OA. At the same time, the mechanisms familiar to HMI developers are preserved, making the use of new widgets "native". Perspectives for further integration between the realms of WinCC OA and Web development are also discussed.
User’s guide for MapMark4GUI—A graphical user interface for the MapMark4 R package
Shapiro, Jason
2018-05-29
MapMark4GUI is an R graphical user interface (GUI) developed by the U.S. Geological Survey to support user implementation of the MapMark4 R statistical software package. MapMark4 was developed by the U.S. Geological Survey to implement probability calculations for simulating undiscovered mineral resources in quantitative mineral resource assessments. The GUI provides an easy-to-use tool to input data, run simulations, and format output results for the MapMark4 package. The GUI is written and accessed in the R statistical programming language. This user’s guide includes instructions on installing and running MapMark4GUI and descriptions of the statistical output processes, output files, and test data files.
High-Performance Signal Detection for Adverse Drug Events using MapReduce Paradigm.
Fan, Kai; Sun, Xingzhi; Tao, Ying; Xu, Linhao; Wang, Chen; Mao, Xianling; Peng, Bo; Pan, Yue
2010-11-13
Post-marketing pharmacovigilance is important for public health, as many Adverse Drug Events (ADEs) are unknown when those drugs were approved for marketing. However, due to the large number of reported drugs and drug combinations, detecting ADE signals by mining these reports is becoming a challenging task in terms of computational complexity. Recently, a parallel programming model, MapReduce has been introduced by Google to support large-scale data intensive applications. In this study, we proposed a MapReduce-based algorithm, for common ADE detection approach, Proportional Reporting Ratio (PRR), and tested it in mining spontaneous ADE reports from FDA. The purpose is to investigate the possibility of using MapReduce principle to speed up biomedical data mining tasks using this pharmacovigilance case as one specific example. The results demonstrated that MapReduce programming model could improve the performance of common signal detection algorithm for pharmacovigilance in a distributed computation environment at approximately liner speedup rates.
Affective Interface Adaptations in the Musickiosk Interactive Entertainment Application
NASA Astrophysics Data System (ADS)
Malatesta, L.; Raouzaiou, A.; Pearce, L.; Karpouzis, K.
The current work presents the affective interface adaptations in the Musickiosk application. Adaptive interaction poses several open questions since there is no unique way of mapping affective factors of user behaviour to the output of the system. Musickiosk uses a non-contact interface and implicit interaction through emotional affect rather than explicit interaction where a gesture, sound or other input directly maps to an output behaviour - as in traditional entertainment applications. PAD model is used for characterizing the different affective states and emotions.
Extensible Probabilistic Repository Technology (XPRT)
2004-10-01
projects, such as, Centaurus , Evidence Data Base (EDB), etc., others were fabricated, such as INS and FED, while others contain data from the open...Google Web Report Unlimited SOAP API News BBC News Unlimited WEB RSS 1.0 Centaurus Person Demographics 204,402 people from 240 countries...objects of the domain ontology map to the various simulated data-sources. For example, the PersonDemographics are stored in the Centaurus database, while
Reaching Forward in the War against the Islamic State
2016-12-07
every week with U.S.- and Coalition-advised ISOF troops taking the lead in combat operations using cellular communications systems that link them...tions—Offline Maps, Google Earth , and Viber, to name a few—which allowed them to bring tablets and phones on their operations to help communicate ...provided an initial Remote Advise and Assist capability that enabled the special forces advisors to track, communicate , and share limited data with
Sato, H.P.; Harp, E.L.
2009-01-01
The 12 May 2008 M7.9 Wenchuan earthquake in the People's Republic of China represented a unique opportunity for the international community to use commonly available GIS (Geographic Information System) tools, like Google Earth (GE), to rapidly evaluate and assess landslide hazards triggered by the destructive earthquake and its aftershocks. In order to map earthquake-triggered landslides, we provide details on the applicability and limitations of publicly available 3-day-post- and pre-earthquake imagery provided by GE from the FORMOSAT-2 (formerly ROCSAT-2; Republic of China Satellite 2). We interpreted landslides on the 8-m-resolution FORMOSAT-2 image by GE; as a result, 257 large landslides were mapped with the highest concentration along the Beichuan fault. An estimated density of 0.3 landslides/km2 represents a minimum bound on density given the resolution of available imagery; higher resolution data would have identified more landslides. This is a preliminary study, and further study is needed to understand the landslide characteristics in detail. Although it is best to obtain landslide locations and measurements from satellite imagery having high resolution, it was found that GE is an effective and rapid reconnaissance tool. ?? 2009 Springer-Verlag.
GeneOnEarth: fitting genetic PC plots on the globe.
Torres-Sánchez, Sergio; Medina-Medina, Nuria; Gignoux, Chris; Abad-Grau, María M; González-Burchard, Esteban
2013-01-01
Principal component (PC) plots have become widely used to summarize genetic variation of individuals in a sample. The similarity between genetic distance in PC plots and geographical distance has shown to be quite impressive. However, in most situations, individual ancestral origins are not precisely known or they are heterogeneously distributed; hence, they are hardly linked to a geographical area. We have developed GeneOnEarth, a user-friendly web-based tool to help geneticists to understand whether a linear isolation-by-distance model may apply to a genetic data set; thus, genetic distances among a set of individuals resemble geographical distances among their origins. Its main goal is to allow users to first apply a by-view Procrustes method to visually learn whether this model holds. To do that, the user can choose the exact geographical area from an on line 2D or 3D world map by using, respectively, Google Maps or Google Earth, and rotate, flip, and resize the images. GeneOnEarth can also compute the optimal rotation angle using Procrustes analysis and assess statistical evidence of similarity when a different rotation angle has been chosen by the user. An online version of GeneOnEarth is available for testing and using purposes at http://bios.ugr.es/GeneOnEarth.
NASA Astrophysics Data System (ADS)
Sareen, Sanjay; Gupta, Sunil Kumar; Sood, Sandeep K.
2017-10-01
Zika virus is a mosquito-borne disease that spreads very quickly in different parts of the world. In this article, we proposed a system to prevent and control the spread of Zika virus disease using integration of Fog computing, cloud computing, mobile phones and the Internet of things (IoT)-based sensor devices. Fog computing is used as an intermediary layer between the cloud and end users to reduce the latency time and extra communication cost that is usually found high in cloud-based systems. A fuzzy k-nearest neighbour is used to diagnose the possibly infected users, and Google map web service is used to provide the geographic positioning system (GPS)-based risk assessment to prevent the outbreak. It is used to represent each Zika virus (ZikaV)-infected user, mosquito-dense sites and breeding sites on the Google map that help the government healthcare authorities to control such risk-prone areas effectively and efficiently. The proposed system is deployed on Amazon EC2 cloud to evaluate its performance and accuracy using data set for 2 million users. Our system provides high accuracy of 94.5% for initial diagnosis of different users according to their symptoms and appropriate GPS-based risk assessment.
Debnath, Manish; Kharumnuid, Graciously; Thongnibah, Welfrank; Tandon, Veena
2016-01-01
Most metazoan parasites that invade vertebrate hosts belong to three phyla: Platyhelminthes, Nematoda and Acanthocephala. Many of the parasitic members of these phyla are collectively known as helminths and are causative agents of many debilitating, deforming and lethal diseases of humans and animals. The North-East India Helminth Parasite Information Database (NEIHPID) project aimed to document and characterise the spectrum of helminth parasites in the north-eastern region of India, providing host, geographical distribution, diagnostic characters and image data. The morphology-based taxonomic data are supplemented with information on DNA sequences of nuclear, ribosomal and mitochondrial gene marker regions that aid in parasite identification. In addition, the database contains raw next generation sequencing (NGS) data for 3 foodborne trematode parasites, with more to follow. The database will also provide study material for students interested in parasite biology. Users can search the database at various taxonomic levels (phylum, class, order, superfamily, family, genus, and species), or by host, habitat and geographical location. Specimen collection locations are noted as co-ordinates in a MySQL database and can be viewed on Google maps, using Google Maps JavaScript API v3. The NEIHPID database has been made freely available at http://nepiac.nehu.ac.in/index.php PMID:27285615
An Interactive Web System for Field Data Sharing and Collaboration
NASA Astrophysics Data System (ADS)
Weng, Y.; Sun, F.; Grigsby, J. D.
2010-12-01
A Web 2.0 system is designed and developed to facilitate data collection for the field studies in the Geological Sciences department at Ball State University. The system provides a student-centered learning platform that enables the users to first upload their collected data in various formats, interact and collaborate dynamically online, and ultimately create a shared digital repository of field experiences. The data types considered for the system and their corresponding format and requirements are listed in the table below. The system has six main functionalities as follows. (1) Only the registered users can access the system with confidential identification and password. (2) Each user can upload/revise/delete data in various formats such as image, audio, video, and text files to the system. (3) Interested users are allowed to co-edit the contents and join the collaboration whiteboard for further discussion. (4) The system integrates with Google, Yahoo, or Flickr to search for similar photos with same tags. (5) Users can search the web system according to the specific key words. (6) Photos with recorded GPS readings can be mashed and mapped to Google Maps/Earth for visualization. Application of the system to geology field trips at Ball State University will be demonstrated to assess the usability of the system.Data Requirements
Biswal, Devendra Kumar; Debnath, Manish; Kharumnuid, Graciously; Thongnibah, Welfrank; Tandon, Veena
2016-01-01
Most metazoan parasites that invade vertebrate hosts belong to three phyla: Platyhelminthes, Nematoda and Acanthocephala. Many of the parasitic members of these phyla are collectively known as helminths and are causative agents of many debilitating, deforming and lethal diseases of humans and animals. The North-East India Helminth Parasite Information Database (NEIHPID) project aimed to document and characterise the spectrum of helminth parasites in the north-eastern region of India, providing host, geographical distribution, diagnostic characters and image data. The morphology-based taxonomic data are supplemented with information on DNA sequences of nuclear, ribosomal and mitochondrial gene marker regions that aid in parasite identification. In addition, the database contains raw next generation sequencing (NGS) data for 3 foodborne trematode parasites, with more to follow. The database will also provide study material for students interested in parasite biology. Users can search the database at various taxonomic levels (phylum, class, order, superfamily, family, genus, and species), or by host, habitat and geographical location. Specimen collection locations are noted as co-ordinates in a MySQL database and can be viewed on Google maps, using Google Maps JavaScript API v3. The NEIHPID database has been made freely available at http://nepiac.nehu.ac.in/index.php.
High-Resolution Air Pollution Mapping with Google Street View Cars: Exploiting Big Data.
Apte, Joshua S; Messier, Kyle P; Gani, Shahzad; Brauer, Michael; Kirchstetter, Thomas W; Lunden, Melissa M; Marshall, Julian D; Portier, Christopher J; Vermeulen, Roel C H; Hamburg, Steven P
2017-06-20
Air pollution affects billions of people worldwide, yet ambient pollution measurements are limited for much of the world. Urban air pollution concentrations vary sharply over short distances (≪1 km) owing to unevenly distributed emission sources, dilution, and physicochemical transformations. Accordingly, even where present, conventional fixed-site pollution monitoring methods lack the spatial resolution needed to characterize heterogeneous human exposures and localized pollution hotspots. Here, we demonstrate a measurement approach to reveal urban air pollution patterns at 4-5 orders of magnitude greater spatial precision than possible with current central-site ambient monitoring. We equipped Google Street View vehicles with a fast-response pollution measurement platform and repeatedly sampled every street in a 30-km 2 area of Oakland, CA, developing the largest urban air quality data set of its type. Resulting maps of annual daytime NO, NO 2 , and black carbon at 30 m-scale reveal stable, persistent pollution patterns with surprisingly sharp small-scale variability attributable to local sources, up to 5-8× within individual city blocks. Since local variation in air quality profoundly impacts public health and environmental equity, our results have important implications for how air pollution is measured and managed. If validated elsewhere, this readily scalable measurement approach could address major air quality data gaps worldwide.
Planetary Data Systems (PDS) Imaging Node Atlas II
NASA Technical Reports Server (NTRS)
Stanboli, Alice; McAuley, James M.
2013-01-01
The Planetary Image Atlas (PIA) is a Rich Internet Application (RIA) that serves planetary imaging data to the science community and the general public. PIA also utilizes the USGS Unified Planetary Coordinate system (UPC) and the on-Mars map server. The Atlas was designed to provide the ability to search and filter through greater than 8 million planetary image files. This software is a three-tier Web application that contains a search engine backend (MySQL, JAVA), Web service interface (SOAP) between server and client, and a GWT Google Maps API client front end. This application allows for the search, retrieval, and download of planetary images and associated meta-data from the following missions: 2001 Mars Odyssey, Cassini, Galileo, LCROSS, Lunar Reconnaissance Orbiter, Mars Exploration Rover, Mars Express, Magellan, Mars Global Surveyor, Mars Pathfinder, Mars Reconnaissance Orbiter, MESSENGER, Phoe nix, Viking Lander, Viking Orbiter, and Voyager. The Atlas utilizes the UPC to translate mission-specific coordinate systems into a unified coordinate system, allowing the end user to query across missions of similar targets. If desired, the end user can also use a mission-specific view of the Atlas. The mission-specific views rely on the same code base. This application is a major improvement over the initial version of the Planetary Image Atlas. It is a multi-mission search engine. This tool includes both basic and advanced search capabilities, providing a product search tool to interrogate the collection of planetary images. This tool lets the end user query information about each image, and ignores the data that the user has no interest in. Users can reduce the number of images to look at by defining an area of interest with latitude and longitude ranges.
Côté, Richard G; Jones, Philip; Martens, Lennart; Kerrien, Samuel; Reisinger, Florian; Lin, Quan; Leinonen, Rasko; Apweiler, Rolf; Hermjakob, Henning
2007-10-18
Each major protein database uses its own conventions when assigning protein identifiers. Resolving the various, potentially unstable, identifiers that refer to identical proteins is a major challenge. This is a common problem when attempting to unify datasets that have been annotated with proteins from multiple data sources or querying data providers with one flavour of protein identifiers when the source database uses another. Partial solutions for protein identifier mapping exist but they are limited to specific species or techniques and to a very small number of databases. As a result, we have not found a solution that is generic enough and broad enough in mapping scope to suit our needs. We have created the Protein Identifier Cross-Reference (PICR) service, a web application that provides interactive and programmatic (SOAP and REST) access to a mapping algorithm that uses the UniProt Archive (UniParc) as a data warehouse to offer protein cross-references based on 100% sequence identity to proteins from over 70 distinct source databases loaded into UniParc. Mappings can be limited by source database, taxonomic ID and activity status in the source database. Users can copy/paste or upload files containing protein identifiers or sequences in FASTA format to obtain mappings using the interactive interface. Search results can be viewed in simple or detailed HTML tables or downloaded as comma-separated values (CSV) or Microsoft Excel (XLS) files suitable for use in a local database or a spreadsheet. Alternatively, a SOAP interface is available to integrate PICR functionality in other applications, as is a lightweight REST interface. We offer a publicly available service that can interactively map protein identifiers and protein sequences to the majority of commonly used protein databases. Programmatic access is available through a standards-compliant SOAP interface or a lightweight REST interface. The PICR interface, documentation and code examples are available at http://www.ebi.ac.uk/Tools/picr.
Côté, Richard G; Jones, Philip; Martens, Lennart; Kerrien, Samuel; Reisinger, Florian; Lin, Quan; Leinonen, Rasko; Apweiler, Rolf; Hermjakob, Henning
2007-01-01
Background Each major protein database uses its own conventions when assigning protein identifiers. Resolving the various, potentially unstable, identifiers that refer to identical proteins is a major challenge. This is a common problem when attempting to unify datasets that have been annotated with proteins from multiple data sources or querying data providers with one flavour of protein identifiers when the source database uses another. Partial solutions for protein identifier mapping exist but they are limited to specific species or techniques and to a very small number of databases. As a result, we have not found a solution that is generic enough and broad enough in mapping scope to suit our needs. Results We have created the Protein Identifier Cross-Reference (PICR) service, a web application that provides interactive and programmatic (SOAP and REST) access to a mapping algorithm that uses the UniProt Archive (UniParc) as a data warehouse to offer protein cross-references based on 100% sequence identity to proteins from over 70 distinct source databases loaded into UniParc. Mappings can be limited by source database, taxonomic ID and activity status in the source database. Users can copy/paste or upload files containing protein identifiers or sequences in FASTA format to obtain mappings using the interactive interface. Search results can be viewed in simple or detailed HTML tables or downloaded as comma-separated values (CSV) or Microsoft Excel (XLS) files suitable for use in a local database or a spreadsheet. Alternatively, a SOAP interface is available to integrate PICR functionality in other applications, as is a lightweight REST interface. Conclusion We offer a publicly available service that can interactively map protein identifiers and protein sequences to the majority of commonly used protein databases. Programmatic access is available through a standards-compliant SOAP interface or a lightweight REST interface. The PICR interface, documentation and code examples are available at . PMID:17945017
Mapping the Americanization of English in space and time
Gonçalves, Bruno; Loureiro-Porto, Lucía; Ramasco, José J.
2018-01-01
As global political preeminence gradually shifted from the United Kingdom to the United States, so did the capacity to culturally influence the rest of the world. In this work, we analyze how the world-wide varieties of written English are evolving. We study both the spatial and temporal variations of vocabulary and spelling of English using a large corpus of geolocated tweets and the Google Books datasets corresponding to books published in the US and the UK. The advantage of our approach is that we can address both standard written language (Google Books) and the more colloquial forms of microblogging messages (Twitter). We find that American English is the dominant form of English outside the UK and that its influence is felt even within the UK borders. Finally, we analyze how this trend has evolved over time and the impact that some cultural events have had in shaping it. PMID:29799872
Mapping the Americanization of English in space and time.
Gonçalves, Bruno; Loureiro-Porto, Lucía; Ramasco, José J; Sánchez, David
2018-01-01
As global political preeminence gradually shifted from the United Kingdom to the United States, so did the capacity to culturally influence the rest of the world. In this work, we analyze how the world-wide varieties of written English are evolving. We study both the spatial and temporal variations of vocabulary and spelling of English using a large corpus of geolocated tweets and the Google Books datasets corresponding to books published in the US and the UK. The advantage of our approach is that we can address both standard written language (Google Books) and the more colloquial forms of microblogging messages (Twitter). We find that American English is the dominant form of English outside the UK and that its influence is felt even within the UK borders. Finally, we analyze how this trend has evolved over time and the impact that some cultural events have had in shaping it.
Neuhaus, Philipp; Doods, Justin; Dugas, Martin
2015-01-01
Automatic coding of medical terms is an important, but highly complicated and laborious task. To compare and evaluate different strategies a framework with a standardized web-interface was created. Two UMLS mapping strategies are compared to demonstrate the interface. The framework is a Java Spring application running on a Tomcat application server. It accepts different parameters and returns results in JSON format. To demonstrate the framework, a list of medical data items was mapped by two different methods: similarity search in a large table of terminology codes versus search in a manually curated repository. These mappings were reviewed by a specialist. The evaluation shows that the framework is flexible (due to standardized interfaces like HTTP and JSON), performant and reliable. Accuracy of automatically assigned codes is limited (up to 40%). Combining different semantic mappers into a standardized Web-API is feasible. This framework can be easily enhanced due to its modular design.
Felfer, Peter; Cairney, Julie
2018-06-01
Analysing the distribution of selected chemical elements with respect to interfaces is one of the most common tasks in data mining in atom probe tomography. This can be represented by 1D concentration profiles, 2D concentration maps or proximity histograms, which represent concentration, density etc. of selected species as a function of the distance from a reference surface/interface. These are some of the most useful tools for the analysis of solute distributions in atom probe data. In this paper, we present extensions to the proximity histogram in the form of 'local' proximity histograms, calculated for selected parts of a surface, and pseudo-2D concentration maps, which are 2D concentration maps calculated on non-flat surfaces. This way, local concentration changes at interfaces or and other structures can be assessed more effectively. Copyright © 2018 Elsevier B.V. All rights reserved.
AlphaSpace: Fragment-Centric Topographical Mapping To Target Protein–Protein Interaction Interfaces
2016-01-01
Inhibition of protein–protein interactions (PPIs) is emerging as a promising therapeutic strategy despite the difficulty in targeting such interfaces with drug-like small molecules. PPIs generally feature large and flat binding surfaces as compared to typical drug targets. These features pose a challenge for structural characterization of the surface using geometry-based pocket-detection methods. An attractive mapping strategy—that builds on the principles of fragment-based drug discovery (FBDD)—is to detect the fragment-centric modularity at the protein surface and then characterize the large PPI interface as a set of localized, fragment-targetable interaction regions. Here, we introduce AlphaSpace, a computational analysis tool designed for fragment-centric topographical mapping (FCTM) of PPI interfaces. Our approach uses the alpha sphere construct, a geometric feature of a protein’s Voronoi diagram, to map out concave interaction space at the protein surface. We introduce two new features—alpha-atom and alpha-space—and the concept of the alpha-atom/alpha-space pair to rank pockets for fragment-targetability and to facilitate the evaluation of pocket/fragment complementarity. The resulting high-resolution interfacial map of targetable pocket space can be used to guide the rational design and optimization of small molecule or biomimetic PPI inhibitors. PMID:26225450
Assessment of a User Guide for One Semi-Automated Forces (OneSAF) Version 2.0
2009-09-01
OneSAF uses a two-dimensional feature named a Plan View Display ( PVD ) as the primary graphical interface. The PVD replicates a map with a series...primary interface, the PVD is how the user watches the scenario unfold and requires the most interaction with the user. As seen in Table 3, all...participant indicated never using these seven map-related functions. Graphic control measures. Graphic control measures are applied to the PVD map to
NASA Astrophysics Data System (ADS)
Tellman, B.; Sullivan, J.; Kettner, A.; Brakenridge, G. R.; Slayback, D. A.; Kuhn, C.; Doyle, C.
2016-12-01
There is an increasing need to understand flood vulnerability as the societal and economic effects of flooding increases. Risk models from insurance companies and flood models from hydrologists must be calibrated based on flood observations in order to make future predictions that can improve planning and help societies reduce future disasters. Specifically, to improve these models both traditional methods of flood prediction from physically based models as well as data-driven techniques, such as machine learning, require spatial flood observation to validate model outputs and quantify uncertainty. A key dataset that is missing for flood model validation is a global historical geo-database of flood event extents. Currently, the most advanced database of historical flood extent is hosted and maintained at the Dartmouth Flood Observatory (DFO) that has catalogued 4320 floods (1985-2015) but has only mapped 5% of these floods. We are addressing this data gap by mapping the inventory of floods in the DFO database to create a first-of- its-kind, comprehensive, global and historical geospatial database of flood events. To do so, we combine water detection algorithms on MODIS and Landsat 5,7 and 8 imagery in Google Earth Engine to map discrete flood events. The created database will be available in the Earth Engine Catalogue for download by country, region, or time period. This dataset can be leveraged for new data-driven hydrologic modeling using machine learning algorithms in Earth Engine's highly parallelized computing environment, and we will show examples for New York and Senegal.
Toubal, Abderrezak Kamel; Achite, Mohammed; Ouillon, Sylvain; Dehni, Abdelatif
2018-03-12
Soil losses must be quantified over watersheds in order to set up protection measures against erosion. The main objective of this paper is to quantify and to map soil losses in the Wadi Sahouat basin (2140 km 2 ) in the north-west of Algeria, using the Revised Universal Soil Loss Equation (RUSLE) model assisted by a Geographic Information System (GIS) and remote sensing. The Model Builder of the GIS allowed the automation of the different operations for establishing thematic layers of the model parameters: the erosivity factor (R), the erodibility factor (K), the topographic factor (LS), the crop management factor (C), and the conservation support practice factor (P). The average annual soil loss rate in the Wadi Sahouat basin ranges from 0 to 255 t ha -1 year -1 , maximum values being observed over steep slopes of more than 25% and between 600 and 1000 m elevations. 3.4% of the basin is classified as highly susceptible to erosion, 4.9% with a medium risk, and 91.6% at a low risk. Google Earth reveals a clear conformity with the degree of zones to erosion sensitivity. Based on the soil loss map, 32 sub-basins were classified into three categories by priority of intervention: high, moderate, and low. This priority is available to sustain a management plan against sediment filling of the Ouizert dam at the basin outlet. The method enhancing the RUSLE model and confrontation with Google Earth can be easily adapted to other watersheds.
A two-way interface between limited Systems Biology Markup Language and R.
Radivoyevitch, Tomas
2004-12-07
Systems Biology Markup Language (SBML) is gaining broad usage as a standard for representing dynamical systems as data structures. The open source statistical programming environment R is widely used by biostatisticians involved in microarray analyses. An interface between SBML and R does not exist, though one might be useful to R users interested in SBML, and SBML users interested in R. A model structure that parallels SBML to a limited degree is defined in R. An interface between this structure and SBML is provided through two function definitions: write.SBML() which maps this R model structure to SBML level 2, and read.SBML() which maps a limited range of SBML level 2 files back to R. A published model of purine metabolism is provided in this SBML-like format and used to test the interface. The model reproduces published time course responses before and after its mapping through SBML. List infrastructure preexisting in R makes it well-suited for manipulating SBML models. Further developments of this SBML-R interface seem to be warranted.
A two-way interface between limited Systems Biology Markup Language and R
Radivoyevitch, Tomas
2004-01-01
Background Systems Biology Markup Language (SBML) is gaining broad usage as a standard for representing dynamical systems as data structures. The open source statistical programming environment R is widely used by biostatisticians involved in microarray analyses. An interface between SBML and R does not exist, though one might be useful to R users interested in SBML, and SBML users interested in R. Results A model structure that parallels SBML to a limited degree is defined in R. An interface between this structure and SBML is provided through two function definitions: write.SBML() which maps this R model structure to SBML level 2, and read.SBML() which maps a limited range of SBML level 2 files back to R. A published model of purine metabolism is provided in this SBML-like format and used to test the interface. The model reproduces published time course responses before and after its mapping through SBML. Conclusions List infrastructure preexisting in R makes it well-suited for manipulating SBML models. Further developments of this SBML-R interface seem to be warranted. PMID:15585059
ERIC Educational Resources Information Center
Kopcha, Theodore J.; Otumfuor, Beryl A.; Wang, Lu
2015-01-01
This study examines the effects of spatial ability, gender differences, and pictorial training on fourth grade students' ability to recall landmark locations from memory. Ninety-six students used Google Earth over a 3-week period to locate landmarks (3-D) and mark their location on a 2-D topographical map. Analysis of covariance on posttest scores…
VizieR Online Data Catalog: Orion Integral Filament ALMA+IRAM30m N2H+(1-0) data (Hacar+, 2018)
NASA Astrophysics Data System (ADS)
Hacar, A.; Tafalla, M.; Forbrich, J.; Alves, J.; Meingast, S.; Grossschedl, J.; Teixeira, P. S.
2018-01-01
Combined ALMA+IRAM30m large-scale N2H+(1-0) emission in the Orion ISF. Two datasets are presented here in FITS format: 1.- Full data cube: spectral resolution = 0.1 kms-1 2.- Total integrated line intensity (moment 0) map Units are in Jy/beam See also: https://sites.google.com/site/orion4dproject/home (2 data files).
NASA Technical Reports Server (NTRS)
Xiong, Jun; Thenkabail, Prasad S.; Tilton, James C.; Gumma, Murali K.; Teluguntla, Pardhasaradhi; Oliphant, Adam; Congalton, Russell G.; Yadav, Kamini; Gorelick, Noel
2017-01-01
A satellite-derived cropland extent map at high spatial resolution (30-m or better) is a must for food and water security analysis. Precise and accurate global cropland extent maps, indicating cropland and non-cropland areas, is a starting point to develop high-level products such as crop watering methods (irrigated or rainfed), cropping intensities (e.g., single, double, or continuous cropping), crop types, cropland fallows, as well as assessment of cropland productivity (productivity per unit of land), and crop water productivity (productivity per unit of water). Uncertainties associated with the cropland extent map have cascading effects on all higher-level cropland products. However, precise and accurate cropland extent maps at high spatial resolution over large areas (e.g., continents or the globe) are challenging to produce due to the small-holder dominant agricultural systems like those found in most of Africa and Asia. Cloud-based Geospatial computing platforms and multi-date, multi-sensor satellite image inventories on Google Earth Engine offer opportunities for mapping croplands with precision and accuracy over large areas that satisfy the requirements of broad range of applications. Such maps are expected to provide highly significant improvements compared to existing products, which tend to be coarser in resolution, and often fail to capture fragmented small-holder farms especially in regions with high dynamic change within and across years. To overcome these limitations, in this research we present an approach for cropland extent mapping at high spatial resolution (30-m or better) using the 10-day, 10 to 20-m, Sentinel-2 data in combination with 16-day, 30-m, Landsat-8 data on Google Earth Engine (GEE). First, nominal 30-m resolution satellite imagery composites were created from 36,924 scenes of Sentinel-2 and Landsat-8 images for the entire African continent in 2015-2016. These composites were generated using a median-mosaic of five bands (blue, green, red, near-infrared, NDVI) during each of the two periods (period 1: January-June 2016 and period 2: July-December 2015) plus a 30-m slope layer derived from the Shuttle Radar Topographic Mission (SRTM) elevation dataset. Second, we selected Cropland/Non-cropland training samples (sample size 9791) from various sources in GEE to create pixel-based classifications. As supervised classification algorithm, Random Forest (RF) was used as the primary classifier because of its efficiency, and when over-fitting issues of RF happened due to the noise of input training data, Support Vector Machine (SVM) was applied to compensate for such defects in specific areas. Third, the Recursive Hierarchical Segmentation (RHSeg) algorithm was employed to generate an object-oriented segmentation layer based on spectral and spatial properties from the same input data. This layer was merged with the pixel-based classification to improve segmentation accuracy. Accuracies of the merged 30-m crop extent product were computed using an error matrix approach in which 1754 independent validation samples were used. In addition, a comparison was performed with other available cropland maps as well as with LULC maps to show spatial similarity. Finally, the cropland area results derived from the map were compared with UN FAO statistics. The independent accuracy assessment showed a weighted overall accuracy of 94, with a producers accuracy of 85.9 (or omission error of 14.1), and users accuracy of 68.5 (commission error of 31.5) for the cropland class. The total net cropland area (TNCA) of Africa was estimated as 313 Mha for the nominal year 2015.
Research of cartographer laser SLAM algorithm
NASA Astrophysics Data System (ADS)
Xu, Bo; Liu, Zhengjun; Fu, Yiran; Zhang, Changsai
2017-11-01
As the indoor is a relatively closed and small space, total station, GPS, close-range photogrammetry technology is difficult to achieve fast and accurate indoor three-dimensional space reconstruction task. LIDAR SLAM technology does not rely on the external environment a priori knowledge, only use their own portable lidar, IMU, odometer and other sensors to establish an independent environment map, a good solution to this problem. This paper analyzes the Google Cartographer laser SLAM algorithm from the point cloud matching and closed loop detection. Finally, the algorithm is presented in the 3D visualization tool RViz from the data acquisition and processing to create the environment map, complete the SLAM technology and realize the process of indoor threedimensional space reconstruction
Integration of Bim, Web Maps and Iot for Supporting Comfort Analysis
NASA Astrophysics Data System (ADS)
Gunduz, M.; Isikdag, U.; Basaraner, M.
2017-11-01
The use of the Internet is expanding and the technological capabilities of electronic devices are evolving. Today, Internet of Things (IoT) solutions can be developed that were never even imaginable before. In this paper, a case study is presented on the joint use of Building Information Model (BIM), Geographical Information Systems (GIS) and Internet of Things (IoT) technologies. It is a part of an ongoing study that intends to overcome some problems about the management of complex facilities. In the study, a BIM has been converted and displayed in 2D on Google Maps, and information on various sensors have been represented on the web with geographic coordinates in real-time.
Using structure locations as a basis for mapping the wildland urban interface
Avi Bar-Massada; Susan I. Stewart; Roger B. Hammer; Miranda H. Mockrin; Volker C. Radeloff
2013-01-01
The wildland urban interface (WUI) delineates the areas where wildland fire hazard most directly impacts human communities and threatens lives and property, and where houses exert the strongest influence on the natural environment. Housing data are a major problem for WUI mapping. When housing data are zonal, the concept of a WUI neighborhood can be captured easily in...
NASA Astrophysics Data System (ADS)
Giardino, Marco; Magagna, Alessandra; Ferrero, Elena; Perrone, Gianluigi
2015-04-01
Digital field mapping has certainly provided geoscientists with the opportunity to map and gather data in the field directly using digital tools and software rather than using paper maps, notebooks and analogue devices and then subsequently transferring the data to a digital format for subsequent analysis. But, the same opportunity has to be recognized for Geoscience education, as well as for stimulating and helping students in the recognition of landforms and interpretation of the geological and geomorphological components of a landscape. More, an early exposure to mapping during school and prior to university can optimise the ability to "read" and identify uncertainty in 3d models. During 2014, about 200 Secondary School students (aged 12-15) of the Piedmont region (NW Italy) participated in a research program involving the use of mobile devices (smartphone and tablet) in the field. Students, divided in groups, used the application Trimble Outdoors Navigators for tracking a geological trail in the Sangone Valley and for taking georeferenced pictures and notes. Back to school, students downloaded the digital data in a .kml file for the visualization on Google Earth. This allowed them: to compare the hand tracked trail on a paper map with the digital trail, and to discuss about the functioning and the precision of the tools; to overlap a digital/semitransparent version of the 2D paper map (a Regional Technical Map) used during the field trip on the 2.5D landscape of Google Earth, as to help them in the interpretation of conventional symbols such as contour lines; to perceive the landforms seen during the field trip as a part of a more complex Pleistocene glacial landscape; to understand the classical and innovative contributions from different geoscientific disciplines to the generation of a 3D structural geological model of the Rivoli-Avigliana Morainic Amphitheatre. In 2013 and 2014, some other pilot projects have been carried out in different areas of the Piedmont region, and in the Sesia Val Grande Geopark, for testing the utility of digital field mapping in Geoscience education. Feedback from students are positive: they are stimulated and involved by the use of ICT for learning Geoscience, and they voluntary choose to work with their personal mobile device (more than 90% of them own a smartphone); they are interested in knowing the features of GPS, and of software for the visualization of satellite and aerial images, but they recognize the importance of integrating and comparing traditional and innovative methods in the field.
NASA Astrophysics Data System (ADS)
Bajo, J. V.; Martinez-Hackert, B.; Polio, C.; Gutierrez, E.
2015-12-01
Santa Ana (Ilamatepec) Volcano is an active composite volcano located in the Apaneca Volcanic Field located in western part of El Salvador, Central America. The volcano is surrounded by rural communities in its proximal areas and the second (Santa Ana, 13 km) and fourth (Sonsosante, 15 km) largest cities of the country. On October 1st, 2005, the volcano erupted after months of increased activity. Following the eruption, volcanic mitigation projects were conducted in the region, but the communities had little or no input on them. This project consisted in the creation of lahar volcanic hazard map for the Canton Buanos Aires on the northern part of the volcano by incorporating the community's knowledge from prior events to model parameters and results. The work with the community consisted in several meetings where the community members recounted past events. They were asked to map the outcomes of those events using either a topographic map of the area, a Google Earth image, or a blank paper poster size. These maps have been used to identify hazard and vulnerable areas, and for model validation. These maps were presented to the communities and they accepted their results and the maps.
NaviCell Web Service for network-based data visualization.
Bonnet, Eric; Viara, Eric; Kuperstein, Inna; Calzone, Laurence; Cohen, David P A; Barillot, Emmanuel; Zinovyev, Andrei
2015-07-01
Data visualization is an essential element of biological research, required for obtaining insights and formulating new hypotheses on mechanisms of health and disease. NaviCell Web Service is a tool for network-based visualization of 'omics' data which implements several data visual representation methods and utilities for combining them together. NaviCell Web Service uses Google Maps and semantic zooming to browse large biological network maps, represented in various formats, together with different types of the molecular data mapped on top of them. For achieving this, the tool provides standard heatmaps, barplots and glyphs as well as the novel map staining technique for grasping large-scale trends in numerical values (such as whole transcriptome) projected onto a pathway map. The web service provides a server mode, which allows automating visualization tasks and retrieving data from maps via RESTful (standard HTTP) calls. Bindings to different programming languages are provided (Python and R). We illustrate the purpose of the tool with several case studies using pathway maps created by different research groups, in which data visualization provides new insights into molecular mechanisms involved in systemic diseases such as cancer and neurodegenerative diseases. © The Author(s) 2015. Published by Oxford University Press on behalf of Nucleic Acids Research.
NASA Astrophysics Data System (ADS)
Dong, Weihua; Liao, Hua
2016-06-01
Despite the now-ubiquitous two-dimensional (2D) maps, photorealistic three-dimensional (3D) representations of cities (e.g., Google Earth) have gained much attention by scientists and public users as another option. However, there is no consistent evidence on the influences of 3D photorealism on pedestrian navigation. Whether 3D photorealism can communicate cartographic information for navigation with higher effectiveness and efficiency and lower cognitive workload compared to the traditional symbolic 2D maps remains unknown. This study aims to explore whether the photorealistic 3D representation can facilitate processes of map reading and navigation in digital environments using a lab-based eye tracking approach. Here we show the differences of symbolic 2D maps versus photorealistic 3D representations depending on users' eye-movement and navigation behaviour data. We found that the participants using the 3D representation were less effective, less efficient and were required higher cognitive workload than using the 2D map for map reading. However, participants using the 3D representation performed more efficiently in self-localization and orientation at the complex decision points. The empirical results can be helpful to improve the usability of pedestrian navigation maps in future designs.
NaviCell Web Service for network-based data visualization
Bonnet, Eric; Viara, Eric; Kuperstein, Inna; Calzone, Laurence; Cohen, David P. A.; Barillot, Emmanuel; Zinovyev, Andrei
2015-01-01
Data visualization is an essential element of biological research, required for obtaining insights and formulating new hypotheses on mechanisms of health and disease. NaviCell Web Service is a tool for network-based visualization of ‘omics’ data which implements several data visual representation methods and utilities for combining them together. NaviCell Web Service uses Google Maps and semantic zooming to browse large biological network maps, represented in various formats, together with different types of the molecular data mapped on top of them. For achieving this, the tool provides standard heatmaps, barplots and glyphs as well as the novel map staining technique for grasping large-scale trends in numerical values (such as whole transcriptome) projected onto a pathway map. The web service provides a server mode, which allows automating visualization tasks and retrieving data from maps via RESTful (standard HTTP) calls. Bindings to different programming languages are provided (Python and R). We illustrate the purpose of the tool with several case studies using pathway maps created by different research groups, in which data visualization provides new insights into molecular mechanisms involved in systemic diseases such as cancer and neurodegenerative diseases. PMID:25958393
The Live Access Server Scientific Product Generation Through Workflow Orchestration
NASA Astrophysics Data System (ADS)
Hankin, S.; Calahan, J.; Li, J.; Manke, A.; O'Brien, K.; Schweitzer, R.
2006-12-01
The Live Access Server (LAS) is a well-established Web-application for display and analysis of geo-science data sets. The software, which can be downloaded and installed by anyone, gives data providers an easy way to establish services for their on-line data holdings, so their users can make plots; create and download data sub-sets; compare (difference) fields; and perform simple analyses. Now at version 7.0, LAS has been in operation since 1994. The current "Armstrong" release of LAS V7 consists of three components in a tiered architecture: user interface, workflow orchestration and Web Services. The LAS user interface (UI) communicates with the LAS Product Server via an XML protocol embedded in an HTTP "get" URL. Libraries (APIs) have been developed in Java, JavaScript and perl that can readily generate this URL. As a result of this flexibility it is common to find LAS user interfaces of radically different character, tailored to the nature of specific datasets or the mindset of specific users. When a request is received by the LAS Product Server (LPS -- the workflow orchestration component), business logic converts this request into a series of Web Service requests invoked via SOAP. These "back- end" Web services perform data access and generate products (visualizations, data subsets, analyses, etc.). LPS then packages these outputs into final products (typically HTML pages) via Jakarta Velocity templates for delivery to the end user. "Fine grained" data access is performed by back-end services that may utilize JDBC for data base access; the OPeNDAP "DAPPER" protocol; or (in principle) the OGC WFS protocol. Back-end visualization services are commonly legacy science applications wrapped in Java or Python (or perl) classes and deployed as Web Services accessible via SOAP. Ferret is the default visualization application used by LAS, though other applications such as Matlab, CDAT, and GrADS can also be used. Other back-end services may include generation of Google Earth layers using KML; generation of maps via WMS or ArcIMS protocols; and data manipulation with Unix utilities.
A virtual tour of geological heritage: Valourising geodiversity using Google Earth and QR code
NASA Astrophysics Data System (ADS)
Martínez-Graña, A. M.; Goy, J. L.; Cimarra, C. A.
2013-12-01
When making land-use plans, it is necessary to inventory and catalogue the geological heritage and geodiversity of a site to establish an apolitical conservation protection plan to meet the educational and social needs of society. New technologies make it possible to create virtual databases using virtual globes - e.g., Google Earth - and other personal-use geomatics applications (smartphones, tablets, PDAs) for accessing geological heritage information in “real time” for scientific, educational, and cultural purposes via a virtual geological itinerary. Seventeen mapped and georeferenced geosites have been created in Keyhole Markup Language for use in map layers used in geological itinerary stops for different applications. A virtual tour has been developed for Las Quilamas Natural Park, which is located in the Spanish Central System, using geological layers and topographic and digital terrain models that can be overlaid in a 3D model. The Google Earth application was used to import the geosite placemarks. For each geosite, a tab has been developed that shows a description of the geology with photographs and diagrams and that evaluates the scientific, educational, and tourism quality. Augmented reality allows the user to access these georeferenced thematic layers and overlay data, images, and graphics in real time on their mobile devices. These virtual tours can be incorporated into subject guides designed by public. Seven educational and interpretive panels describing some of the geosites were designed and tagged with a QR code that could be printed at each stop or in the printed itinerary. These QR codes can be scanned with the camera found on most mobile devices, and video virtual tours can be viewed on these devices. The virtual tour of the geological heritage can be used to show tourists the geological history of the Las Quilamas Natural Park using new geomatics technologies (virtual globes, augmented reality, and QR codes).
Google Earth Grand Tour Themes
NASA Astrophysics Data System (ADS)
De Paor, D. G.; Whitmeyer, S. J.; Bentley, C.; Dordevic, M. M.
2014-12-01
As part of an NSF TUES Type 3 project entitled "Google Earth for Onsite and Distance Education (GEODE)," we are assembling a "Grand Tour" of locations on Earth and other terrestrial bodies that every geoscience student should know about and visit at least in virtual reality. Based on feedback from colleagues at previous meetings, we have identified nine Grand Tour themes: "Plates and Plumes," "Rocks and Regions," "Geology Through Time," "The Mapping Challenge*," "U.S. National Parks*," "The Magical Mystery Tour*," "Resources and Hazards," "Planets and Moons," and "Top of the Pops." Themes marked with an asterisk are most developed at this stage and will be demonstrated in real time. The Mapping Challenge invites students to trace geological contacts, measure bedding strike and dip and the plunge, trend, and facing of a fold. There is an advanced tool for modeling periclinal folds. The challenge is presented in a game-like format with an emphasis on puzzle-solving that will appeal to students regardless of gender. For the tour of U.S. national parks, we divided the most geologically important parks into four groups—Western Pacific, West Coast, Rockies, and East Coast. We are combining our own team's GigaPan imagery with imagery already available on the Internet. There is a great deal of imagery just waiting to be annotated for geological education purposes. The Magical Mystery Tour takes students to Google Streetview locations selected by instructors. Students are presented with questions or tasks and are given automatic feedback. Other themes are under development. Within each theme, we are crowd-sourcing contributions from colleagues and inviting colleagues to vote for or against proposed locations and student interactions. The GEODE team includes the authors and: Heather Almquist, Stephen Burgin, Cinzia Cervato, Gene Cooper, Paul Karabinos, Terry Pavlis, Jen Piatek, Bill Richards, Jeff Ryan, Ron Schott, Kristen St. John, and Barb Tewksbury.
Next Generation Landsat Products Delivered Using Virtual Globes and OGC Standard Services
NASA Astrophysics Data System (ADS)
Neiers, M.; Dwyer, J.; Neiers, S.
2008-12-01
The Landsat Data Continuity Mission (LDCM) is the next in the series of Landsat satellite missions and is tasked with the objective of delivering data acquired by the Operational Land Imager (OLI). The OLI instrument will provide data continuity to over 30 years of global multispectral data collected by the Landsat series of satellites. The U.S. Geological Survey Earth Resources Observation and Science (USGS EROS) Center has responsibility for the development and operation of the LDCM ground system. One of the mission objectives of the LDCM is to distribute OLI data products electronically over the Internet to the general public on a nondiscriminatory basis and at no cost. To ensure the user community and general public can easily access LDCM data from multiple clients, the User Portal Element (UPE) of the LDCM ground system will use OGC standards and services such as Keyhole Markup Language (KML), Web Map Service (WMS), Web Coverage Service (WCS), and Geographic encoding of Really Simple Syndication (GeoRSS) feeds for both access to and delivery of LDCM products. The USGS has developed and tested the capabilities of several successful UPE prototypes for delivery of Landsat metadata, full resolution browse, and orthorectified (L1T) products from clients such as Google Earth, Google Maps, ESRI ArcGIS Explorer, and Microsoft's Virtual Earth. Prototyping efforts included the following services: using virtual globes to search the historical Landsat archive by dynamic generation of KML; notification of and access to new Landsat acquisitions and L1T downloads from GeoRSS feeds; Google indexing of KML files containing links to full resolution browse and data downloads; WMS delivery of reduced resolution browse, full resolution browse, and cloud mask overlays; and custom data downloads using WCS clients. These various prototypes will be demonstrated and LDCM service implementation plans will be discussed during this session.
Walsh, Gregory J.
2014-01-01
The bedrock geology of the 7.5-minute Uxbridge quadrangle consists of Neoproterozoic metamorphic and igneous rocks of the Avalon zone. In this area, rocks of the Avalon zone lie within the core of the Milford antiform, south and east of the terrane-bounding Bloody Bluff fault zone. Permian pegmatite dikes and quartz veins occur throughout the quadrangle. The oldest metasedimentary rocks include the Blackstone Group, which represents a Neoproterozoic peri-Gondwanan marginal shelf sequence. The metasedimentary rocks are intruded by Neoproterozoic arc-related plutonic rocks of the Rhode Island batholith. This report presents mapping by G.J. Walsh. The complete report consists of a map, text pamphlet, and GIS database. The map and text pamphlet are available only as downloadable files (see frame at right). The GIS database is available for download in ESRI™ shapefile and Google Earth™ formats, and includes contacts of bedrock geologic units, faults, outcrops, structural geologic information, geochemical data, and photographs.
Mapping 2000 2010 Impervious Surface Change in India Using Global Land Survey Landsat Data
NASA Technical Reports Server (NTRS)
Wang, Panshi; Huang, Chengquan; Brown De Colstoun, Eric C.
2017-01-01
Understanding and monitoring the environmental impacts of global urbanization requires better urban datasets. Continuous field impervious surface change (ISC) mapping using Landsat data is an effective way to quantify spatiotemporal dynamics of urbanization. It is well acknowledged that Landsat-based estimation of impervious surface is subject to seasonal and phenological variations. The overall goal of this paper is to map 200-02010 ISC for India using Global Land Survey datasets and training data only available for 2010. To this end, a method was developed that could transfer the regression tree model developed for mapping 2010 impervious surface to 2000 using an iterative training and prediction (ITP) approach An independent validation dataset was also developed using Google Earth imagery. Based on the reference ISC from the validation dataset, the RMSE of predicted ISC was estimated to be 18.4%. At 95% confidence, the total estimated ISC for India between 2000 and 2010 is 2274.62 +/- 7.84 sq km.
Krystosik, Amy R; Curtis, Andrew; Buritica, Paola; Ajayakumar, Jayakrishnan; Squires, Robert; Dávalos, Diana; Pacheco, Robinson; Bhatta, Madhav P; James, Mark A
2017-01-01
Cali, Colombia has experienced chikungunya and Zika outbreaks and hypoendemic dengue. Studies have explained Cali's dengue patterns but lack the sub-neighborhood-scale detail investigated here. Spatial-video geonarratives (SVG) with Ministry of Health officials and Community Health Workers were collected in hotspots, providing perspective on perceptions of why dengue, chikungunya and Zika hotspots exist, impediments to control, and social outcomes. Using spatial video and Google Street View, sub-neighborhood features possibly contributing to incidence were mapped to create risk surfaces, later compared with dengue, chikungunya and Zika case data. SVG captured insights in 24 neighborhoods. Trash and water risks in Calipso were mapped using SVG results. Perceived risk factors included proximity to standing water, canals, poverty, invasions, localized violence and military migration. These risks overlapped case density maps and identified areas that are suitable for transmission but are possibly underreporting to the surveillance system. Resulting risk maps with local context could be leveraged to increase vector-control efficiency- targeting key areas of environmental risk.
Integrating Socioeconomic and Earth Science Data Using Geobrowsers and Web Services: A Demonstration
NASA Astrophysics Data System (ADS)
Schumacher, J. A.; Yetman, G. G.
2007-12-01
The societal benefit areas identified as the focus for the Global Earth Observing System of Systems (GEOSS) 10- year implementation plan are an indicator of the importance of integrating socioeconomic data with earth science data to support decision makers. To aid this integration, CIESIN is delivering its global and U.S. demographic data to commercial and open source Geobrowsers and providing open standards based services for data access. Currently, data on population distribution, poverty, and detailed census data for the U.S. are available for visualization and access in Google Earth, NASA World Wind, and a browser-based 2-dimensional mapping client. The mapping client allows for the creation of web map documents that pull together layers from distributed servers and can be saved and shared. Visualization tools with Geobrowsers, user-driven map creation and sharing via browser-based clients, and a prototype for characterizing populations at risk to predicted precipitation deficits will be demonstrated.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Durcan, Chris A.; Balsano, Robert; LaBella, Vincent P., E-mail: vlabella@albany.edu
2015-06-28
The W/Si(001) Schottky barrier height is mapped to nanoscale dimensions using ballistic electron emission microscopy (BEEM) over a period of 21 days to observe changes in the interface electrostatics. Initially, the average spectrum is fit to a Schottky barrier height of 0.71 eV, and the map is uniform with 98% of the spectra able to be fit. After 21 days, the average spectrum is fit to a Schottky barrier height of 0.62 eV, and the spatial map changes dramatically with only 27% of the spectra able to be fit. Transmission electron microscopy shows the formation of an ultra-thin tungsten silicide at themore » interface, which increases in thickness over the 21 days. This increase is attributed to an increase in electron scattering and the changes are observed in the BEEM measurements. Interestingly, little to no change is observed in the I-V measurements throughout the 21 day period.« less
Savel, Thomas G; Bronstein, Alvin; Duck, William; Rhodes, M Barry; Lee, Brian; Stinn, John; Worthen, Katherine
2010-01-01
Real-time surveillance systems are valuable for timely response to public health emergencies. It has been challenging to leverage existing surveillance systems in state and local communities, and, using a centralized architecture, add new data sources and analytical capacity. Because this centralized model has proven to be difficult to maintain and enhance, the US Centers for Disease Control and Prevention (CDC) has been examining the ability to use a federated model based on secure web services architecture, with data stewardship remaining with the data provider. As a case study for this approach, the American Association of Poison Control Centers and the CDC extended an existing data warehouse via a secure web service, and shared aggregate clinical effects and case counts data by geographic region and time period. To visualize these data, CDC developed a web browser-based interface, Quicksilver, which leveraged the Google Maps API and Flot, a javascript plotting library. Two iterations of the NPDS web service were completed in 12 weeks. The visualization client, Quicksilver, was developed in four months. This implementation of web services combined with a visualization client represents incremental positive progress in transitioning national data sources like BioSense and NPDS to a federated data exchange model. Quicksilver effectively demonstrates how the use of secure web services in conjunction with a lightweight, rapidly deployed visualization client can easily integrate isolated data sources for biosurveillance.
Proposal for optimal placement platform of bikes using queueing networks.
Mizuno, Shinya; Iwamoto, Shogo; Seki, Mutsumi; Yamaki, Naokazu
2016-01-01
In recent social experiments, rental motorbikes and rental bicycles have been arranged at nodes, and environments where users can ride these bikes have been improved. When people borrow bikes, they return them to nearby nodes. Some experiments have been conducted using the models of Hamachari of Yokohama, the Niigata Rental Cycle, and Bicing. However, from these experiments, the effectiveness of distributing bikes was unclear, and many models were discontinued midway. Thus, we need to consider whether these models are effectively designed to represent the distribution system. Therefore, we construct a model to arrange the nodes for distributing bikes using a queueing network. To adopt realistic values for our model, we use the Google Maps application program interface. Thus, we can easily obtain values of distance and transit time between nodes in various places in the world. Moreover, we apply the distribution of a population to a gravity model and we compute the effective transition probability for this queueing network. If the arrangement of the nodes and number of bikes at each node is known, we can precisely design the system. We illustrate our system using convenience stores as nodes and optimize the node configuration. As a result, we can optimize simultaneously the number of nodes, node places, and number of bikes for each node, and we can construct a base for a rental cycle business to use our system.
Quantification of Plant Chlorophyll Content Using Google Glass
Cortazar, Bingen; Koydemir, Hatice Ceylan; Tseng, Derek; Feng, Steve; Ozcan, Aydogan
2015-01-01
Measuring plant chlorophyll concentration is a well-known and commonly used method in agriculture and environmental applications for monitoring plant health, which also correlates with many other plant parameters including, e.g., carotenoids, nitrogen, maximum green fluorescence, etc. Direct chlorophyll measurement using chemical extraction is destructive, complex and time-consuming, which has led to the development of mobile optical readers, providing non-destructive but at the same time relatively expensive tools for evaluation of plant chlorophyll levels. Here we demonstrate accurate measurement of chlorophyll concentration in plant leaves using Google Glass and a custom-developed software application together with a cost-effective leaf holder and multi-spectral illuminator device. Two images, taken using Google Glass, of a leaf placed in our portable illuminator device under red and white (i.e., broadband) light-emitting-diode (LED) illumination are uploaded to our servers for remote digital processing and chlorophyll quantification, with results returned to the user in less than 10 seconds. Intensity measurements extracted from the uploaded images are mapped against gold-standard colorimetric measurements made through a commercially available reader to generate calibration curves for plant leaf chlorophyll concentration. Using five plant species to calibrate our system, we demonstrate that our approach can accurately and rapidly estimate chlorophyll concentration of fifteen different plant species under both indoor and outdoor lighting conditions. This Google Glass based chlorophyll measurement platform can display the results in spatiotemporal and tabular forms and would be highly useful for monitoring of plant health in environmental and agriculture related applications, including e.g., urban plant monitoring, indirect measurements of the effects of climate change, and as an early indicator for water, soil, and air quality degradation. PMID:25669673
Quantification of plant chlorophyll content using Google Glass.
Cortazar, Bingen; Koydemir, Hatice Ceylan; Tseng, Derek; Feng, Steve; Ozcan, Aydogan
2015-04-07
Measuring plant chlorophyll concentration is a well-known and commonly used method in agriculture and environmental applications for monitoring plant health, which also correlates with many other plant parameters including, e.g., carotenoids, nitrogen, maximum green fluorescence, etc. Direct chlorophyll measurement using chemical extraction is destructive, complex and time-consuming, which has led to the development of mobile optical readers, providing non-destructive but at the same time relatively expensive tools for evaluation of plant chlorophyll levels. Here we demonstrate accurate measurement of chlorophyll concentration in plant leaves using Google Glass and a custom-developed software application together with a cost-effective leaf holder and multi-spectral illuminator device. Two images, taken using Google Glass, of a leaf placed in our portable illuminator device under red and white (i.e., broadband) light-emitting-diode (LED) illumination are uploaded to our servers for remote digital processing and chlorophyll quantification, with results returned to the user in less than 10 seconds. Intensity measurements extracted from the uploaded images are mapped against gold-standard colorimetric measurements made through a commercially available reader to generate calibration curves for plant leaf chlorophyll concentration. Using five plant species to calibrate our system, we demonstrate that our approach can accurately and rapidly estimate chlorophyll concentration of fifteen different plant species under both indoor and outdoor lighting conditions. This Google Glass based chlorophyll measurement platform can display the results in spatiotemporal and tabular forms and would be highly useful for monitoring of plant health in environmental and agriculture related applications, including e.g., urban plant monitoring, indirect measurements of the effects of climate change, and as an early indicator for water, soil, and air quality degradation.
Kaewpitoon, Soraya J; Rujirakul, Ratana; Sangkudloa, Amnat; Kaewthani, Sarochinee; Khemplila, Kritsakorn; Cherdjirapong, Karuna; Kujapun, Jirawoot; Norkaew, Jun; Chavengkun, Wasugree; Ponphimai, Sukanya; Polsripradist, Poowadol; Padchasuwan, Natnapa; Joosiri, Apinya; Wakkhuwattapong, Parichart; Loyd, Ryan A; Matrakool, Likit; Tongtawee, Taweesak; Panpimanmas, Sukij; Kaewpitoon, Natthawut
2016-01-01
Cholangiocarcinoma (CCA), a major problem of health in Thailand, particularly in Northeastern and Northern regions, is generally incurable and rapidly lethal because of presentation in stage 3 or 4. Early diagnosis of stage 1 and 2 could allow better survival. Therefore, this study aimed to provide a distribution map of populations at risk for CCA in BuaYai district of Nakhon Ratchasima province, Northeast Thailand. A cross-sectional survey was carried out in 10 sub-districts and 122 villages, during June and November 2015. The populations at risk for CCA were screened using the Korat CCA verbal screening test (KCVST) and then risk areas were displayed by using Google map (GM). A total of 11,435 individuals from a 26,198 population completed the KCVST. The majority had a low score of risk for CCA (1-4 points; 93.3%). High scores with 6, 7 and 8 points accounted for 1.20%, 0.13% and 0.02%. The population at risk was found frequently in sub-district municipalities, followed by sub-district administrative organization and town municipalities, (F=396.220, P-value=0.000). Distribution mapping comprised 11 layers: 1, district; 2, local administrative organization; 3, hospital; 4, KCVST opisthorchiasis; 5, KCVST praziquantel used; 6, KCVST cholelithiasis; 7, KCVST raw fish consumption; 8, KCVST alcohol consumption; 9, KCVST pesticide used; 10, KCVST relative family with CCA; and 11, KCVST naive northeastern people. Geovisual display is now available online. This study indicated that the population at high risk of CCA in Bua Yai district is low, therefore setting a zero model project is possible. Key success factors for disease prevention and control need further study. GM production is suitable for further CCA surveillance and monitoring of the population with a high risk score in this area.
Kumar, Rajendra; Sobhy, Haitham
2017-01-01
Abstract Hi-C experiments generate data in form of large genome contact maps (Hi-C maps). These show that chromosomes are arranged in a hierarchy of three-dimensional compartments. But to understand how these compartments form and by how much they affect genetic processes such as gene regulation, biologists and bioinformaticians need efficient tools to visualize and analyze Hi-C data. However, this is technically challenging because these maps are big. In this paper, we remedied this problem, partly by implementing an efficient file format and developed the genome contact map explorer platform. Apart from tools to process Hi-C data, such as normalization methods and a programmable interface, we made a graphical interface that let users browse, scroll and zoom Hi-C maps to visually search for patterns in the Hi-C data. In the software, it is also possible to browse several maps simultaneously and plot related genomic data. The software is openly accessible to the scientific community. PMID:28973466
Cloud-based Computing and Applications of New Snow Metrics for Societal Benefit
NASA Astrophysics Data System (ADS)
Nolin, A. W.; Sproles, E. A.; Crumley, R. L.; Wilson, A.; Mar, E.; van de Kerk, M.; Prugh, L.
2017-12-01
Seasonal and interannual variability in snow cover affects socio-environmental systems including water resources, forest ecology, freshwater and terrestrial habitat, and winter recreation. We have developed two new seasonal snow metrics: snow cover frequency (SCF) and snow disappearance date (SDD). These metrics are calculated at 500-m resolution using NASA's Moderate Resolution Imaging Spectroradiometer (MODIS) snow cover data (MOD10A1). SCF is the number of times snow is observed in a pixel over the user-defined observation period. SDD is the last date of observed snow in a water year. These pixel-level metrics are calculated rapidly and globally in the Google Earth Engine cloud-based environment. SCF and SDD can be interactively visualized in a map-based interface, allowing users to explore spatial and temporal snowcover patterns from 2000-present. These metrics are especially valuable in regions where snow data are sparse or non-existent. We have used these metrics in several ongoing projects. When SCF was linked with a simple hydrologic model in the La Laguna watershed in northern Chile, it successfully predicted summer low flows with a Nash-Sutcliffe value of 0.86. SCF has also been used to help explain changes in Dall sheep populations in Alaska where sheep populations are negatively impacted by late snow cover and low snowline elevation during the spring lambing season. In forest management, SCF and SDD appear to be valuable predictors of post-wildfire vegetation growth. We see a positive relationship between winter SCF and subsequent summer greening for several years post-fire. For western US winter recreation, we are exploring trends in SDD and SCF for regions where snow sports are economically important. In a world with declining snowpacks and increasing uncertainty, these metrics extend across elevations and fill data gaps to provide valuable information for decision-making. SCF and SDD are being produced so that anyone with Internet access and a Google account can access, visualize, and download the data with a minimum of technical expertise and no need for proprietary software.
WebViz:A Web-based Collaborative Interactive Visualization System for large-Scale Data Sets
NASA Astrophysics Data System (ADS)
Yuen, D. A.; McArthur, E.; Weiss, R. M.; Zhou, J.; Yao, B.
2010-12-01
WebViz is a web-based application designed to conduct collaborative, interactive visualizations of large data sets for multiple users, allowing researchers situated all over the world to utilize the visualization services offered by the University of Minnesota’s Laboratory for Computational Sciences and Engineering (LCSE). This ongoing project has been built upon over the last 3 1/2 years .The motivation behind WebViz lies primarily with the need to parse through an increasing amount of data produced by the scientific community as a result of larger and faster multicore and massively parallel computers coming to the market, including the use of general purpose GPU computing. WebViz allows these large data sets to be visualized online by anyone with an account. The application allows users to save time and resources by visualizing data ‘on the fly’, wherever he or she may be located. By leveraging AJAX via the Google Web Toolkit (http://code.google.com/webtoolkit/), we are able to provide users with a remote, web portal to LCSE's (http://www.lcse.umn.edu) large-scale interactive visualization system already in place at the University of Minnesota. LCSE’s custom hierarchical volume rendering software provides high resolution visualizations on the order of 15 million pixels and has been employed for visualizing data primarily from simulations in astrophysics to geophysical fluid dynamics . In the current version of WebViz, we have implemented a highly extensible back-end framework built around HTTP "server push" technology. The web application is accessible via a variety of devices including netbooks, iPhones, and other web and javascript-enabled cell phones. Features in the current version include the ability for users to (1) securely login (2) launch multiple visualizations (3) conduct collaborative visualization sessions (4) delegate control aspects of a visualization to others and (5) engage in collaborative chats with other users within the user interface of the web application. These features are all in addition to a full range of essential visualization functions including 3-D camera and object orientation, position manipulation, time-stepping control, and custom color/alpha mapping.
Walsh, Gregory J.; Jahns, Richard H.; Aleinikoff, John N.
2013-01-01
The bedrock geology of the 7.5-minute Nashua South quadrangle consists primarily of deformed Silurian metasedimentary rocks of the Berwick Formation. The metasedimentary rocks are intruded by a Late Silurian to Early Devonian diorite-gabbro suite, Devonian rocks of the Ayer Granodiorite, Devonian granitic rocks of the New Hampshire Plutonic Suite including pegmatite and the Chelmsford Granite, and Jurassic diabase dikes. The bedrock geology was mapped to study the tectonic history of the area and to provide a framework for ongoing hydrogeologic characterization of the fractured bedrock of Massachusetts and New Hampshire. This report presents mapping by G.J. Walsh and R.H. Jahns and zircon U-Pb geochronology by J.N. Aleinikoff. The complete report consists of a map, text pamphlet, and GIS database. The map and text pamphlet are only available as downloadable files (see frame at right). The GIS database is available for download in ESRITM shapefile and Google EarthTM formats, and includes contacts of bedrock geologic units, faults, outcrops, structural geologic information, photographs, and a three-dimensional model.
BatMis: a fast algorithm for k-mismatch mapping.
Tennakoon, Chandana; Purbojati, Rikky W; Sung, Wing-Kin
2012-08-15
Second-generation sequencing (SGS) generates millions of reads that need to be aligned to a reference genome allowing errors. Although current aligners can efficiently map reads allowing a small number of mismatches, they are not well suited for handling a large number of mismatches. The efficiency of aligners can be improved using various heuristics, but the sensitivity and accuracy of the alignments are sacrificed. In this article, we introduce Basic Alignment tool for Mismatches (BatMis)--an efficient method to align short reads to a reference allowing k mismatches. BatMis is a Burrows-Wheeler transformation based aligner that uses a seed and extend approach, and it is an exact method. Benchmark tests show that BatMis performs better than competing aligners in solving the k-mismatch problem. Furthermore, it can compete favorably even when compared with the heuristic modes of the other aligners. BatMis is a useful alternative for applications where fast k-mismatch mappings, unique mappings or multiple mappings of SGS data are required. BatMis is written in C/C++ and is freely available from http://code.google.com/p/batmis/
A Fast, Minimalist Search Tool for Remote Sensing Data
NASA Astrophysics Data System (ADS)
Lynnes, C. S.; Macharrie, P. G.; Elkins, M.; Joshi, T.; Fenichel, L. H.
2005-12-01
We present a tool that emphasizes speed and simplicity in searching remotely sensed Earth Science data. The tool, nicknamed "Mirador" (Spanish for a scenic overlook), provides only four freetext search form fields, for Keywords, Location, Data Start and Data Stop. This contrasts with many current Earth Science search tools that offer highly structured interfaces in order to ensure precise, non-zero results. The disadvantages of the structured approach lie in its complexity and resultant learning curve, as well as the time it takes to formulate and execute the search, thus discouraging iterative discovery. On the other hand, the success of the basic Google search interface shows that many users are willing to forgo high search precision if the search process is fast enough to enable rapid iteration. Therefore, we employ several methods to increase the speed of search formulation and execution. Search formulation is expedited by the minimalist search form, with only one required field. Also, a gazetteer enables the use of geographic terms as shorthand for latitude/longitude coordinates. The search execution is accelerated by initially presenting dataset results (returned from a Google Mini appliance) with an estimated number of "hits" for each dataset based on the user's space-time constraints. The more costly file-level search is executed against a PostGres database only when the user "drills down", and then covering only the fraction of the time period needed to return the next page of results. The simplicity of the search form makes the tool easy to learn and use, and the speed of the searches enables an iterative form of data discovery.
Optimizing Distributed Sensor Placement for Border Patrol Interdiction Using Microsoft Excel
2007-04-01
weather conditions and they can be evaded by using techniques which minimize heat signatures use of lasers and other technologies day or night (26:8...technologies which can be used for border security. Maier [2004] developed a seismic intrusion sensor technology which uses fiber optic cables, lasers , and...needed to create the is used as the base map for the network. program originally developed by Keyhole by Google Inc. It provides satellite images of
On Internet Symmetry and its Impact on Society
NASA Astrophysics Data System (ADS)
Wolff, S. S.
2014-12-01
The end-to-end principle, enunciated by Clark and Saltzer in 1981 enabled an Internet implementation in which there was a symmetry among the network nodes in the sense that no node was architecturally distinguished. Each interface to the network had a unique and accessible address and could communicate on equal terms with any other interface or collection of interfaces. In this egalitarian implementation there was in consequence no architectural distinction between providers and consumers of content - any network node could play either role. As the Internet spread to university campuses, incoming students found 10 megabit Ethernet in the dorm - while their parents at home were still stuck with 56 kilobit dialup. In the two decades bisected by the millenium, this combination of speed and symmetry on campus and beyond led to a panoply of transformational Internet applications such as Internet video conferencing and billion dollar industries like Google, Yahoo!, and Facebook. This talk places early Internet history in a social context, elaborates on the social and economic outcomes, defines"middlebox friction", discusses its erosive consequences, and suggests a solution to restore symmetry to the Internet-at-large.
NASA Astrophysics Data System (ADS)
Girvetz, E. H.; Zganjar, C.; Raber, G. T.; Hoekstra, J.; Lawler, J. J.; Kareiva, P.
2008-12-01
Now that there is overwhelming evidence of global climate change, scientists, managers and planners (i.e. practitioners) need to assess the potential impacts of climate change on particular ecological systems, within specific geographic areas, and at spatial scales they care about, in order to make better land management, planning, and policy decisions. Unfortunately, this application of climate science to real world decisions and planning has proceeded too slowly because we lack tools for translating cutting-edge climate science and climate-model outputs into something managers and planners can work with at local or regional scales (CCSP 2008). To help increase the accessibility of climate information, we have developed a freely-available, easy-to-use, web-based climate-change analysis toolbox, called ClimateWizard, for assessing how climate has and is projected to change at specific geographic locations throughout the world. The ClimateWizard uses geographic information systems (GIS), web-services (SOAP/XML), statistical analysis platforms (e.g. R- project), and web-based mapping services (e.g. Google Earth/Maps, KML/GML) to provide a variety of different analyses (e.g. trends and departures) and outputs (e.g. maps, graphs, tables, GIS layers). Because ClimateWizard analyzes large climate datasets stored remotely on powerful computers, users of the tool do not need to have fast computers or expensive software, but simply need access to the internet. The analysis results are then provided to users in a Google Maps webpage tailored to the specific climate-change question being asked. The ClimateWizard is not a static product, but rather a framework to be built upon and modified to suit the purposes of specific scientific, management, and policy questions. For example, it can be expanded to include bioclimatic variables (e.g. evapotranspiration) and marine data (e.g. sea surface temperature), as well as improved future climate projections, and climate-change impact analyses involving hydrology, vegetation, wildfire, disease, and food security. By harnessing the power of computer and web- based technologies, the ClimateWizard puts local, regional, and global climate-change analyses in the hands of a wider array of managers, planners, and scientists.
Environmental asbestos exposure sources in Korea
2016-01-01
Background Because of the long asbestos-related disease latencies (10–50 years), detection, diagnosis, and epidemiologic studies require asbestos exposure history. However, environmental asbestos exposure source (EAES) data are lacking. Objectives To survey the available data for past EAES and supplement these data with interviews. Methods We constructed an EAES database using a literature review and interviews of experts, former traders, and workers. Exposure sources by time period and type were visualized using a geographic information system (ArcGIS), web-based mapping (Google Maps), and OpenWeatherMap. The data were mounted in the GIS to show the exposure source location and trend. Results The majority of asbestos mines, factories, and consumption was located in Chungnam; Gyeonggi, Busan, and Gyeongnam; and Gyeonggi, Daejeon, and Busan, respectively. Shipbuilding and repair companies were mostly located in Busan and Gyeongnam. Conclusions These tools might help evaluate past exposure from EAES and estimate the future asbestos burden in Korea. PMID:27726756
Environmental asbestos exposure sources in Korea.
Kang, Dong-Mug; Kim, Jong-Eun; Kim, Ju-Young; Lee, Hyun-Hee; Hwang, Young-Sik; Kim, Young-Ki; Lee, Yong-Jin
2016-10-01
Because of the long asbestos-related disease latencies (10-50 years), detection, diagnosis, and epidemiologic studies require asbestos exposure history. However, environmental asbestos exposure source (EAES) data are lacking. To survey the available data for past EAES and supplement these data with interviews. We constructed an EAES database using a literature review and interviews of experts, former traders, and workers. Exposure sources by time period and type were visualized using a geographic information system (ArcGIS), web-based mapping (Google Maps), and OpenWeatherMap. The data were mounted in the GIS to show the exposure source location and trend. The majority of asbestos mines, factories, and consumption was located in Chungnam; Gyeonggi, Busan, and Gyeongnam; and Gyeonggi, Daejeon, and Busan, respectively. Shipbuilding and repair companies were mostly located in Busan and Gyeongnam. These tools might help evaluate past exposure from EAES and estimate the future asbestos burden in Korea.
Map of Life - A Dashboard for Monitoring Planetary Species Distributions
NASA Astrophysics Data System (ADS)
Jetz, W.
2016-12-01
Geographic information about biodiversity is vital for understanding the many services nature provides and their potential changes, yet remains unreliable and often insufficient. By integrating a wide range of knowledge about species distributions and their dynamics over time, Map of Life supports global biodiversity education, monitoring, research and decision-making. Built on a scalable web platform geared for large biodiversity and environmental data, Map of Life endeavors provides species range information globally and species lists for any area. With data and technology provided by NASA and Google Earth Engine, tools under development use remote sensing-based environmental layers to enable on-the-fly predictions of species distributions, range changes, and early warning signals for threatened species. The ultimate vision is a globally connected, collaborative knowledge- and tool-base for regional and local biodiversity decision-making, education, monitoring, and projection. For currently available tools, more information and to follow progress, go to MOL.org.
Prinos, Scott T.
2017-07-11
The inland extent of saltwater at the base of the Biscayne aquifer in the Model Land Area of Miami-Dade County, Florida, was mapped in 2011. Since that time, the saltwater interface has continued to move inland. The interface is near several active well fields; therefore, an updated approximation of the inland extent of saltwater and an improved understanding of the rate of movement of the saltwater interface are necessary. A geographic information system was used to create a map using the data collected by the organizations that monitor water salinity in this area. An average rate of saltwater interface movement of 140 meters per year was estimated by dividing the distance between two monitoring wells (TPGW-7L and Sec34-MW-02-FS) by the travel time. The travel time was determined by estimating the dates of arrival of the saltwater interface at the wells and computing the difference. This estimate assumes that the interface is traveling east to west between the two monitoring wells. Although monitoring is spatially limited in this area and some of the wells are not ideally designed for salinity monitoring, the monitoring network in this area is improving in spatial distribution and most of the new wells are well designed for salinity monitoring. The approximation of the inland extent of the saltwater interface and the estimated rate of movement of the interface are dependent on existing data. Improved estimates could be obtained by installing uniformly designed monitoring wells in systematic transects extending landward of the advancing saltwater interface.
Optimal design method to minimize users' thinking mapping load in human-machine interactions.
Huang, Yanqun; Li, Xu; Zhang, Jie
2015-01-01
The discrepancy between human cognition and machine requirements/behaviors usually results in serious mental thinking mapping loads or even disasters in product operating. It is important to help people avoid human-machine interaction confusions and difficulties in today's mental work mastered society. Improving the usability of a product and minimizing user's thinking mapping and interpreting load in human-machine interactions. An optimal human-machine interface design method is introduced, which is based on the purpose of minimizing the mental load in thinking mapping process between users' intentions and affordance of product interface states. By analyzing the users' thinking mapping problem, an operating action model is constructed. According to human natural instincts and acquired knowledge, an expected ideal design with minimized thinking loads is uniquely determined at first. Then, creative alternatives, in terms of the way human obtains operational information, are provided as digital interface states datasets. In the last, using the cluster analysis method, an optimum solution is picked out from alternatives, by calculating the distances between two datasets. Considering multiple factors to minimize users' thinking mapping loads, a solution nearest to the ideal value is found in the human-car interaction design case. The clustering results show its effectiveness in finding an optimum solution to the mental load minimizing problems in human-machine interaction design.
Galaxy Portal: interacting with the galaxy platform through mobile devices.
Børnich, Claus; Grytten, Ivar; Hovig, Eivind; Paulsen, Jonas; Čech, Martin; Sandve, Geir Kjetil
2016-06-01
: We present Galaxy Portal app, an open source interface to the Galaxy system through smart phones and tablets. The Galaxy Portal provides convenient and efficient monitoring of job completion, as well as opportunities for inspection of results and execution history. In addition to being useful to the Galaxy community, we believe that the app also exemplifies a useful way of exploiting mobile interfaces for research/high-performance computing resources in general. The source is freely available under a GPL license on GitHub, along with user documentation and pre-compiled binaries and instructions for several platforms: https://github.com/Tarostar/QMLGalaxyPortal It is available for iOS version 7 (and newer) through the Apple App Store, and for Android through Google Play for version 4.1 (API 16) or newer. geirksa@ifi.uio.no. © The Author 2016. Published by Oxford University Press.
Interface shape and crystallinity in LEC GaAs
NASA Astrophysics Data System (ADS)
Tower, J. P.; Tobin, R.; Pearah, P. J.; Ware, R. M.
1991-12-01
Growth striation mapping was used to relate the growth interface shape to crystallinity failure modes in LEC growth of undoped <100> GaAs. The onset of twinning and polycrystallinity were both found to depend on the interface shape near the crystal periphery. The origins of polycrystalline growth were investigated in 8 kg, 3-inch and 4-inch diameter crystals. Interface maps of these crystals show that polycrystalline growth begins when the growth interface periphery turns down, independent of the shape of the central portions. The cause of initial grain boundary formation was found to be included gallium droplets which originate on the surface and migrate through the crystal toward the growth interface. Twinning occurs on {111} facets, usually during shoulder growth. Growth striations show that the sequence of events leading to twin formation consists of deep facet growth, followed by meltback and rapid regrowth. We found it possible to avoid twinning by reducing melt instabilities or by reducing the extent of facet growth.
The North America tapestry of time and terrain
Barton, Kate E.; Howell, David G.; Vigil, Jose F.
2003-01-01
The North America Tapestry of Time and Terrain (1:8,000,000 scale) is a product of the US Geological Survey in the I-map series (I-2781). This map was prepared in collaboration with the Geological Survey of Canada and the Mexican Consejo Recursos de Minerales. This cartographic Tapestry is woven from a geologic map and a shaded relief image. This digital combination reveals the geologic history of North America through the interrelation of rock type, topography and time. Regional surface processes as well as continent-scale tectonic events are exposed in the three dimensions of space and the fourth dimension, geologic time. The large map shows the varying age of bedrock underlying North America, while four smaller maps show the distribution of four principal types of rock: sedimentary, volcanic, plutonic and metamorphic.This map expands the original concept of the 2000 Tapestry of Time and Terrain, by José F. Vigil, Richard J. Pike and David G. Howell, which covered the conterminous United States. The U.S. Tapestry poster and website have been popular in classrooms, homes, and even the Google office building, and we anticipate the North America Tapestry will have a similarly wide appeal, and to a larger audience.
Applying Cognitive Psychology to User Interfaces
NASA Astrophysics Data System (ADS)
Durrani, Sabeen; Durrani, Qaiser S.
This paper explores some key aspects of cognitive psychology that may be mapped onto user interfaces. Major focus in existing user interface guidelines is on consistency, simplicity, feedback, system messages, display issues, navigation, colors, graphics, visibility and error prevention [8-10]. These guidelines are effective indesigning user interfaces. However, these guidelines do not handle the issues that may arise due to the innate structure of human brain and human limitations. For example, where to place graphics on the screen so that user can easily process them and what kind of background should be given on the screen according to the limitation of human motor system. In this paper we have collected some available guidelines from the area of cognitive psychology [1, 5, 7]. In addition, we have extracted few guidelines from theories and studies of cognitive psychology [3, 11] which may be mapped to user interfaces.
LOD map--A visual interface for navigating multiresolution volume visualization.
Wang, Chaoli; Shen, Han-Wei
2006-01-01
In multiresolution volume visualization, a visual representation of level-of-detail (LOD) quality is important for us to examine, compare, and validate different LOD selection algorithms. While traditional methods rely on ultimate images for quality measurement, we introduce the LOD map--an alternative representation of LOD quality and a visual interface for navigating multiresolution data exploration. Our measure for LOD quality is based on the formulation of entropy from information theory. The measure takes into account the distortion and contribution of multiresolution data blocks. A LOD map is generated through the mapping of key LOD ingredients to a treemap representation. The ordered treemap layout is used for relative stable update of the LOD map when the view or LOD changes. This visual interface not only indicates the quality of LODs in an intuitive way, but also provides immediate suggestions for possible LOD improvement through visually-striking features. It also allows us to compare different views and perform rendering budget control. A set of interactive techniques is proposed to make the LOD adjustment a simple and easy task. We demonstrate the effectiveness and efficiency of our approach on large scientific and medical data sets.
Development of decision support system for oil spill management in the Mediterranean Sea
NASA Astrophysics Data System (ADS)
Liubartseva, Svitlana; Coppini, Giovanni; Pinardi, Nadia; De Dominicis, Michela; Marra, Palmalisa; Lecci, Rita; Turrisi, Giuseppe; Creti, Sergio; Martinelli, Sara; Agostini, Paola; Palermo, Francesco
2016-04-01
Possible oil spill accidents and operational pollution could have severe impacts on the Mediterranean basin. It is therefore crucial to provide decision makers, stakeholders, and public with trustworthy DSS (Decision Support System) based on the environmental monitoring, state-of-the-art modeling and innovative technology platforms. Innovative web-based decision support system, called WITOL (Where Is The Oil http://www.witoil.com), has been developed to maintain emergency management in case of oil spill accidents. WITOIL embraces (1) Lagrangian oil spill model MEDSLIK-II (De Dominicis et al., 2013 http://medslikii.bo.ingv.it) coupled with the basin-scale and regional operational oceanographic services; (2) two-modular block of oil spill forecast and uncertainty evaluation; (3) user visualization tool including web and mobile interface with visualization of geospatial information by means of Google Maps. Service-oriented approach plays a key role in the WITOIL DSS development. The system meets the real-time requirements in terms of performance and in dynamic service delivery. Client part of WITOIL is presented by a 8-language GUI (Graphical User Interface) supplied with a great variety of user services including a video tutorial (https://www.youtube.com/watch?v=qj_GokYy8MU). GUI allows users to configure and activate the system, visualize the results using Google Maps, and save them afterwards. Not only does a new generation of DSS require the oil spill forecast, but it also needs the evaluation of uncertainty, which is critical for efficient response, recovery, and mitigation. Uncertainty in prediction of the oil transport and transformation stems from the uncertain environment and data-sparse. A new methodology of uncertainty calculation with respect to initial conditions is incorporated in WITOIL DSS. The results are presented in probability terms. Special application to Android has been implemented to support users involved in the field operations. The system is developed as a part of TESSA Project portfolio providing the unified access to others services. Thus, SEACONDITIONS (http://www.sea-conditions.com) performs visualization and on-line delivery of forecast of surface currents, sea surface temperature, significant wave height and direction, wave period and direction; air temperature, surface pressure, precipitation, cloud coverage, wind speed, etc. Apart from the basin scale visualization SEACONDITIONS supports the zooming capability. User feedback reports from fishermen, port authorities including Coast Guard, offshore companies, aquatic and coastal tourism managers, and academia have been collected and used for the system improvements. User-friendliness of GUI, tooltips, an opportunity to vary the advanced parameters, efficiency of the visualization tool, and a help section were appreciated in these reports. In accordance with the users' requirements, a to-do list is composed for the further development of WITOIL. This work was performed in the framework of the TESSA Project (Sviluppo di TEcnologie per la Situational Sea Awareness) supported by PON (Ricerca & Competitività 2007-2013) cofunded by UE (Fondo Europeo di sviluppo regionale), MIUR (Ministero Italiano dell'Università e della Ricerca), and MSE (Ministero dello Sviluppo Economico). References De Dominicis, M., Pinardi, N., Zodiatis, G., and Lardner, R., 2013. MEDSLIK-II, a Lagrangian marine surface oil spill model for short term forecasting - Part 1: Theory. Geosci. Model Dev. 6, 1851-1869.
The excitement of Google Scholar, the worry of Google Print
Banks, Marcus A
2005-01-01
In late 2004 Google announced two major projects, the unveiling of Google Scholar and a major expansion of the Google Print digitization program. Both projects have generated discussion within the library and research communities, and Google Print has received significant media attention. This commentary describes exciting educational possibilities stimulated by Google Scholar, and argues for caution regarding the Google Print project. PMID:15784147
Usability evaluation of mobile applications using ISO 9241 and ISO 25062 standards.
Moumane, Karima; Idri, Ali; Abran, Alain
2016-01-01
This paper presents an empirical study based on a set of measures to evaluate the usability of mobile applications running on different mobile operating systems, including Android, iOS and Symbian. The aim is to evaluate empirically a framework that we have developed on the use of the Software Quality Standard ISO 9126 in mobile environments, especially the usability characteristic. To do that, 32 users had participated in the experiment and we have used ISO 25062 and ISO 9241 standards for objective measures by working with two widely used mobile applications: Google Apps and Google Maps. The QUIS 7.0 questionnaire have been used to collect measures assessing the users' level of satisfaction when using these two mobile applications. By analyzing the results we highlighted a set of mobile usability issues that are related to the hardware as well as to the software and that need to be taken into account by designers and developers in order to improve the usability of mobile applications.
orthAgogue: an agile tool for the rapid prediction of orthology relations.
Ekseth, Ole Kristian; Kuiper, Martin; Mironov, Vladimir
2014-03-01
The comparison of genes and gene products across species depends on high-quality tools to determine the relationships between gene or protein sequences from various species. Although some excellent applications are available and widely used, their performance leaves room for improvement. We developed orthAgogue: a multithreaded C application for high-speed estimation of homology relations in massive datasets, operated via a flexible and easy command-line interface. The orthAgogue software is distributed under the GNU license. The source code and binaries compiled for Linux are available at https://code.google.com/p/orthagogue/.
3D Immersive Visualization with Astrophysical Data
NASA Astrophysics Data System (ADS)
Kent, Brian R.
2017-01-01
We present the refinement of a new 3D immersion technique for astrophysical data visualization.Methodology to create 360 degree spherical panoramas is reviewed. The 3D software package Blender coupled with Python and the Google Spatial Media module are used together to create the final data products. Data can be viewed interactively with a mobile phone or tablet or in a web browser. The technique can apply to different kinds of astronomical data including 3D stellar and galaxy catalogs, images, and planetary maps.
NASA Astrophysics Data System (ADS)
de Paor, D. G.; Whitmeyer, S. J.; Gobert, J.
2009-12-01
We previously reported on innovative techniques for presenting data on virtual globes such as Google Earth using emergent Collada models that reveal subsurface geology and geophysics. We here present several new and enhanced models and linked lesson plans to aid deployment in undergraduate geoscience courses, along with preliminary results from our assessment of their effectiveness. The new Collada models are created with Google SketchUp, Bonzai3D, and MeshLab software, and are grouped to cover (i) small scale field mapping areas; (ii) regional scale studies of the North Atlantic Ocean Basin, the Appalachian Orogen, and the Pacific Ring of Fire; and (iii) global scale studies of terrestrial planets, moons, and asteroids. Enhancements include emergent block models with three-dimensional surface topography; models that conserve structural orientation data; interactive virtual specimens; models that animate plate movements on the virtual globe; exploded 3-D views of planetary mantles and cores; and server-generated dynamic KML. We tested volunteer students and professors using Silverback monitoring software, think-aloud verbalizations, and questionnaires designed to assess their understanding of the underlying geo-scientific phenomena. With the aid of a cohort of instructors across the U.S., we are continuing to assess areas in which users encounter difficulties with both the software and geoscientific concepts. Preliminary results suggest that it is easy to overestimate the computer expertise of novice users even when they are content knowledge experts (i.e., instructors), and that a detailed introduction to virtual globe manipulation is essential before moving on to geoscience applications. Tasks that seem trivial to developers may present barriers to non-technical users and technicalities that challenge instructors may block adoption in the classroom. We have developed new models using the Google Earth API which permits enhanced interaction and dynamic feedback and are assessing their relative merits versus the Google Earth APP. Overall, test students and professors value the models very highly. There are clear pedagogical opportunities for using materials such as these to create engaging in-course research opportunities for undergraduates.
NASA Astrophysics Data System (ADS)
Singh, Anil Kumar; Gupta, Anjan K.
2018-05-01
Evolution of electronic inhomogeneities with back-gate voltage in graphene on SiO2 was studied using room temperature scanning tunneling microscopy and spectroscopy. Reversal of contrast in some places in the conductance maps and sharp changes in cross correlations between topographic and conductance maps, when graphene Fermi energy approaches its Dirac point, are attributed to the change in charge state of interface defects. The spatial correlations in the conductance maps, described by two length scales, and their growth during approach to Dirac point, show a qualitative agreement with the predictions of the screening theory of graphene. Thus a sharp change in the two length scales close to the Dirac point, seen in our experiments, is interpreted in terms of the change in charge state of some of the interface defects. A systematic understanding and control of the charge state of defects can help in memory applications of graphene.
Assessing the methods needed for improved dengue mapping: a SWOT analysis
Attaway, David Frost; Jacobsen, Kathryn H; Falconer, Allan; Manca, Germana; Waters, Nigel M
2014-01-01
Introduction Dengue fever, a mosquito-borne viral infection, is a growing threat to human health in tropical and subtropical areas worldwide. There is a demand from public officials for maps that capture the current distribution of dengue and maps that analyze risk factors to predict the future burden of disease. Methods To identify relevant articles, we searched Google Scholar, PubMed, BioMed Central, and WHOLIS (World Health Organization Library Database) for published articles with a specific set of dengue criteria between January 2002 and July 2013. Results After evaluating the currently available dengue models, we identified four key barriers to the creation of high-quality dengue maps: (1) data limitations related to the expense of diagnosing and reporting dengue cases in places where health information systems are underdeveloped; (2) issues related to the use of socioeconomic proxies in places with limited dengue incidence data; (3) mosquito ranges which may be changing as a result of climate changes; and (4) the challenges of mapping dengue events at a variety of scales. Conclusion An ideal dengue map will present endemic and epidemic dengue information from both rural and urban areas. Overcoming the current barriers requires expanded collaboration and data sharing by geographers, epidemiologists, and entomologists. Enhanced mapping techniques would allow for improved visualizations of dengue rates and risks. PMID:25328585
Assessing the methods needed for improved dengue mapping: a SWOT analysis.
Attaway, David Frost; Jacobsen, Kathryn H; Falconer, Allan; Manca, Germana; Waters, Nigel M
2014-01-01
Dengue fever, a mosquito-borne viral infection, is a growing threat to human health in tropical and subtropical areas worldwide. There is a demand from public officials for maps that capture the current distribution of dengue and maps that analyze risk factors to predict the future burden of disease. To identify relevant articles, we searched Google Scholar, PubMed, BioMed Central, and WHOLIS (World Health Organization Library Database) for published articles with a specific set of dengue criteria between January 2002 and July 2013. After evaluating the currently available dengue models, we identified four key barriers to the creation of high-quality dengue maps: (1) data limitations related to the expense of diagnosing and reporting dengue cases in places where health information systems are underdeveloped; (2) issues related to the use of socioeconomic proxies in places with limited dengue incidence data; (3) mosquito ranges which may be changing as a result of climate changes; and (4) the challenges of mapping dengue events at a variety of scales. An ideal dengue map will present endemic and epidemic dengue information from both rural and urban areas. Overcoming the current barriers requires expanded collaboration and data sharing by geographers, epidemiologists, and entomologists. Enhanced mapping techniques would allow for improved visualizations of dengue rates and risks.
VMSbase: an R-package for VMS and logbook data management and analysis in fisheries ecology.
Russo, Tommaso; D'Andrea, Lorenzo; Parisi, Antonio; Cataudella, Stefano
2014-01-01
VMSbase is an R package devised to manage, process and visualize information about fishing vessels activity (provided by the vessel monitoring system--VMS) and catches/landings (as reported in the logbooks). VMSbase is primarily conceived to be user-friendly; to this end, a suite of state-of-the-art analyses is accessible via a graphical interface. In addition, the package uses a database platform allowing large datasets to be stored, managed and processed vey efficiently. Methodologies include data cleaning, that is removal of redundant or evidently erroneous records, and data enhancing, that is interpolation and merging with external data sources. In particular, VMSbase is able to estimate sea bottom depth for single VMS pings using an on-line connection to the National Oceanic and Atmospheric Administration (NOAA) database. It also allows VMS pings to be assigned to whatever geographic partitioning has been selected by users. Standard analyses comprise: 1) métier identification (using a modified CLARA clustering approach on Logbook data or Artificial Neural Networks on VMS data); 2) linkage between VMS and Logbook records, with the former organized into fishing trips; 3) discrimination between steaming and fishing points; 4) computation of spatial effort with respect to user-selected grids; 5) calculation of standard fishing effort indicators within Data Collection Framework; 6) a variety of mapping tools, including an interface for Google viewer; 7) estimation of trawled area. Here we report a sample workflow for the accessory sample datasets (available with the package) in order to explore the potentialities of VMSbase. In addition, the results of some performance tests on two large datasets (1×10(5) and 1×10(6) VMS signals, respectively) are reported to inform about the time required for the analyses. The results, although merely illustrative, indicate that VMSbase can represent a step forward in extracting and enhancing information from VMS/logbook data for fisheries studies.
Metadata Creation, Management and Search System for your Scientific Data
NASA Astrophysics Data System (ADS)
Devarakonda, R.; Palanisamy, G.
2012-12-01
Mercury Search Systems is a set of tools for creating, searching, and retrieving of biogeochemical metadata. Mercury toolset provides orders of magnitude improvements in search speed, support for any metadata format, integration with Google Maps for spatial queries, multi-facetted type search, search suggestions, support for RSS (Really Simple Syndication) delivery of search results, and enhanced customization to meet the needs of the multiple projects that use Mercury. Mercury's metadata editor provides a easy way for creating metadata and Mercury's search interface provides a single portal to search for data and information contained in disparate data management systems, each of which may use any metadata format including FGDC, ISO-19115, Dublin-Core, Darwin-Core, DIF, ECHO, and EML. Mercury harvests metadata and key data from contributing project servers distributed around the world and builds a centralized index. The search interfaces then allow the users to perform a variety of fielded, spatial, and temporal searches across these metadata sources. This centralized repository of metadata with distributed data sources provides extremely fast search results to the user, while allowing data providers to advertise the availability of their data and maintain complete control and ownership of that data. Mercury is being used more than 14 different projects across 4 federal agencies. It was originally developed for NASA, with continuing development funded by NASA, USGS, and DOE for a consortium of projects. Mercury search won the NASA's Earth Science Data Systems Software Reuse Award in 2008. References: R. Devarakonda, G. Palanisamy, B.E. Wilson, and J.M. Green, "Mercury: reusable metadata management data discovery and access system", Earth Science Informatics, vol. 3, no. 1, pp. 87-94, May 2010. R. Devarakonda, G. Palanisamy, J.M. Green, B.E. Wilson, "Data sharing and retrieval using OAI-PMH", Earth Science Informatics DOI: 10.1007/s12145-010-0073-0, (2010);
Submicron mapping of strained silicon-on-insulator features induced
NASA Astrophysics Data System (ADS)
Murray, Conal E.; Sankarapandian, M.; Polvino, S. M.; Noyan, I. C.; Lai, B.; Cai, Z.
2007-04-01
Real-space maps of strain within silicon-on-insulator (SOI) features induced by adjacent, embedded shallow-trench-isolation (STI) SiO2 regions were obtained using x-ray microbeam diffraction. The quantitative strain mapping indicated that the SOI strain was largest at the SOI/STI interface and decreased as a function of distance from this interface. An out-of-plane residual strain of approximately -31μɛ was observed in the blanket regions of the SOI. A comparison of the depth-averaged strain distributions to the strain profiles calculated from an Eshelby inclusion model indicated an equivalent eigenstrain of -0.55% in the STI regions acting on the SOI features.
Cropland Capture: A Game to Improve Global Cropland through Crowdsourcing
NASA Astrophysics Data System (ADS)
Fritz, Steffen; Sturn, Tobias; See, Linda; Perger, Christoph; Schill, Christian; McCallum, Ian; Schepaschenko, Dmitry; Karner, Mathias; Dueruer, Martina; Kraxner, Florian; Obersteiner, Michael
2014-05-01
Accurate and reliable global cropland extent maps are essential for estimating and forecasting crop yield, in particular losses due to drought and production anomalies. Major questions surrounding energy futures and environmental change (EU and US biofuel target setting, determination of greenhouse gas emissions, REDD initiatives, and implications of climate change on crop production and productivity patterns) also require reliable information on the spatial distribution of cropland as well as crop types. Although global land cover maps identify cropland (which exist as one or more land cover categories), this information is currently not accurate enough for many applications. There are several ways of improving current cropland extent though hybrid approaches and by integrating information collected though Geo-Wiki (a global crowdsourcing platform) from very high resolution imagery such as that found on Google Earth. Another way of getting improved cropland extent maps would be to classify all very high resolution images found on Google Earth and to create a wall-to-wall map of cropland. This is a very ambitious task that would require a large number of individuals, like that found in massive multiplayer online games. For this reason we have developed a game called 'Cropland Capture'. The game can be played on a desktop, on a tablet (iPad or Android) or mobile phone (iPhone or Android) where the game mechanics are very simple. The player is provided with a satellite image or in-situ photo and they must determine if the image contains cropland or not. The game was launched in the middle of November 2013 and will run for 6 months, after which the weekly winners will be entered into a draw to win large prizes. To date we have collected more than 2.5 million areas, where we will continue to expand the sample to more locations around the world. Eventually the data will be used to calibrate and validate a new version of our global cropland map, where the latest version is available from http://beta-hybrid.geo-wiki.org. If we find, however, that a large number of people participate in the game, we will aim to make wall-to-wall cropland maps for those countries where no national maps exist. This paper will present an overview of the game and a summary of the crowdsourced data from the game, including information about quality and user performance. If successful, this gaming approach could be used to gather information about other land cover types in the future in order to improve global land cover information more generally.
Hyperspectral Soil Mapper (HYSOMA) software interface: Review and future plans
NASA Astrophysics Data System (ADS)
Chabrillat, Sabine; Guillaso, Stephane; Eisele, Andreas; Rogass, Christian
2014-05-01
With the upcoming launch of the next generation of hyperspectral satellites that will routinely deliver high spectral resolution images for the entire globe (e.g. EnMAP, HISUI, HyspIRI, HypXIM, PRISMA), an increasing demand for the availability/accessibility of hyperspectral soil products is coming from the geoscience community. Indeed, many robust methods for the prediction of soil properties based on imaging spectroscopy already exist and have been successfully used for a wide range of soil mapping airborne applications. Nevertheless, these methods require expert know-how and fine-tuning, which makes them used sparingly. More developments are needed toward easy-to-access soil toolboxes as a major step toward the operational use of hyperspectral soil products for Earth's surface processes monitoring and modelling, to allow non-experienced users to obtain new information based on non-expensive software packages where repeatability of the results is an important prerequisite. In this frame, based on the EU-FP7 EUFAR (European Facility for Airborne Research) project and EnMAP satellite science program, higher performing soil algorithms were developed at the GFZ German Research Center for Geosciences as demonstrators for end-to-end processing chains with harmonized quality measures. The algorithms were built-in into the HYSOMA (Hyperspectral SOil MApper) software interface, providing an experimental platform for soil mapping applications of hyperspectral imagery that gives the choice of multiple algorithms for each soil parameter. The software interface focuses on fully automatic generation of semi-quantitative soil maps such as soil moisture, soil organic matter, iron oxide, clay content, and carbonate content. Additionally, a field calibration option calculates fully quantitative soil maps provided ground truth soil data are available. Implemented soil algorithms have been tested and validated using extensive in-situ ground truth data sets. The source of the HYSOMA code was developed as standalone IDL software to allow easy implementation in the hyperspectral and non-hyperspectral communities. Indeed, within the hyperspectral community, IDL language is very widely used, and for non-expert users that do not have an ENVI license, such software can be executed as a binary version using the free IDL virtual machine under various operating systems. Based on the growing interest of users in the software interface, the experimental software was adapted for public release version in 2012, and since then ~80 users of hyperspectral soil products downloaded the soil algorithms at www.gfz-potsdam.de/hysoma. The software interface was distributed for free as IDL plug-ins under the IDL-virtual machine. Up-to-now distribution of HYSOMA was based on a close source license model, for non-commercial and educational purposes. Currently, the HYSOMA is being under further development in the context of the EnMAP satellite mission, for extension and implementation in the EnMAP Box as EnSoMAP (EnMAP SOil MAPper). The EnMAP Box is a freely available, platform-independent software distributed under an open source license. In the presentation we will focus on an update of the HYSOMA software interface status and upcoming implementation in the EnMAP Box. Scientific software validation, associated publication record and users responses as well as software management and transition to open source will be discussed.
Control-display mapping in brain-computer interfaces.
Thurlings, Marieke E; van Erp, Jan B F; Brouwer, Anne-Marie; Blankertz, Benjamin; Werkhoven, Peter
2012-01-01
Event-related potential (ERP) based brain-computer interfaces (BCIs) employ differences in brain responses to attended and ignored stimuli. When using a tactile ERP-BCI for navigation, mapping is required between navigation directions on a visual display and unambiguously corresponding tactile stimuli (tactors) from a tactile control device: control-display mapping (CDM). We investigated the effect of congruent (both display and control horizontal or both vertical) and incongruent (vertical display, horizontal control) CDMs on task performance, the ERP and potential BCI performance. Ten participants attended to a target (determined via CDM), in a stream of sequentially vibrating tactors. We show that congruent CDM yields best task performance, enhanced the P300 and results in increased estimated BCI performance. This suggests a reduced availability of attentional resources when operating an ERP-BCI with incongruent CDM. Additionally, we found an enhanced N2 for incongruent CDM, which indicates a conflict between visual display and tactile control orientations. Incongruency in control-display mapping reduces task performance. In this study, brain responses, task and system performance are related to (in)congruent mapping of command options and the corresponding stimuli in a brain-computer interface (BCI). Directional congruency reduces task errors, increases available attentional resources, improves BCI performance and thus facilitates human-computer interaction.
SWATMOD-PREP: Graphical user interface for preparing coupled SWAT-modflow simulations
USDA-ARS?s Scientific Manuscript database
This paper presents SWATMOD-Prep, a graphical user interface that couples a SWAT watershed model with a MODFLOW groundwater flow model. The interface is based on a recently published SWAT-MODFLOW code that couples the models via mapping schemes. The spatial layout of SWATMOD-Prep guides the user t...
Extra-terra incognita: Martian maps in the digital age.
Messeri, Lisa
2017-02-01
Science and technology studies (STS) and critical cartography are both asking questions about the ontological fixity of maps and other scientific objects. This paper examines how a group of NASA computer scientists who call themselves The Mapmakers conceptualizes and creates maps in service of different commitments. The maps under construction are those of alien Mars, produced through partnerships that NASA has established with Google and Microsoft. With the goal of bringing an experience of Mars to as many people as possible, these maps influence how we imagine our neighbouring planet. This paper analyzes two attributes of the map, evident in both its representation and the attending cartographic practices: a sense of Mars as dynamic and a desire for a democratic experience of Mars in which up-to-date Mars data can be intuitively accessed not only by scientists but by lay users as well. Whereas a democratic Mars promises users the ability to decide how to interact with the map and understand Mars, dynamic Mars imposes a more singular sense of Mars as a target of continued robotic and maybe even human exploration. Because maps of Mars have a different (and arguably less complex) set of social and political commitments than those of Earth, they help us see how different goals contradict and complement each other in matters of exploration and state-craft relevant both to other worlds and our own.
The Cellular Automata for modelling of spreading of lava flow on the earth surface
NASA Astrophysics Data System (ADS)
Jarna, A.
2012-12-01
Volcanic risk assessment is a very important scientific, political and economic issue in densely populated areas close to active volcanoes. Development of effective tools for early prediction of a potential volcanic hazard and management of crises are paramount. However, to this date volcanic hazard maps represent the most appropriate way to illustrate the geographical area that can potentially be affected by a volcanic event. Volcanic hazard maps are usually produced by mapping out old volcanic deposits, however dynamic lava flow simulation gaining popularity and can give crucial information to corroborate other methodologies. The methodology which is used here for the generation of volcanic hazard maps is based on numerical simulation of eruptive processes by the principle of Cellular Automata (CA). The python script is integrated into ArcToolbox in ArcMap (ESRI) and the user can select several input and output parameters which influence surface morphology, size and shape of the flow, flow thickness, flow velocity and length of lava flows. Once the input parameters are selected, the software computes and generates hazard maps on the fly. The results can be exported to Google Maps (.klm format) to visualize the results of the computation. For validation of the simulation code are used data from a real lava flow. Comparison of the simulation results with real lava flows mapped out from satellite images will be presented.
Using Google Earth to Explore Multiple Data Sets and Plate Tectonic Concepts
NASA Astrophysics Data System (ADS)
Goodell, L. P.
2015-12-01
Google Earth (GE) offers an engaging and dynamic environment for exploration of earth science data. While GIS software offers higher-level analytical capability, it comes with a steep learning curve and complex interface that is not easy for the novice, and in many cases the instructor, to negotiate. In contrast, the intuitive interface of GE makes it easy for students to quickly become proficient in manipulating the globe and independently exploring relationships between multiple data sets at a wide range of scales. Inquiry-based, data-rich exercises have been developed for both introductory and upper-level activities including: exploration of plate boundary characteristics and relative motion across plate boundaries; determination and comparison of short-term and long-term average plate velocities; crustal strain analysis (modeled after the UNAVCO activity); and determining earthquake epicenters, body-wave magnitudes, and focal plane solutions. Used successfully in undergraduate course settings, for TA training and for professional development programs for middle and high school teachers, the exercises use the following GE data sets (with sources) that have been collected/compiled by the author and are freely available for non-commercial use: 1) tectonic plate boundaries and plate names (Bird, 2003 model); 2) real-time earthquakes (USGS); 3) 30 years of M>=5.0 earthquakes, plotted by depth (USGS); 4) seafloor age (Mueller et al., 1997, 2008); 5) location and age data for hot spot tracks (published literature); 6) Holocene volcanoes (Smithsonian Global Volcanism Program); 7) GPS station locations with links to times series (JPL, NASA, UNAVCO); 8) short-term motion vectors derived from GPS times series; 9) long-term average motion vectors derived from plate motion models (UNAVCO plate motion calculator); 10) earthquake data sets consisting of seismic station locations and links to relevant seismograms (Rapid Earthquake Viewer, USC/IRIS/DELESE).
The wildland-urban interface raster dataset of Catalonia.
Alcasena, Fermín J; Evers, Cody R; Vega-Garcia, Cristina
2018-04-01
We provide the wildland urban interface (WUI) map of the autonomous community of Catalonia (Northeastern Spain). The map encompasses an area of some 3.21 million ha and is presented as a 150-m resolution raster dataset. Individual housing location, structure density and vegetation cover data were used to spatially assess in detail the interface, intermix and dispersed rural WUI communities with a geographical information system. Most WUI areas concentrate in the coastal belt where suburban sprawl has occurred nearby or within unmanaged forests. This geospatial information data provides an approximation of residential housing potential for loss given a wildfire, and represents a valuable contribution to assist landscape and urban planning in the region.
Can Satellite Remote Sensing be Applied in Geological Mapping in Tropics?
NASA Astrophysics Data System (ADS)
Magiera, Janusz
2018-03-01
Remote sensing (RS) techniques are based on spectral data registered by RS scanners as energy reflected from the Earth's surface or emitted by it. In "geological" RS the reflectance (or emittence) should come from rock or sediment. The problem in tropical and subtropical areas is a dense vegetation. Spectral response from the rocks and sediments is gathered only from the gaps among the trees and shrubs. Images of high resolution are appreciated here, therefore. New generation of satellites and scanners (Digital Globe WV2, WV3 and WV4) yield imagery of spatial resolution of 2 m and up to 16 spectral bands (WV3). Images acquired by Landsat (TM, ETM+, OLI) and Sentinel 2 have good spectral resolution too (6-12 bands in visible and infrared) and, despite lower spatial resolution (10-60 m of pixel size) are useful in extracting lithological information too. Lithological RS map may reveal good precision (down to a single rock or outcrop of a meter size). Supplemented with the analysis of Digital Elevation Model and high resolution ortophotomaps (Google Maps, Bing etc.) allows for quick and cheap mapping of unsurveyed areas.
Curtis, Andrew; Buritica, Paola; Ajayakumar, Jayakrishnan; Squires, Robert; Dávalos, Diana; Pacheco, Robinson; Bhatta, Madhav P.; James, Mark A.
2017-01-01
Background Cali, Colombia has experienced chikungunya and Zika outbreaks and hypoendemic dengue. Studies have explained Cali’s dengue patterns but lack the sub-neighborhood-scale detail investigated here. Methods Spatial-video geonarratives (SVG) with Ministry of Health officials and Community Health Workers were collected in hotspots, providing perspective on perceptions of why dengue, chikungunya and Zika hotspots exist, impediments to control, and social outcomes. Using spatial video and Google Street View, sub-neighborhood features possibly contributing to incidence were mapped to create risk surfaces, later compared with dengue, chikungunya and Zika case data. Results SVG captured insights in 24 neighborhoods. Trash and water risks in Calipso were mapped using SVG results. Perceived risk factors included proximity to standing water, canals, poverty, invasions, localized violence and military migration. These risks overlapped case density maps and identified areas that are suitable for transmission but are possibly underreporting to the surveillance system. Conclusion Resulting risk maps with local context could be leveraged to increase vector-control efficiency- targeting key areas of environmental risk. PMID:28767730
NASA Astrophysics Data System (ADS)
Bouiflane, Mustapha; Manar, Ahmed; Medina, Fida; Youbi, Nasrrddine; Rimi, Abdelkrim
2017-06-01
A high-resolution aeromagnetic survey was carried out in the Anti- Atlas, Morocco covering the main areas traversed by the Great CAMP Foum Zguid dyke (FZD). This ;doleritic; dyke belongs to the Central Atlantic Magmatic Province (CAMP), a Large Igneous Province which is associated with the fragmentation of the supercontinent Pangaea and the initial stages of rifting of the Central Atlantic Ocean. It also coincides in time with the mass extinction of the Triassic - Jurassic boundary. Based on the study of geological maps and Google Earth satellite images, it appears that the FZD is poorly exposed and, often covered by Quaternary deposits. This work proposes aeromagnetic modelling and interpretation of the FZD in order to better constrain its structural extent. The data have allowed (i) mapping of the dyke over great distances, under the Quaternary deposits and through areas where it was poorly characterized on the geological map; (ii) identifying major tectonic lineaments interpreted as faults; (iii) recognizing magnetic anomalies related to mafic intrusive bodies; and (iv) informing about regional structural context.
Virtual GEOINT Center: C2ISR through an avatar's eyes
NASA Astrophysics Data System (ADS)
Seibert, Mark; Tidbal, Travis; Basil, Maureen; Muryn, Tyler; Scupski, Joseph; Williams, Robert
2013-05-01
As the number of devices collecting and sending data in the world are increasing, finding ways to visualize and understand that data is becoming more and more of a problem. This has often been coined as the problem of "Big Data." The Virtual Geoint Center (VGC) aims to aid in solving that problem by providing a way to combine the use of the virtual world with outside tools. Using open-source software such as OpenSim and Blender, the VGC uses a visually stunning 3D environment to display the data sent to it. The VGC is broken up into two major components: The Kinect Minimap, and the Geoint Map. The Kinect Minimap uses the Microsoft Kinect and its open-source software to make a miniature display of people the Kinect detects in front of it. The Geoint Map collect smartphone sensor information from online databases and displays them in real time onto a map generated by Google Maps. By combining outside tools and the virtual world, the VGC can help a user "visualize" data, and provide additional tools to "understand" the data.
Direct k-space mapping of the electronic structure in an oxide-oxide interface.
Berner, G; Sing, M; Fujiwara, H; Yasui, A; Saitoh, Y; Yamasaki, A; Nishitani, Y; Sekiyama, A; Pavlenko, N; Kopp, T; Richter, C; Mannhart, J; Suga, S; Claessen, R
2013-06-14
The interface between LaAlO(3) and SrTiO(3) hosts a two-dimensional electron system of itinerant carriers, although both oxides are band insulators. Interface ferromagnetism coexisting with superconductivity has been found and attributed to local moments. Experimentally, it has been established that Ti 3d electrons are confined to the interface. Using soft x-ray angle-resolved resonant photoelectron spectroscopy we have directly mapped the interface states in k space. Our data demonstrate a charge dichotomy. A mobile fraction contributes to Fermi surface sheets, whereas a localized portion at higher binding energies is tentatively attributed to electrons trapped by O vacancies in the SrTiO(3). While photovoltage effects in the polar LaAlO(3) layers cannot be excluded, the apparent absence of surface-related Fermi surface sheets could also be fully reconciled in a recently proposed electronic reconstruction picture where the built-in potential in the LaAlO(3) is compensated by surface O vacancies serving also as a charge reservoir.
MapApp: A Java(TM) Applet for Accessing Geographic Databases
NASA Astrophysics Data System (ADS)
Haxby, W.; Carbotte, S.; Ryan, W. B.; OHara, S.
2001-12-01
MapApp (http://coast.ldeo.columbia.edu/help/MapApp.html) is a prototype Java(TM) applet that is intended to give easy and versatile access to geographic data sets through a web browser. It was developed initially to interface with the RIDGE Multibeam Synthesis. Subsequently, interfaces with other geophysical databases were added. At present, multibeam bathymetry grids, underway geophysics along ship tracks, and the LDEO Borehole Research Group's ODP well logging database are accessible through MapApp. We plan to add an interface with the Ridge Petrology Database in the near future. The central component of MapApp is a world physiographic map. Users may navigate around the map (zoom/pan) without waiting for HTTP requests to a remote server to be processed. A focus request loads image tiles from the server to compose a new map at the current viewing resolution. Areas in which multibeam grids are available may be focused to a pixel resolution of about 200 m. These areas may be identified by toggling a mask. Databases may be accessed through menus, and selected data objects may be loaded into MapApp by selecting items from tables. Once loaded, a bathymetry grid may be contoured or used to create bathymetric profiles; ship tracks and ODP sites may be overlain on the map and their geophysical data plotted in X-Y graphs. The advantage of applets over traditional web pages is that they permit dynamic interaction with data sets, while limiting time consuming interaction with a remote server. Users may customize the graphics display by modifying the scale, or the symbol or line characteristics of rendered data, contour interval, etc. The ease with which users can select areas, view the physiography of areas, and preview data sets and evaluate them for quality and applicability, makes MapApp a valuable tool for education and research.
Analysis of hand contact areas and interaction capabilities during manipulation and exploration.
Gonzalez, Franck; Gosselin, Florian; Bachta, Wael
2014-01-01
Manual human-computer interfaces for virtual reality are designed to allow an operator interacting with a computer simulation as naturally as possible. Dexterous haptic interfaces are the best suited for this goal. They give intuitive and efficient control on the environment with haptic and tactile feedback. This paper is aimed at helping in the choice of the interaction areas to be taken into account in the design of such interfaces. The literature dealing with hand interactions is first reviewed in order to point out the contact areas involved in exploration and manipulation tasks. Their frequencies of use are then extracted from existing recordings. The results are gathered in an original graphical interaction map allowing for a simple visualization of the way the hand is used, and compared with a map of mechanoreceptors densities. Then an interaction tree, mapping the relative amount of actions made available through the use of a given contact area, is built and correlated with the losses of hand function induced by amputations. A rating of some existing haptic interfaces and guidelines for their design are finally achieved to illustrate a possible use of the developed graphical tools.
Mapping atomic contact between pentacene and a Au surface using scanning tunneling spectroscopy.
Song, Young Jae; Lee, Kyuho; Kim, Seong Heon; Choi, Byoung-Young; Yu, Jaejun; Kuk, Young
2010-03-10
We mapped spatially varying intramolecular electronic structures on a pentacene-gold interface using scanning tunneling spectroscopy. Along with ab initio calculations based on density functional theory, we found that the directional nature of the d orbitals of Au atoms plays an important role in the interaction at the pentacene-gold contact. The gold-induced interface states are broadened and shifted by various pentacene-gold distances determined by the various registries of a pentacene molecule on a gold substrate.
A Different Web-Based Geocoding Service Using Fuzzy Techniques
NASA Astrophysics Data System (ADS)
Pahlavani, P.; Abbaspour, R. A.; Zare Zadiny, A.
2015-12-01
Geocoding - the process of finding position based on descriptive data such as address or postal code - is considered as one of the most commonly used spatial analyses. Many online map providers such as Google Maps, Bing Maps and Yahoo Maps present geocoding as one of their basic capabilities. Despite the diversity of geocoding services, users usually face some limitations when they use available online geocoding services. In existing geocoding services, proximity and nearness concept is not modelled appropriately as well as these services search address only by address matching based on descriptive data. In addition there are also some limitations in display searching results. Resolving these limitations can enhance efficiency of the existing geocoding services. This paper proposes the idea of integrating fuzzy technique with geocoding process to resolve these limitations. In order to implement the proposed method, a web-based system is designed. In proposed method, nearness to places is defined by fuzzy membership functions and multiple fuzzy distance maps are created. Then these fuzzy distance maps are integrated using fuzzy overlay technique for obtain the results. Proposed methods provides different capabilities for users such as ability to search multi-part addresses, searching places based on their location, non-point representation of results as well as displaying search results based on their priority.
The Status of Topographic Mapping in the World a Unggim-Isprs Project 2012-2015
NASA Astrophysics Data System (ADS)
Konecny, G.; Breitkopf, U.; Radtke, A.
2016-06-01
In December 2011, UNGGIM initiated a cooperative project with ISPRS to resume the former UN Secretariat studies on the status of topographic mapping in the world, conducted between 1968 and 1986. After the design of a questionnaire with 27 questions, the UNGGIM Secretariat sent the questionnaires to the UN member states. 115 replies were received from the 193 member states and regions thereof. Regarding the global data coverage and age, the UN questionnaire survey was supplemented by data from the Eastview database. For each of the 27 questions, an interactive viewer was programmed permitting the analysis of the results. The authoritative data coverage at the various scale ranges has greatly increased between 1986 and 2012. Now, a 30 % 1 : 25 000 map data coverage and a 75 % 1 : 50 000 map data coverage has been completed. Nevertheless, there is still an updating problem, as data for some countries is 10 to 30 years old. Private Industry, with Google, Microsoft and Navigation system providers, have undertaken huge efforts to supplement authoritative mapping. For critical areas on the globe, MGCP committed to military mapping at 1 : 50 000. ISPRS has decided to make such surveys a sustainable issue by establishing a working group.
Wang, Jie; Xiao, Xiangming; Qin, Yuanwei; Dong, Jinwei; Zhang, Geli; Kou, Weili; Jin, Cui; Zhou, Yuting; Zhang, Yao
2015-05-12
As farmland systems vary over space and time (season and year), accurate and updated maps of paddy rice are needed for studies of food security and environmental problems. We selected a wheat-rice double-cropped area from fragmented landscapes along the rural-urban complex (Jiangsu Province, China) and explored the potential utility of integrating time series optical images (Landsat-8, MODIS) and radar images (PALSAR) in mapping paddy rice planting areas. We first identified several main types of non-cropland land cover and then identified paddy rice fields by selecting pixels that were inundated only during paddy rice flooding periods. These key temporal windows were determined based on MODIS Land Surface Temperature and vegetation indices. The resultant paddy rice map was evaluated using regions of interest (ROIs) drawn from multiple high-resolution images, Google Earth, and in-situ cropland photos. The estimated overall accuracy and Kappa coefficient were 89.8% and 0.79, respectively. In comparison with the National Land Cover Data (China) from 2010, the resultant map better detected changes in the paddy rice fields and revealed more details about their distribution. These results demonstrate the efficacy of using images from multiple sources to generate paddy rice maps for two-crop rotation systems.