Sample records for web log mining

  1. The design and implementation of web mining in web sites security

    NASA Astrophysics Data System (ADS)

    Li, Jian; Zhang, Guo-Yin; Gu, Guo-Chang; Li, Jian-Li

    2003-06-01

    The backdoor or information leak of Web servers can be detected by using Web Mining techniques on some abnormal Web log and Web application log data. The security of Web servers can be enhanced and the damage of illegal access can be avoided. Firstly, the system for discovering the patterns of information leakages in CGI scripts from Web log data was proposed. Secondly, those patterns for system administrators to modify their codes and enhance their Web site security were provided. The following aspects were described: one is to combine web application log with web log to extract more information, so web data mining could be used to mine web log for discovering the information that firewall and Information Detection System cannot find. Another approach is to propose an operation module of web site to enhance Web site security. In cluster server session, Density-Based Clustering technique is used to reduce resource cost and obtain better efficiency.

  2. Web usage data mining agent

    NASA Astrophysics Data System (ADS)

    Madiraju, Praveen; Zhang, Yanqing

    2002-03-01

    When a user logs in to a website, behind the scenes the user leaves his/her impressions, usage patterns and also access patterns in the web servers log file. A web usage mining agent can analyze these web logs to help web developers to improve the organization and presentation of their websites. They can help system administrators in improving the system performance. Web logs provide invaluable help in creating adaptive web sites and also in analyzing the network traffic analysis. This paper presents the design and implementation of a Web usage mining agent for digging in to the web log files.

  3. Effect of Temporal Relationships in Associative Rule Mining for Web Log Data

    PubMed Central

    Mohd Khairudin, Nazli; Mustapha, Aida

    2014-01-01

    The advent of web-based applications and services has created such diverse and voluminous web log data stored in web servers, proxy servers, client machines, or organizational databases. This paper attempts to investigate the effect of temporal attribute in relational rule mining for web log data. We incorporated the characteristics of time in the rule mining process and analysed the effect of various temporal parameters. The rules generated from temporal relational rule mining are then compared against the rules generated from the classical rule mining approach such as the Apriori and FP-Growth algorithms. The results showed that by incorporating the temporal attribute via time, the number of rules generated is subsequently smaller but is comparable in terms of quality. PMID:24587757

  4. Working with Data: Discovering Knowledge through Mining and Analysis; Systematic Knowledge Management and Knowledge Discovery; Text Mining; Methodological Approach in Discovering User Search Patterns through Web Log Analysis; Knowledge Discovery in Databases Using Formal Concept Analysis; Knowledge Discovery with a Little Perspective.

    ERIC Educational Resources Information Center

    Qin, Jian; Jurisica, Igor; Liddy, Elizabeth D.; Jansen, Bernard J; Spink, Amanda; Priss, Uta; Norton, Melanie J.

    2000-01-01

    These six articles discuss knowledge discovery in databases (KDD). Topics include data mining; knowledge management systems; applications of knowledge discovery; text and Web mining; text mining and information retrieval; user search patterns through Web log analysis; concept analysis; data collection; and data structure inconsistency. (LRW)

  5. Web Mining for Web Image Retrieval.

    ERIC Educational Resources Information Center

    Chen, Zheng; Wenyin, Liu; Zhang, Feng; Li, Mingjing; Zhang, Hongjiang

    2001-01-01

    Presents a prototype system for image retrieval from the Internet using Web mining. Discusses the architecture of the Web image retrieval prototype; document space modeling; user log mining; and image retrieval experiments to evaluate the proposed system. (AEF)

  6. Visualization of usability and functionality of a professional website through web-mining.

    PubMed

    Jones, Josette F; Mahoui, Malika; Gopa, Venkata Devi Pragna

    2007-10-11

    Functional interface design requires understanding of the information system structure and the user. Web logs record user interactions with the interface, and thus provide some insight into user search behavior and efficiency of the search process. The present study uses a data-mining approach with techniques such as association rules, clustering and classification, to visualize the usability and functionality of a digital library through in depth analyses of web logs.

  7. Web usage mining at an academic health sciences library: an exploratory study.

    PubMed

    Bracke, Paul J

    2004-10-01

    This paper explores the potential of multinomial logistic regression analysis to perform Web usage mining for an academic health sciences library Website. Usage of database-driven resource gateway pages was logged for a six-month period, including information about users' network addresses, referring uniform resource locators (URLs), and types of resource accessed. It was found that referring URL did vary significantly by two factors: whether a user was on-campus and what type of resource was accessed. Although the data available for analysis are limited by the nature of the Web and concerns for privacy, this method demonstrates the potential for gaining insight into Web usage that supplements Web log analysis. It can be used to improve the design of static and dynamic Websites today and could be used in the design of more advanced Web systems in the future.

  8. A Clustering Methodology of Web Log Data for Learning Management Systems

    ERIC Educational Resources Information Center

    Valsamidis, Stavros; Kontogiannis, Sotirios; Kazanidis, Ioannis; Theodosiou, Theodosios; Karakos, Alexandros

    2012-01-01

    Learning Management Systems (LMS) collect large amounts of data. Data mining techniques can be applied to analyse their web data log files. The instructors may use this data for assessing and measuring their courses. In this respect, we have proposed a methodology for analysing LMS courses and students' activity. This methodology uses a Markov…

  9. Mining Formative Evaluation Rules Using Web-Based Learning Portfolios for Web-Based Learning Systems

    ERIC Educational Resources Information Center

    Chen, Chih-Ming; Hong, Chin-Ming; Chen, Shyuan-Yi; Liu, Chao-Yu

    2006-01-01

    Learning performance assessment aims to evaluate what knowledge learners have acquired from teaching activities. Objective technical measures of learning performance are difficult to develop, but are extremely important for both teachers and learners. Learning performance assessment using learning portfolios or web server log data is becoming an…

  10. Online Persistence in Higher Education Web-Supported Courses

    ERIC Educational Resources Information Center

    Hershkovitz, Arnon; Nachmias, Rafi

    2011-01-01

    This research consists of an empirical study of online persistence in Web-supported courses in higher education, using Data Mining techniques. Log files of 58 Moodle websites accompanying Tel Aviv University courses were drawn, recording the activity of 1189 students in 1897 course enrollments during the academic year 2008/9, and were analyzed…

  11. Query Classification and Study of University Students' Search Trends

    ERIC Educational Resources Information Center

    Maabreh, Majdi A.; Al-Kabi, Mohammed N.; Alsmadi, Izzat M.

    2012-01-01

    Purpose: This study is an attempt to develop an automatic identification method for Arabic web queries and divide them into several query types using data mining. In addition, it seeks to evaluate the impact of the academic environment on using the internet. Design/methodology/approach: The web log files were collected from one of the higher…

  12. Data Mining for Web Site Evaluation: An Exploration of Site Usage by Graduate Social Work Students

    ERIC Educational Resources Information Center

    Santhiveeran, Janaki

    2006-01-01

    This paper evaluates the actual use of a course Website by graduate social work students. The study utilized data mining techniques to discover meaningful trends by using the data from server logs. The course Website was accessed 24,730 times by all 49 graduate students during a semester. The students utilized the course Website 23 hours a day, 7…

  13. Automated data mining: an innovative and efficient web-based approach to maintaining resident case logs.

    PubMed

    Bhattacharya, Pratik; Van Stavern, Renee; Madhavan, Ramesh

    2010-12-01

    Use of resident case logs has been considered by the Residency Review Committee for Neurology of the Accreditation Council for Graduate Medical Education (ACGME). This study explores the effectiveness of a data-mining program for creating resident logs and compares the results to a manual data-entry system. Other potential applications of data mining to enhancing resident education are also explored. Patient notes dictated by residents were extracted from the Hospital Information System and analyzed using an unstructured mining program. History, examination and ICD codes were obtained and compared to the existing manual log. The automated data History, examination, and ICD codes were gathered for a 30-day period and compared to manual case logs. The automated method extracted all resident dictations with the dates of encounter and transcription. The automated data-miner processed information from all 19 residents, while only 4 residents logged manually. The manual method identified only broad categories of diseases; the major categories were stroke or vascular disorder 53 (27.6%), epilepsy 28 (14.7%), and pain syndromes 26 (13.5%). In the automated method, epilepsy 114 (21.1%), cerebral atherosclerosis 114 (21.1%), and headache 105 (19.4%) were the most frequent primary diagnoses, and headache 89 (16.5%), seizures 94 (17.4%), and low back pain 47 (9%) were the most common chief complaints. More detailed patient information such as tobacco use 227 (42%), alcohol use 205 (38%), and drug use 38 (7%) were extracted by the data-mining method. Manual case logs are time-consuming, provide limited information, and may be unpopular with residents. Data mining is a time-effective tool that may aid in the assessment of resident experience or the ACGME core competencies or in resident clinical research. More study of this method in larger numbers of residency programs is needed.

  14. Three Interaction Patterns on Asynchronous Online Discussion Behaviours: A Methodological Comparison

    ERIC Educational Resources Information Center

    Jo, I.; Park, Y.; Lee, H.

    2017-01-01

    An asynchronous online discussion (AOD) is one format of instructional methods that facilitate student-centered learning. In the wealth of AOD research, this study evaluated how students' behavior on AOD influences their academic outcomes. This case study compared the differential analytic methods including web log mining, social network analysis…

  15. Using Cluster Analysis for Data Mining in Educational Technology Research

    ERIC Educational Resources Information Center

    Antonenko, Pavlo D.; Toy, Serkan; Niederhauser, Dale S.

    2012-01-01

    Cluster analysis is a group of statistical methods that has great potential for analyzing the vast amounts of web server-log data to understand student learning from hyperlinked information resources. In this methodological paper we provide an introduction to cluster analysis for educational technology researchers and illustrate its use through…

  16. Modeling User Behavior and Attention in Search

    ERIC Educational Resources Information Center

    Huang, Jeff

    2013-01-01

    In Web search, query and click log data are easy to collect but they fail to capture user behaviors that do not lead to clicks. As search engines reach the limits inherent in click data and are hungry for more data in a competitive environment, mining cursor movements, hovering, and scrolling becomes important. This dissertation investigates how…

  17. Text mining for adverse drug events: the promise, challenges, and state of the art.

    PubMed

    Harpaz, Rave; Callahan, Alison; Tamang, Suzanne; Low, Yen; Odgers, David; Finlayson, Sam; Jung, Kenneth; LePendu, Paea; Shah, Nigam H

    2014-10-01

    Text mining is the computational process of extracting meaningful information from large amounts of unstructured text. It is emerging as a tool to leverage underutilized data sources that can improve pharmacovigilance, including the objective of adverse drug event (ADE) detection and assessment. This article provides an overview of recent advances in pharmacovigilance driven by the application of text mining, and discusses several data sources-such as biomedical literature, clinical narratives, product labeling, social media, and Web search logs-that are amenable to text mining for pharmacovigilance. Given the state of the art, it appears text mining can be applied to extract useful ADE-related information from multiple textual sources. Nonetheless, further research is required to address remaining technical challenges associated with the text mining methodologies, and to conclusively determine the relative contribution of each textual source to improving pharmacovigilance.

  18. The new idea of transporting tailings-logs in tailings slurry pipeline and the innovation of technology of mining waste-fill method

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Lin Yu; Wang Fuji; Tao Yan

    2000-07-01

    This paper introduced a new idea of transporting mine tailings-logs in mine tailings-slurry pipeline and a new technology of mine cemented filing of tailings-logs with tailings-slurry. The hydraulic principles, the compaction of tailings-logs and the mechanic function of fillbody of tailings-logs cemented by tailings-slurry have been discussed.

  19. Resource Management Scheme Based on Ubiquitous Data Analysis

    PubMed Central

    Lee, Heung Ki; Jung, Jaehee

    2014-01-01

    Resource management of the main memory and process handler is critical to enhancing the system performance of a web server. Owing to the transaction delay time that affects incoming requests from web clients, web server systems utilize several web processes to anticipate future requests. This procedure is able to decrease the web generation time because there are enough processes to handle the incoming requests from web browsers. However, inefficient process management results in low service quality for the web server system. Proper pregenerated process mechanisms are required for dealing with the clients' requests. Unfortunately, it is difficult to predict how many requests a web server system is going to receive. If a web server system builds too many web processes, it wastes a considerable amount of memory space, and thus performance is reduced. We propose an adaptive web process manager scheme based on the analysis of web log mining. In the proposed scheme, the number of web processes is controlled through prediction of incoming requests, and accordingly, the web process management scheme consumes the least possible web transaction resources. In experiments, real web trace data were used to prove the improved performance of the proposed scheme. PMID:25197692

  20. Using Web Server Logs in Evaluating Instructional Web Sites.

    ERIC Educational Resources Information Center

    Ingram, Albert L.

    2000-01-01

    Web server logs contain a great deal of information about who uses a Web site and how they use it. This article discusses the analysis of Web logs for instructional Web sites; reviews the data stored in most Web server logs; demonstrates what further information can be gleaned from the logs; and discusses analyzing that information for the…

  1. Using Open Web APIs in Teaching Web Mining

    ERIC Educational Resources Information Center

    Chen, Hsinchun; Li, Xin; Chau, M.; Ho, Yi-Jen; Tseng, Chunju

    2009-01-01

    With the advent of the World Wide Web, many business applications that utilize data mining and text mining techniques to extract useful business information on the Web have evolved from Web searching to Web mining. It is important for students to acquire knowledge and hands-on experience in Web mining during their education in information systems…

  2. Web mining in soft computing framework: relevance, state of the art and future directions.

    PubMed

    Pal, S K; Talwar, V; Mitra, P

    2002-01-01

    The paper summarizes the different characteristics of Web data, the basic components of Web mining and its different types, and the current state of the art. The reason for considering Web mining, a separate field from data mining, is explained. The limitations of some of the existing Web mining methods and tools are enunciated, and the significance of soft computing (comprising fuzzy logic (FL), artificial neural networks (ANNs), genetic algorithms (GAs), and rough sets (RSs) are highlighted. A survey of the existing literature on "soft Web mining" is provided along with the commercially available systems. The prospective areas of Web mining where the application of soft computing needs immediate attention are outlined with justification. Scope for future research in developing "soft Web mining" systems is explained. An extensive bibliography is also provided.

  3. Who Goes There? Measuring Library Web Site Usage.

    ERIC Educational Resources Information Center

    Bauer, Kathleen

    2000-01-01

    Discusses how libraries can gather data on the use of their Web sites. Highlights include Web server log files, including the common log file, referrer log file, and agent log file; log file limitations; privacy concerns; and choosing log analysis software, both free and commercial. (LRW)

  4. The Value of Web Log Data in Use-based Design and Testing.

    ERIC Educational Resources Information Center

    Burton, Mary C.; Walther, Joseph B.

    2001-01-01

    Suggests Web-based logs contain useful empirical data with which World Wide Web designers and design theorists can assess usability and effectiveness of design choices. Enumerates identification of types of Web server logs, client logs, types and uses of log data, and issues associated with the validity of these data. Presents an approach to…

  5. Applying Web Usage Mining for Personalizing Hyperlinks in Web-Based Adaptive Educational Systems

    ERIC Educational Resources Information Center

    Romero, Cristobal; Ventura, Sebastian; Zafra, Amelia; de Bra, Paul

    2009-01-01

    Nowadays, the application of Web mining techniques in e-learning and Web-based adaptive educational systems is increasing exponentially. In this paper, we propose an advanced architecture for a personalization system to facilitate Web mining. A specific Web mining tool is developed and a recommender engine is integrated into the AHA! system in…

  6. Logging Residue Available for Mine-Timber Production

    Treesearch

    Floyd G. Timson

    1978-01-01

    Hardwood logging residue was examined as a source of raw material in the manufacture of sawn, split, and round timbers for use in underground coal mines. Forty-four percent of the total logging residue (residue !U 4 inches in diameter outside bark (dob), small end, and 4 feet long) from sawlog-only harvests was suitable for mine-timber production. Only 26 percent of...

  7. DOE Office of Scientific and Technical Information (OSTI.GOV)

    W. Lynn Watney; John H. Doveton

    GEMINI (Geo-Engineering Modeling through Internet Informatics) is a public-domain web application focused on analysis and modeling of petroleum reservoirs and plays (http://www.kgs.ukans.edu/Gemini/index.html). GEMINI creates a virtual project by ''on-the-fly'' assembly and analysis of on-line data either from the Kansas Geological Survey or uploaded from the user. GEMINI's suite of geological and engineering web applications for reservoir analysis include: (1) petrofacies-based core and log modeling using an interactive relational rock catalog and log analysis modules; (2) a well profile module; (3) interactive cross sections to display ''marked'' wireline logs; (4) deterministic gridding and mapping of petrophysical data; (5) calculation and mappingmore » of layer volumetrics; (6) material balance calculations; (7) PVT calculator; (8) DST analyst, (9) automated hydrocarbon association navigator (KHAN) for database mining, and (10) tutorial and help functions. The Kansas Hydrocarbon Association Navigator (KHAN) utilizes petrophysical databases to estimate hydrocarbon pay or other constituent at a play- or field-scale. Databases analyzed and displayed include digital logs, core analysis and photos, DST, and production data. GEMINI accommodates distant collaborations using secure password protection and authorized access. Assembled data, analyses, charts, and maps can readily be moved to other applications. GEMINI's target audience includes small independents and consultants seeking to find, quantitatively characterize, and develop subtle and bypassed pays by leveraging the growing base of digital data resources. Participating companies involved in the testing and evaluation of GEMINI included Anadarko, BP, Conoco-Phillips, Lario, Mull, Murfin, and Pioneer Resources.« less

  8. Fate and Trophic Transfer of Rare Earth Elements in Temperate Lake Food Webs.

    PubMed

    Amyot, Marc; Clayden, Meredith G; MacMillan, Gwyneth A; Perron, Tania; Arscott-Gauvin, Alexandre

    2017-06-06

    Many mining projects targeting rare earth elements (REE) are in development in North America, but the background concentrations and trophic transfer of these elements in natural environments have not been well characterized. We sampled abiotic and food web components in 14 Canadian temperate lakes unaffected by mines to assess the natural ecosystem fate of REE. Individual REE and total REE concentrations (sum of individual element concentrations, ΣREE) were strongly related with each other throughout different components of lake food webs. Dissolved organic carbon and dissolved oxygen in the water column, as well as ΣREE in sediments, were identified as potential drivers of aqueous ΣREE. Log 10 of median bioaccumulation factors ranged from 1.3, 3.7, 4.0, and 4.4 L/kg (wet weight) for fish muscle, zooplankton, predatory invertebrates, and nonpredatory invertebrates, respectively. [ΣREE] in fish, benthic macroinvertebrates, and zooplankton declined as a function of their trophic position, as determined by functional feeding groups and isotopic signatures of nitrogen (δ 15 N), indicating that REE were subject to trophic dilution. Low concentrations of REE in freshwater fish muscle compared to their potential invertebrate prey suggest that fish fillet consumption is unlikely to be a significant source of REE to humans in areas unperturbed by mining activities. However, other fish predators (e.g., piscivorous birds and mammals) may accumulate REE from whole fish as they are more concentrated than muscle. Overall, this study provides key information on the baseline concentrations and trophic patterns for REE in freshwater temperate lakes in Quebec, Canada.

  9. Study on online community user motif using web usage mining

    NASA Astrophysics Data System (ADS)

    Alphy, Meera; Sharma, Ajay

    2016-04-01

    The Web usage mining is the application of data mining, which is used to extract useful information from the online community. The World Wide Web contains at least 4.73 billion pages according to Indexed Web and it contains at least 228.52 million pages according Dutch Indexed web on 6th august 2015, Thursday. It’s difficult to get needed data from these billions of web pages in World Wide Web. Here is the importance of web usage mining. Personalizing the search engine helps the web user to identify the most used data in an easy way. It reduces the time consumption; automatic site search and automatic restore the useful sites. This study represents the old techniques to latest techniques used in pattern discovery and analysis in web usage mining from 1996 to 2015. Analyzing user motif helps in the improvement of business, e-commerce, personalisation and improvement of websites.

  10. A Framework for Web Usage Mining in Electronic Government

    NASA Astrophysics Data System (ADS)

    Zhou, Ping; Le, Zhongjian

    Web usage mining has been a major component of management strategy to enhance organizational analysis and decision. The literature on Web usage mining that deals with strategies and technologies for effectively employing Web usage mining is quite vast. In recent years, E-government has received much attention from researchers and practitioners. Huge amounts of user access data are produced in Electronic government Web site everyday. The role of these data in the success of government management cannot be overstated because they affect government analysis, prediction, strategies, tactical, operational planning and control. Web usage miming in E-government has an important role to play in setting government objectives, discovering citizen behavior, and determining future courses of actions. Web usage mining in E-government has not received adequate attention from researchers or practitioners. We developed a framework to promote a better understanding of the importance of Web usage mining in E-government. Using the current literature, we developed the framework presented herein, in hopes that it would stimulate more interest in this important area.

  11. Introduction to the JASIST Special Topic Issue on Web Retrieval and Mining: A Machine Learning Perspective.

    ERIC Educational Resources Information Center

    Chen, Hsinchun

    2003-01-01

    Discusses information retrieval techniques used on the World Wide Web. Topics include machine learning in information extraction; relevance feedback; information filtering and recommendation; text classification and text clustering; Web mining, based on data mining techniques; hyperlink structure; and Web size. (LRW)

  12. Visual Based Retrieval Systems and Web Mining--Introduction.

    ERIC Educational Resources Information Center

    Iyengar, S. S.

    2001-01-01

    Briefly discusses Web mining and image retrieval techniques, and then presents a summary of articles in this special issue. Articles focus on Web content mining, artificial neural networks as tools for image retrieval, content-based image retrieval systems, and personalizing the Web browsing experience using media agents. (AEF)

  13. West Virginia wood waste from uncharted sources: log landings and active surface mines

    Treesearch

    Shawn T. Grushecky; Lawrence E. Osborn

    2013-01-01

    Traditionally, biomass availability estimates from West Virginia have focused on primary and secondary mill byproducts and logging residues. Other sources of woody biomass are available that have not been surveyed. Through a series of field studies during 2010 and 2011, biomass availability estimates were developed for surface mine sites and log landings in West...

  14. Analysis of Hospital Processes with Process Mining Techniques.

    PubMed

    Orellana García, Arturo; Pérez Alfonso, Damián; Larrea Armenteros, Osvaldo Ulises

    2015-01-01

    Process mining allows for discovery, monitoring, and improving processes identified in information systems from their event logs. In hospital environments, process analysis has been a crucial factor for cost reduction, control and proper use of resources, better patient care, and achieving service excellence. This paper presents a new component for event logs generation in the Hospital Information System or HIS, developed at University of Informatics Sciences. The event logs obtained are used for analysis of hospital processes with process mining techniques. The proposed solution intends to achieve the generation of event logs in the system with high quality. The performed analyses allowed for redefining functions in the system and proposed proper flow of information. The study exposed the need to incorporate process mining techniques in hospital systems to analyze the processes execution. Moreover, we illustrate its application for making clinical and administrative decisions for the management of hospital activities.

  15. Using Web Server Logs to Track Users through the Electronic Forest

    ERIC Educational Resources Information Center

    Coombs, Karen A.

    2005-01-01

    This article analyzes server logs, providing helpful information in making decisions about Web-based services. The author indicates, as a result of analyzing server logs, several interesting things about the users' behavior were learned. The resulting findings are discussed in this article. Certain pages of the author's Web site, for instance, are…

  16. Analyzing Information Seeking and Drug-Safety Alert Response by Health Care Professionals as New Methods for Surveillance

    PubMed Central

    Pernek, Igor; Stiglic, Gregor; Leskovec, Jure; Strasberg, Howard R; Shah, Nigam Haresh

    2015-01-01

    Background Patterns in general consumer online search logs have been used to monitor health conditions and to predict health-related activities, but the multiple contexts within which consumers perform online searches make significant associations difficult to interpret. Physician information-seeking behavior has typically been analyzed through survey-based approaches and literature reviews. Activity logs from health care professionals using online medical information resources are thus a valuable yet relatively untapped resource for large-scale medical surveillance. Objective To analyze health care professionals’ information-seeking behavior and assess the feasibility of measuring drug-safety alert response from the usage logs of an online medical information resource. Methods Using two years (2011-2012) of usage logs from UpToDate, we measured the volume of searches related to medical conditions with significant burden in the United States, as well as the seasonal distribution of those searches. We quantified the relationship between searches and resulting page views. Using a large collection of online mainstream media articles and Web log posts we also characterized the uptake of a Food and Drug Administration (FDA) alert via changes in UpToDate search activity compared with general online media activity related to the subject of the alert. Results Diseases and symptoms dominate UpToDate searches. Some searches result in page views of only short duration, while others consistently result in longer-than-average page views. The response to an FDA alert for Celexa, characterized by a change in UpToDate search activity, differed considerably from general online media activity. Changes in search activity appeared later and persisted longer in UpToDate logs. The volume of searches and page view durations related to Celexa before the alert also differed from those after the alert. Conclusions Understanding the information-seeking behavior associated with online evidence sources can offer insight into the information needs of health professionals and enable large-scale medical surveillance. Our Web log mining approach has the potential to monitor responses to FDA alerts at a national level. Our findings can also inform the design and content of evidence-based medical information resources such as UpToDate. PMID:26293444

  17. Analyzing Information Seeking and Drug-Safety Alert Response by Health Care Professionals as New Methods for Surveillance.

    PubMed

    Callahan, Alison; Pernek, Igor; Stiglic, Gregor; Leskovec, Jure; Strasberg, Howard R; Shah, Nigam Haresh

    2015-08-20

    Patterns in general consumer online search logs have been used to monitor health conditions and to predict health-related activities, but the multiple contexts within which consumers perform online searches make significant associations difficult to interpret. Physician information-seeking behavior has typically been analyzed through survey-based approaches and literature reviews. Activity logs from health care professionals using online medical information resources are thus a valuable yet relatively untapped resource for large-scale medical surveillance. To analyze health care professionals' information-seeking behavior and assess the feasibility of measuring drug-safety alert response from the usage logs of an online medical information resource. Using two years (2011-2012) of usage logs from UpToDate, we measured the volume of searches related to medical conditions with significant burden in the United States, as well as the seasonal distribution of those searches. We quantified the relationship between searches and resulting page views. Using a large collection of online mainstream media articles and Web log posts we also characterized the uptake of a Food and Drug Administration (FDA) alert via changes in UpToDate search activity compared with general online media activity related to the subject of the alert. Diseases and symptoms dominate UpToDate searches. Some searches result in page views of only short duration, while others consistently result in longer-than-average page views. The response to an FDA alert for Celexa, characterized by a change in UpToDate search activity, differed considerably from general online media activity. Changes in search activity appeared later and persisted longer in UpToDate logs. The volume of searches and page view durations related to Celexa before the alert also differed from those after the alert. Understanding the information-seeking behavior associated with online evidence sources can offer insight into the information needs of health professionals and enable large-scale medical surveillance. Our Web log mining approach has the potential to monitor responses to FDA alerts at a national level. Our findings can also inform the design and content of evidence-based medical information resources such as UpToDate.

  18. Web Log Analysis: A Study of Instructor Evaluations Done Online

    ERIC Educational Resources Information Center

    Klassen, Kenneth J.; Smith, Wayne

    2004-01-01

    This paper focuses on developing a relatively simple method for analyzing web-logs. It also explores the challenges and benefits of web-log analysis. The study of student behavior on this site provides insights into website design and the effectiveness of this site in particular. Another benefit realized from the paper is the ease with which these…

  19. Evaluation of the Kloswall longwall mining system

    NASA Astrophysics Data System (ADS)

    Guay, P. J.

    1982-04-01

    A new longwal mining system specifically designed to extract a very deep web (48 inches or deeper) from a longwall panel was studied. Productivity and cost analysis comparing the new mining system with a conventional longwall operation taking a 30 inch wide web is presented. It is shown that the new system will increase annual production and return on investment in most cases. Conceptual drawings and specifications for a high capacity three drum shearer and a unique shield type of roof support specifically designed for very wide web operation are reported. The advantages and problems associated with wide web mining in general and as they relate specifically to the equipment selected for the new mining system are discussed.

  20. Web Mining: Machine Learning for Web Applications.

    ERIC Educational Resources Information Center

    Chen, Hsinchun; Chau, Michael

    2004-01-01

    Presents an overview of machine learning research and reviews methods used for evaluating machine learning systems. Ways that machine-learning algorithms were used in traditional information retrieval systems in the "pre-Web" era are described, and the field of Web mining and how machine learning has been used in different Web mining…

  1. User Centric Job Monitoring - a redesign and novel approach in the STAR experiment

    NASA Astrophysics Data System (ADS)

    Arkhipkin, D.; Lauret, J.; Zulkarneeva, Y.

    2014-06-01

    User Centric Monitoring (or UCM) has been a long awaited feature in STAR, whereas programs, workflows and system "events" could be logged, broadcast and later analyzed. UCM allows to collect and filter available job monitoring information from various resources and present it to users in a user-centric view rather than an administrative-centric point of view. The first attempt and implementation of "a" UCM approach was made in STAR 2004 using a log4cxx plug-in back-end and then further evolved with an attempt to push toward a scalable database back-end (2006) and finally using a Web-Service approach (2010, CSW4DB SBIR). The latest showed to be incomplete and not addressing the evolving needs of the experiment where streamlined messages for online (data acquisition) purposes as well as the continuous support for the data mining needs and event analysis need to coexists and unified in a seamless approach. The code also revealed to be hardly maintainable. This paper presents the next evolutionary step of the UCM toolkit, a redesign and redirection of our latest attempt acknowledging and integrating recent technologies and a simpler, maintainable and yet scalable manner. The extended version of the job logging package is built upon three-tier approach based on Task, Job and Event, and features a Web-Service based logging API, a responsive AJAX-powered user interface, and a database back-end relying on MongoDB, which is uniquely suited for STAR needs. In addition, we present details of integration of this logging package with the STAR offline and online software frameworks. Leveraging on the reported experience and work from the ATLAS and CMS experience on using the ESPER engine, we discuss and show how such approach has been implemented in STAR for meta-data event triggering stream processing and filtering. An ESPER based solution seems to fit well into the online data acquisition system where many systems are monitored.

  2. An intelligent tool for activity data collection.

    PubMed

    Sarkar, A M Jehad

    2011-01-01

    Activity recognition systems using simple and ubiquitous sensors require a large variety of real-world sensor data for not only evaluating their performance but also training the systems for better functioning. However, a tremendous amount of effort is required to setup an environment for collecting such data. For example, expertise and resources are needed to design and install the sensors, controllers, network components, and middleware just to perform basic data collections. It is therefore desirable to have a data collection method that is inexpensive, flexible, user-friendly, and capable of providing large and diverse activity datasets. In this paper, we propose an intelligent activity data collection tool which has the ability to provide such datasets inexpensively without physically deploying the testbeds. It can be used as an inexpensive and alternative technique to collect human activity data. The tool provides a set of web interfaces to create a web-based activity data collection environment. It also provides a web-based experience sampling tool to take the user's activity input. The tool generates an activity log using its activity knowledge and the user-given inputs. The activity knowledge is mined from the web. We have performed two experiments to validate the tool's performance in producing reliable datasets.

  3. Historic Mining and Agriculture as Indicators of Occurrence and Abundance of Widespread Invasive Plant Species

    PubMed Central

    Calinger, Kellen; Calhoon, Elisabeth; Chang, Hsiao-chi; Whitacre, James; Wenzel, John; Comita, Liza; Queenborough, Simon

    2015-01-01

    Anthropogenic disturbances often change ecological communities and provide opportunities for non-native species invasion. Understanding the impacts of disturbances on species invasion is therefore crucial for invasive species management. We used generalized linear mixed effects models to explore the influence of land-use history and distance to roads on the occurrence and abundance of two invasive plant species (Rosa multiflora and Berberis thunbergii) in a 900-ha deciduous forest in the eastern U.S.A., the Powdermill Nature Reserve. Although much of the reserve has been continuously forested since at least 1939, aerial photos revealed a variety of land-uses since then including agriculture, mining, logging, and development. By 2008, both R. multiflora and B. thunbergii were widespread throughout the reserve (occurring in 24% and 13% of 4417 10-m diameter regularly-placed vegetation plots, respectively) with occurrence and abundance of each varying significantly with land-use history. Rosa multiflora was more likely to occur in historically farmed, mined, logged or developed plots than in plots that remained forested, (log odds of 1.8 to 3.0); Berberis thunbergii was more likely to occur in plots with agricultural, mining, or logging history than in plots without disturbance (log odds of 1.4 to 2.1). Mining, logging, and agriculture increased the probability that R. multiflora had >10% cover while only past agriculture was related to cover of B. thunbergii. Proximity to roads was positively correlated with the occurrence of R. multiflora (a 0.26 increase in the log odds for every 1-m closer) but not B. thunbergii, and roads had no impact on the abundance of either species. Our results indicated that a wide variety of disturbances may aid the introduction of invasive species into new habitats, while high-impact disturbances such as agriculture and mining increase the likelihood of high abundance post-introduction. PMID:26046534

  4. MineScan: non-image data monitoring and mining from imaging modalities

    NASA Astrophysics Data System (ADS)

    Zaidi, Shayan M.; Huff, Dov; Bhalodia, Pankit; Mongkolwat, Pattanasak; Channin, David S.

    2003-05-01

    This project is intended to capture and interactively display non-image information routinely generated by imaging modalities. This information relates to the device's performance of the individual procedures and is not necessarily available in other information streams such as DICOM headers. While originally intended for use in servicing the modalities, this information can also be presented to radiologists and administrators within the department for both micro- and macro-management purposes. This data can help hospital administrators and radiologists manage available resources and discover clues to indicate what modifications in hospital operations might significantly improve its ability to provide efficient patient care. Data is collected from a departmental CT scanner. The data consists of a running record of exams followed by a list of processing records logged over a 24-hour period. MineScan extracts information from these records and stores it into a database. A statistical program is run once a day to collect relevant metrics. MineScan can be accessed via a Web browser or through an advanced prototype PACS workstation. This information, if provided in real-time, can be used to manage operations in a busy department. Even when provided historically, the data can be used to assess current activity, analyze trends and plan future operations.

  5. Industrial application of semantic process mining

    NASA Astrophysics Data System (ADS)

    Espen Ingvaldsen, Jon; Atle Gulla, Jon

    2012-05-01

    Process mining relates to the extraction of non-trivial and useful information from information system event logs. It is a new research discipline that has evolved significantly since the early work on idealistic process logs. Over the last years, process mining prototypes have incorporated elements from semantics and data mining and targeted visualisation techniques that are more user-friendly to business experts and process owners. In this article, we present a framework for evaluating different aspects of enterprise process flows and address practical challenges of state-of-the-art industrial process mining. We also explore the inherent strengths of the technology for more efficient process optimisation.

  6. The application of PGNAA borehole logging for copper grade estimation at Chuquicamata mine.

    PubMed

    Charbucinski, J; Duran, O; Freraut, R; Heresi, N; Pineyro, I

    2004-05-01

    The field trials of a prompt gamma neutron activation (PGNAA) spectrometric logging method and instrumentation (SIROLOG) for copper grade estimation in production holes of a porphyry type copper ore mine, Chuquicamata in Chile, are described. Examples of data analysis, calibration procedures and copper grade profiles are provided. The field tests have proved the suitability of the PGNAA logging system for in situ quality control of copper ore.

  7. Inference of strata separation and gas emission paths in longwall overburden using continuous wavelet transform of well logs and geostatistical simulation

    NASA Astrophysics Data System (ADS)

    Karacan, C. Özgen; Olea, Ricardo A.

    2014-06-01

    Prediction of potential methane emission pathways from various sources into active mine workings or sealed gobs from longwall overburden is important for controlling methane and for improving mining safety. The aim of this paper is to infer strata separation intervals and thus gas emission pathways from standard well log data. The proposed technique was applied to well logs acquired through the Mary Lee/Blue Creek coal seam of the Upper Pottsville Formation in the Black Warrior Basin, Alabama, using well logs from a series of boreholes aligned along a nearly linear profile. For this purpose, continuous wavelet transform (CWT) of digitized gamma well logs was performed by using Mexican hat and Morlet, as the mother wavelets, to identify potential discontinuities in the signal. Pointwise Hölder exponents (PHE) of gamma logs were also computed using the generalized quadratic variations (GQV) method to identify the location and strength of singularities of well log signals as a complementary analysis. PHEs and wavelet coefficients were analyzed to find the locations of singularities along the logs. Using the well logs in this study, locations of predicted singularities were used as indicators in single normal equation simulation (SNESIM) to generate equi-probable realizations of potential strata separation intervals. Horizontal and vertical variograms of realizations were then analyzed and compared with those of indicator data and training image (TI) data using the Kruskal-Wallis test. A sum of squared differences was employed to select the most probable realization representing the locations of potential strata separations and methane flow paths. Results indicated that singularities located in well log signals reliably correlated with strata transitions or discontinuities within the strata. Geostatistical simulation of these discontinuities provided information about the location and extents of the continuous channels that may form during mining. If there is a gas source within their zone of influence, paths may develop and allow methane movement towards sealed or active gobs under pressure differentials. Knowledge gained from this research will better prepare mine operations for potential methane inflows, thus improving mine safety.

  8. Pattern mining of user interaction logs for a post-deployment usability evaluation of a radiology PACS client.

    PubMed

    Jorritsma, Wiard; Cnossen, Fokie; Dierckx, Rudi A; Oudkerk, Matthijs; van Ooijen, Peter M A

    2016-01-01

    To perform a post-deployment usability evaluation of a radiology Picture Archiving and Communication System (PACS) client based on pattern mining of user interaction log data, and to assess the usefulness of this approach compared to a field study. All user actions performed on the PACS client were logged for four months. A data mining technique called closed sequential pattern mining was used to automatically extract frequently occurring interaction patterns from the log data. These patterns were used to identify usability issues with the PACS. The results of this evaluation were compared to the results of a field study based usability evaluation of the same PACS client. The interaction patterns revealed four usability issues: (1) the display protocols do not function properly, (2) the line measurement tool stays active until another tool is selected, rather than being deactivated after one use, (3) the PACS's built-in 3D functionality does not allow users to effectively perform certain 3D-related tasks, (4) users underuse the PACS's customization possibilities. All usability issues identified based on the log data were also found in the field study, which identified 48 issues in total. Post-deployment usability evaluation based on pattern mining of user interaction log data provides useful insights into the way users interact with the radiology PACS client. However, it reveals few usability issues compared to a field study and should therefore not be used as the sole method of usability evaluation. Copyright © 2015 Elsevier Ireland Ltd. All rights reserved.

  9. Web-video-mining-supported workflow modeling for laparoscopic surgeries.

    PubMed

    Liu, Rui; Zhang, Xiaoli; Zhang, Hao

    2016-11-01

    As quality assurance is of strong concern in advanced surgeries, intelligent surgical systems are expected to have knowledge such as the knowledge of the surgical workflow model (SWM) to support their intuitive cooperation with surgeons. For generating a robust and reliable SWM, a large amount of training data is required. However, training data collected by physically recording surgery operations is often limited and data collection is time-consuming and labor-intensive, severely influencing knowledge scalability of the surgical systems. The objective of this research is to solve the knowledge scalability problem in surgical workflow modeling with a low cost and labor efficient way. A novel web-video-mining-supported surgical workflow modeling (webSWM) method is developed. A novel video quality analysis method based on topic analysis and sentiment analysis techniques is developed to select high-quality videos from abundant and noisy web videos. A statistical learning method is then used to build the workflow model based on the selected videos. To test the effectiveness of the webSWM method, 250 web videos were mined to generate a surgical workflow for the robotic cholecystectomy surgery. The generated workflow was evaluated by 4 web-retrieved videos and 4 operation-room-recorded videos, respectively. The evaluation results (video selection consistency n-index ≥0.60; surgical workflow matching degree ≥0.84) proved the effectiveness of the webSWM method in generating robust and reliable SWM knowledge by mining web videos. With the webSWM method, abundant web videos were selected and a reliable SWM was modeled in a short time with low labor cost. Satisfied performances in mining web videos and learning surgery-related knowledge show that the webSWM method is promising in scaling knowledge for intelligent surgical systems. Copyright © 2016 Elsevier B.V. All rights reserved.

  10. Transaction aware tape-infrastructure monitoring

    NASA Astrophysics Data System (ADS)

    Nikolaidis, Fotios; Kruse, Daniele Francesco

    2014-06-01

    Administrating a large scale, multi protocol, hierarchical tape infrastructure like the CERN Advanced STORage manager (CASTOR)[2], which stores now 100 PB (with an increasing step of 25 PB per year), requires an adequate monitoring system for quick spotting of malfunctions, easier debugging and on demand report generation. The main challenges for such system are: to cope with CASTOR's log format diversity and its information scattered among several log files, the need for long term information archival, the strict reliability requirements and the group based GUI visualization. For this purpose, we have designed, developed and deployed a centralized system consisting of four independent layers: the Log Transfer layer for collecting log lines from all tape servers to a single aggregation server, the Data Mining layer for combining log data into transaction context, the Storage layer for archiving the resulting transactions and finally the Web UI layer for accessing the information. Having flexibility, extensibility and maintainability in mind, each layer is designed to work as a message broker for the next layer, providing a clean and generic interface while ensuring consistency, redundancy and ultimately fault tolerance. This system unifies information previously dispersed over several monitoring tools into a single user interface, using Splunk, which also allows us to provide information visualization based on access control lists (ACL). Since its deployment, it has been successfully used by CASTOR tape operators for quick overview of transactions, performance evaluation, malfunction detection and from managers for report generation.

  11. Geophysical logs for selected wells in the Picher Field, northeast Oklahoma and southeast Kansas

    USGS Publications Warehouse

    Christenson, Scott C.; Thomas, Tom B.; Overton, Myles D.; Goemaat, Robert L.; Havens, John S.

    1991-01-01

    The Roubidoux aquifer in northeastern Oklahoma is used extensively as a source of water for public supplies, commerce, industry, and rural water districts. The Roubidoux aquifer may be subject to contamination from abandoned lead and zinc mines of the Picher field. Water in flooded underground mines contains large concentrations of iron, zinc, cadmium, and lead. The contaminated water may migrate from the mines to the Roubidoux aquifer through abandoned water wells in the Picher field. In late 1984, the Oklahoma Water Resources Board began to locate abandoned wells that might be serving as conduits for the migration of contaminants from the abandoned mines. These wells were cleared of debris and plugged. A total of 66 wells had been located, cleared, and plugged by July 1985. In cooperation with the Oklahoma Water Resources Board, the U.S. Geological Survey took advantage of the opportunity to obtain geophysical data in the study area and provide the Oklahoma Water Resources Board with data that might be useful during the well-plugging operation. Geophysical logs obtained by the U.S. Geological Survey are presented in this report. The geophysical logs include hole diameter, normal, single-point resistance, fluid resistivity, natural-gamma, gamma-gamma, and neutron logs. Depths logged range from 145 to 1,344 feet.

  12. OntoGene web services for biomedical text mining.

    PubMed

    Rinaldi, Fabio; Clematide, Simon; Marques, Hernani; Ellendorff, Tilia; Romacker, Martin; Rodriguez-Esteban, Raul

    2014-01-01

    Text mining services are rapidly becoming a crucial component of various knowledge management pipelines, for example in the process of database curation, or for exploration and enrichment of biomedical data within the pharmaceutical industry. Traditional architectures, based on monolithic applications, do not offer sufficient flexibility for a wide range of use case scenarios, and therefore open architectures, as provided by web services, are attracting increased interest. We present an approach towards providing advanced text mining capabilities through web services, using a recently proposed standard for textual data interchange (BioC). The web services leverage a state-of-the-art platform for text mining (OntoGene) which has been tested in several community-organized evaluation challenges,with top ranked results in several of them.

  13. OntoGene web services for biomedical text mining

    PubMed Central

    2014-01-01

    Text mining services are rapidly becoming a crucial component of various knowledge management pipelines, for example in the process of database curation, or for exploration and enrichment of biomedical data within the pharmaceutical industry. Traditional architectures, based on monolithic applications, do not offer sufficient flexibility for a wide range of use case scenarios, and therefore open architectures, as provided by web services, are attracting increased interest. We present an approach towards providing advanced text mining capabilities through web services, using a recently proposed standard for textual data interchange (BioC). The web services leverage a state-of-the-art platform for text mining (OntoGene) which has been tested in several community-organized evaluation challenges, with top ranked results in several of them. PMID:25472638

  14. Analyzing Web Server Logs to Improve a Site's Usage. The Systems Librarian

    ERIC Educational Resources Information Center

    Breeding, Marshall

    2005-01-01

    This column describes ways to streamline and optimize how a Web site works in order to improve both its usability and its visibility. The author explains how to analyze logs and other system data to measure the effectiveness of the Web site design and search engine.

  15. An Intelligent Tool for Activity Data Collection

    PubMed Central

    Jehad Sarkar, A. M.

    2011-01-01

    Activity recognition systems using simple and ubiquitous sensors require a large variety of real-world sensor data for not only evaluating their performance but also training the systems for better functioning. However, a tremendous amount of effort is required to setup an environment for collecting such data. For example, expertise and resources are needed to design and install the sensors, controllers, network components, and middleware just to perform basic data collections. It is therefore desirable to have a data collection method that is inexpensive, flexible, user-friendly, and capable of providing large and diverse activity datasets. In this paper, we propose an intelligent activity data collection tool which has the ability to provide such datasets inexpensively without physically deploying the testbeds. It can be used as an inexpensive and alternative technique to collect human activity data. The tool provides a set of web interfaces to create a web-based activity data collection environment. It also provides a web-based experience sampling tool to take the user’s activity input. The tool generates an activity log using its activity knowledge and the user-given inputs. The activity knowledge is mined from the web. We have performed two experiments to validate the tool’s performance in producing reliable datasets. PMID:22163832

  16. Automatic generation of Web mining environments

    NASA Astrophysics Data System (ADS)

    Cibelli, Maurizio; Costagliola, Gennaro

    1999-02-01

    The main problem related to the retrieval of information from the world wide web is the enormous number of unstructured documents and resources, i.e., the difficulty of locating and tracking appropriate sources. This paper presents a web mining environment (WME), which is capable of finding, extracting and structuring information related to a particular domain from web documents, using general purpose indices. The WME architecture includes a web engine filter (WEF), to sort and reduce the answer set returned by a web engine, a data source pre-processor (DSP), which processes html layout cues in order to collect and qualify page segments, and a heuristic-based information extraction system (HIES), to finally retrieve the required data. Furthermore, we present a web mining environment generator, WMEG, that allows naive users to generate a WME specific to a given domain by providing a set of specifications.

  17. Mining a Web Citation Database for Author Co-Citation Analysis.

    ERIC Educational Resources Information Center

    He, Yulan; Hui, Siu Cheung

    2002-01-01

    Proposes a mining process to automate author co-citation analysis based on the Web Citation Database, a data warehouse for storing citation indices of Web publications. Describes the use of agglomerative hierarchical clustering for author clustering and multidimensional scaling for displaying author cluster maps, and explains PubSearch, a…

  18. From Log Files to Assessment Metrics: Measuring Students' Science Inquiry Skills Using Educational Data Mining

    ERIC Educational Resources Information Center

    Gobert, Janice D.; Sao Pedro, Michael; Raziuddin, Juelaila; Baker, Ryan S.

    2013-01-01

    We present a method for assessing science inquiry performance, specifically for the inquiry skill of designing and conducting experiments, using educational data mining on students' log data from online microworlds in the Inq-ITS system (Inquiry Intelligent Tutoring System; www.inq-its.org). In our approach, we use a 2-step process: First we use…

  19. Users' Perceptions of the Web As Revealed by Transaction Log Analysis.

    ERIC Educational Resources Information Center

    Moukdad, Haidar; Large, Andrew

    2001-01-01

    Describes the results of a transaction log analysis of a Web search engine, WebCrawler, to analyze user's queries for information retrieval. Results suggest most users do not employ advanced search features, and the linguistic structure often resembles a human-human communication model that is not always successful in human-computer communication.…

  20. Discovering Decision Knowledge from Web Log Portfolio for Managing Classroom Processes by Applying Decision Tree and Data Cube Technology.

    ERIC Educational Resources Information Center

    Chen, Gwo-Dong; Liu, Chen-Chung; Ou, Kuo-Liang; Liu, Baw-Jhiune

    2000-01-01

    Discusses the use of Web logs to record student behavior that can assist teachers in assessing performance and making curriculum decisions for distance learning students who are using Web-based learning systems. Adopts decision tree and data cube information processing methodologies for developing more effective pedagogical strategies. (LRW)

  1. Data Mining for Web-Based Support Systems: A Case Study in e-Custom Systems

    NASA Astrophysics Data System (ADS)

    Razmerita, Liana; Kirchner, Kathrin

    This chapter provides an example of a Web-based support system (WSS) used to streamline trade procedures, prevent potential security threats, and reduce tax-related fraud in cross-border trade. The architecture is based on a service-oriented architecture that includes smart seals and Web services. We discuss the implications and suggest further enhancements to demonstrate how such systems can move toward a Web-based decision support system with the support of data mining methods. We provide a concrete example of how data mining can help to analyze the vast amount of data collected while monitoring the container movements along its supply chain.

  2. Semantic web for integrated network analysis in biomedicine.

    PubMed

    Chen, Huajun; Ding, Li; Wu, Zhaohui; Yu, Tong; Dhanapalan, Lavanya; Chen, Jake Y

    2009-03-01

    The Semantic Web technology enables integration of heterogeneous data on the World Wide Web by making the semantics of data explicit through formal ontologies. In this article, we survey the feasibility and state of the art of utilizing the Semantic Web technology to represent, integrate and analyze the knowledge in various biomedical networks. We introduce a new conceptual framework, semantic graph mining, to enable researchers to integrate graph mining with ontology reasoning in network data analysis. Through four case studies, we demonstrate how semantic graph mining can be applied to the analysis of disease-causal genes, Gene Ontology category cross-talks, drug efficacy analysis and herb-drug interactions analysis.

  3. From IHE Audit Trails to XES Event Logs Facilitating Process Mining.

    PubMed

    Paster, Ferdinand; Helm, Emmanuel

    2015-01-01

    Recently Business Intelligence approaches like process mining are applied to the healthcare domain. The goal of process mining is to gain process knowledge, compliance and room for improvement by investigating recorded event data. Previous approaches focused on process discovery by event data from various specific systems. IHE, as a globally recognized basis for healthcare information systems, defines in its ATNA profile how real-world events must be recorded in centralized event logs. The following approach presents how audit trails collected by the means of ATNA can be transformed to enable process mining. Using the standardized audit trails provides the ability to apply these methods to all IHE based information systems.

  4. “Blogging” About Course Concepts: Using Technology for Reflective Journaling in a Communications Class

    PubMed Central

    Bouldin, Alicia S.; Holmes, Erin R.; Fortenberry, Michael L.

    2006-01-01

    Objective Web log technology was applied to a reflective journaling exercise in a communication course during the second-professional year at the University of Mississippi School of Pharmacy, to encourage students to reflect on course concepts and apply them to the environment outside the classroom, and to assess their communication performance. Design Two Web log entries per week were required for full credit. Web logs were evaluated at three points during the term. At the end of the course, students evaluated the assignment using a 2-page survey instrument. Assessment The assignment contributed to student learning and increased awareness level for approximately 40% of the class. Students had few complaints about the logistics of the assignment. Conclusion The Web log technology was a useful tool for reflective journaling in this communications course. Future versions of the assignment will benefit from student feedback from this initial experience. PMID:17136203

  5. Exploring Online Students' Self-Regulated Learning with Self-Reported Surveys and Log Files: A Data Mining Approach

    ERIC Educational Resources Information Center

    Cho, Moon-Heum; Yoo, Jin Soung

    2017-01-01

    Many researchers who are interested in studying students' online self-regulated learning (SRL) have heavily relied on self-reported surveys. Data mining is an alternative technique that can be used to discover students' SRL patterns from large data logs saved on a course management system. The purpose of this study was to identify students' online…

  6. Web data mining

    NASA Astrophysics Data System (ADS)

    Wibonele, Kasanda J.; Zhang, Yanqing

    2002-03-01

    A web data mining system using granular computing and ASP programming is proposed. This is a web based application, which allows web users to submit survey data for many different companies. This survey is a collection of questions that will help these companies develop and improve their business and customer service with their clients by analyzing survey data. This web application allows users to submit data anywhere. All the survey data is collected into a database for further analysis. An administrator of this web application can login to the system and view all the data submitted. This web application resides on a web server, and the database resides on the MS SQL server.

  7. Using Web Logs in the Science Classroom

    ERIC Educational Resources Information Center

    Duplichan, Staycle C.

    2009-01-01

    As educators we must ask ourselves if we are meeting the needs of today's students. The science world is adapting to our ever-changing society; are the methodology and philosophy of our educational system keeping up? In this article, you'll learn why web logs (also called blogs) are an important Web 2.0 tool in your science classroom and how they…

  8. Statistics, Structures & Satisfied Customers: Using Web Log Data to Improve Site Performance.

    ERIC Educational Resources Information Center

    Peacock, Darren

    This paper explores some of the ways in which the National Museum of Australia is using Web analysis tools to shape its future directions in the delivery of online services. In particular, it explores the potential of quantitative analysis, based on Web server log data, to convert these ephemeral traces of user experience into a strategic…

  9. The Effectiveness of Web-Based Learning Environment: A Case Study of Public Universities in Kenya

    ERIC Educational Resources Information Center

    Kirui, Paul A.; Mutai, Sheila J.

    2010-01-01

    Web mining is emerging in many aspects of e-learning, aiming at improving online learning and teaching processes and making them more transparent and effective. Researchers using Web mining tools and techniques are challenged to learn more about the online students' reshaping online courses and educational websites, and create tools for…

  10. Why, What, and How to Log? Lessons from LISTEN

    ERIC Educational Resources Information Center

    Mostow, Jack; Beck, Joseph E.

    2009-01-01

    The ability to log tutorial interactions in comprehensive, longitudinal, fine-grained detail offers great potential for educational data mining--but what data is logged, and how, can facilitate or impede the realization of that potential. We propose guidelines gleaned over 15 years of logging, exploring, and analyzing millions of events from…

  11. Earth Science Mining Web Services

    NASA Astrophysics Data System (ADS)

    Pham, L. B.; Lynnes, C. S.; Hegde, M.; Graves, S.; Ramachandran, R.; Maskey, M.; Keiser, K.

    2008-12-01

    To allow scientists further capabilities in the area of data mining and web services, the Goddard Earth Sciences Data and Information Services Center (GES DISC) and researchers at the University of Alabama in Huntsville (UAH) have developed a system to mine data at the source without the need of network transfers. The system has been constructed by linking together several pre-existing technologies: the Simple Scalable Script-based Science Processor for Measurements (S4PM), a processing engine at the GES DISC; the Algorithm Development and Mining (ADaM) system, a data mining toolkit from UAH that can be configured in a variety of ways to create customized mining processes; ActiveBPEL, a workflow execution engine based on BPEL (Business Process Execution Language); XBaya, a graphical workflow composer; and the EOS Clearinghouse (ECHO). XBaya is used to construct an analysis workflow at UAH using ADaM components, which are also installed remotely at the GES DISC, wrapped as Web Services. The S4PM processing engine searches ECHO for data using space-time criteria, staging them to cache, allowing the ActiveBPEL engine to remotely orchestrates the processing workflow within S4PM. As mining is completed, the output is placed in an FTP holding area for the end user. The goals are to give users control over the data they want to process, while mining data at the data source using the server's resources rather than transferring the full volume over the internet. These diverse technologies have been infused into a functioning, distributed system with only minor changes to the underlying technologies. The key to this infusion is the loosely coupled, Web- Services based architecture: All of the participating components are accessible (one way or another) through (Simple Object Access Protocol) SOAP-based Web Services.

  12. Earth Science Mining Web Services

    NASA Technical Reports Server (NTRS)

    Pham, Long; Lynnes, Christopher; Hegde, Mahabaleshwa; Graves, Sara; Ramachandran, Rahul; Maskey, Manil; Keiser, Ken

    2008-01-01

    To allow scientists further capabilities in the area of data mining and web services, the Goddard Earth Sciences Data and Information Services Center (GES DISC) and researchers at the University of Alabama in Huntsville (UAH) have developed a system to mine data at the source without the need of network transfers. The system has been constructed by linking together several pre-existing technologies: the Simple Scalable Script-based Science Processor for Measurements (S4PM), a processing engine at he GES DISC; the Algorithm Development and Mining (ADaM) system, a data mining toolkit from UAH that can be configured in a variety of ways to create customized mining processes; ActiveBPEL, a workflow execution engine based on BPEL (Business Process Execution Language); XBaya, a graphical workflow composer; and the EOS Clearinghouse (ECHO). XBaya is used to construct an analysis workflow at UAH using ADam components, which are also installed remotely at the GES DISC, wrapped as Web Services. The S4PM processing engine searches ECHO for data using space-time criteria, staging them to cache, allowing the ActiveBPEL engine to remotely orchestras the processing workflow within S4PM. As mining is completed, the output is placed in an FTP holding area for the end user. The goals are to give users control over the data they want to process, while mining data at the data source using the server's resources rather than transferring the full volume over the internet. These diverse technologies have been infused into a functioning, distributed system with only minor changes to the underlying technologies. The key to the infusion is the loosely coupled, Web-Services based architecture: All of the participating components are accessible (one way or another) through (Simple Object Access Protocol) SOAP-based Web Services.

  13. Lightweight monitoring and control system for coal mine safety using REST style.

    PubMed

    Cheng, Bo; Cheng, Xin; Chen, Junliang

    2015-01-01

    The complex environment of a coal mine requires the underground environment, devices and miners to be constantly monitored to ensure safe coal production. However, existing coal mines do not meet these coverage requirements because blind spots occur when using a wired network. In this paper, we develop a Web-based, lightweight remote monitoring and control platform using a wireless sensor network (WSN) with the REST style to collect temperature, humidity and methane concentration data in a coal mine using sensor nodes. This platform also collects information on personnel positions inside the mine. We implement a RESTful application programming interface (API) that provides access to underground sensors and instruments through the Web such that underground coal mine physical devices can be easily interfaced to remote monitoring and control applications. We also implement three different scenarios for Web-based, lightweight remote monitoring and control of coal mine safety and measure and analyze the system performance. Finally, we present the conclusions from this study and discuss future work. Copyright © 2014 ISA. Published by Elsevier Ltd. All rights reserved.

  14. Service-based analysis of biological pathways

    PubMed Central

    Zheng, George; Bouguettaya, Athman

    2009-01-01

    Background Computer-based pathway discovery is concerned with two important objectives: pathway identification and analysis. Conventional mining and modeling approaches aimed at pathway discovery are often effective at achieving either objective, but not both. Such limitations can be effectively tackled leveraging a Web service-based modeling and mining approach. Results Inspired by molecular recognitions and drug discovery processes, we developed a Web service mining tool, named PathExplorer, to discover potentially interesting biological pathways linking service models of biological processes. The tool uses an innovative approach to identify useful pathways based on graph-based hints and service-based simulation verifying user's hypotheses. Conclusion Web service modeling of biological processes allows the easy access and invocation of these processes on the Web. Web service mining techniques described in this paper enable the discovery of biological pathways linking these process service models. Algorithms presented in this paper for automatically highlighting interesting subgraph within an identified pathway network enable the user to formulate hypothesis, which can be tested out using our simulation algorithm that are also described in this paper. PMID:19796403

  15. Exploiting Recurring Structure in a Semantic Network

    NASA Technical Reports Server (NTRS)

    Wolfe, Shawn R.; Keller, Richard M.

    2004-01-01

    With the growing popularity of the Semantic Web, an increasing amount of information is becoming available in machine interpretable, semantically structured networks. Within these semantic networks are recurring structures that could be mined by existing or novel knowledge discovery methods. The mining of these semantic structures represents an interesting area that focuses on mining both for and from the Semantic Web, with surprising applicability to problems confronting the developers of Semantic Web applications. In this paper, we present representative examples of recurring structures and show how these structures could be used to increase the utility of a semantic repository deployed at NASA.

  16. Analyzing engagement in a web-based intervention platform through visualizing log-data.

    PubMed

    Morrison, Cecily; Doherty, Gavin

    2014-11-13

    Engagement has emerged as a significant cross-cutting concern within the development of Web-based interventions. There have been calls to institute a more rigorous approach to the design of Web-based interventions, to increase both the quantity and quality of engagement. One approach would be to use log-data to better understand the process of engagement and patterns of use. However, an important challenge lies in organizing log-data for productive analysis. Our aim was to conduct an initial exploration of the use of visualizations of log-data to enhance understanding of engagement with Web-based interventions. We applied exploratory sequential data analysis to highlight sequential aspects of the log data, such as time or module number, to provide insights into engagement. After applying a number of processing steps, a range of visualizations were generated from the log-data. We then examined the usefulness of these visualizations for understanding the engagement of individual users and the engagement of cohorts of users. The visualizations created are illustrated with two datasets drawn from studies using the SilverCloud Platform: (1) a small, detailed dataset with interviews (n=19) and (2) a large dataset (n=326) with 44,838 logged events. We present four exploratory visualizations of user engagement with a Web-based intervention, including Navigation Graph, Stripe Graph, Start-Finish Graph, and Next Action Heat Map. The first represents individual usage and the last three, specific aspects of cohort usage. We provide examples of each with a discussion of salient features. Log-data analysis through data visualization is an alternative way of exploring user engagement with Web-based interventions, which can yield different insights than more commonly used summative measures. We describe how understanding the process of engagement through visualizations can support the development and evaluation of Web-based interventions. Specifically, we show how visualizations can (1) allow inspection of content or feature usage in a temporal relationship to the overall program at different levels of granularity, (2) detect different patterns of use to consider personalization in the design process, (3) detect usability issues, (4) enable exploratory analysis to support the design of statistical queries to summarize the data, (5) provide new opportunities for real-time evaluation, and (6) examine assumptions about interactivity that underlie many summative measures in this field.

  17. Analyzing Engagement in a Web-Based Intervention Platform Through Visualizing Log-Data

    PubMed Central

    2014-01-01

    Background Engagement has emerged as a significant cross-cutting concern within the development of Web-based interventions. There have been calls to institute a more rigorous approach to the design of Web-based interventions, to increase both the quantity and quality of engagement. One approach would be to use log-data to better understand the process of engagement and patterns of use. However, an important challenge lies in organizing log-data for productive analysis. Objective Our aim was to conduct an initial exploration of the use of visualizations of log-data to enhance understanding of engagement with Web-based interventions. Methods We applied exploratory sequential data analysis to highlight sequential aspects of the log data, such as time or module number, to provide insights into engagement. After applying a number of processing steps, a range of visualizations were generated from the log-data. We then examined the usefulness of these visualizations for understanding the engagement of individual users and the engagement of cohorts of users. The visualizations created are illustrated with two datasets drawn from studies using the SilverCloud Platform: (1) a small, detailed dataset with interviews (n=19) and (2) a large dataset (n=326) with 44,838 logged events. Results We present four exploratory visualizations of user engagement with a Web-based intervention, including Navigation Graph, Stripe Graph, Start–Finish Graph, and Next Action Heat Map. The first represents individual usage and the last three, specific aspects of cohort usage. We provide examples of each with a discussion of salient features. Conclusions Log-data analysis through data visualization is an alternative way of exploring user engagement with Web-based interventions, which can yield different insights than more commonly used summative measures. We describe how understanding the process of engagement through visualizations can support the development and evaluation of Web-based interventions. Specifically, we show how visualizations can (1) allow inspection of content or feature usage in a temporal relationship to the overall program at different levels of granularity, (2) detect different patterns of use to consider personalization in the design process, (3) detect usability issues, (4) enable exploratory analysis to support the design of statistical queries to summarize the data, (5) provide new opportunities for real-time evaluation, and (6) examine assumptions about interactivity that underlie many summative measures in this field. PMID:25406097

  18. Data Mining Web Services for Science Data Repositories

    NASA Astrophysics Data System (ADS)

    Graves, S.; Ramachandran, R.; Keiser, K.; Maskey, M.; Lynnes, C.; Pham, L.

    2006-12-01

    The maturation of web services standards and technologies sets the stage for a distributed "Service-Oriented Architecture" (SOA) for NASA's next generation science data processing. This architecture will allow members of the scientific community to create and combine persistent distributed data processing services and make them available to other users over the Internet. NASA has initiated a project to create a suite of specialized data mining web services designed specifically for science data. The project leverages the Algorithm Development and Mining (ADaM) toolkit as its basis. The ADaM toolkit is a robust, mature and freely available science data mining toolkit that is being used by several research organizations and educational institutions worldwide. These mining services will give the scientific community a powerful and versatile data mining capability that can be used to create higher order products such as thematic maps from current and future NASA satellite data records with methods that are not currently available. The package of mining and related services are being developed using Web Services standards so that community-based measurement processing systems can access and interoperate with them. These standards-based services allow users different options for utilizing them, from direct remote invocation by a client application to deployment of a Business Process Execution Language (BPEL) solutions package where a complex data mining workflow is exposed to others as a single service. The ability to deploy and operate these services at a data archive allows the data mining algorithms to be run where the data are stored, a more efficient scenario than moving large amounts of data over the network. This will be demonstrated in a scenario in which a user uses a remote Web-Service-enabled clustering algorithm to create cloud masks from satellite imagery at the Goddard Earth Sciences Data and Information Services Center (GES DISC).

  19. Content-Aware DataGuide with Incremental Index Update using Frequently Used Paths

    NASA Astrophysics Data System (ADS)

    Sharma, A. K.; Duhan, Neelam; Khattar, Priyanka

    2010-11-01

    Size of the WWW is increasing day by day. Due to the absence of structured data on the Web, it becomes very difficult for information retrieval tools to fully utilize the Web information. As a solution to this problem, XML pages come into play, which provide structural information to the users to some extent. Without efficient indexes, query processing can be quite inefficient due to an exhaustive traversal on XML data. In this paper an improved content-centric approach of Content-Aware DataGuide, which is an indexing technique for XML databases, is being proposed that uses frequently used paths from historical query logs to improve query performance. The index can be updated incrementally according to the changes in query workload and thus, the overhead of reconstruction can be minimized. Frequently used paths are extracted using any Sequential Pattern mining algorithm on subsequent queries in the query workload. After this, the data structures are incrementally updated. This indexing technique proves to be efficient as partial matching queries can be executed efficiently and users can now get the more relevant documents in results.

  20. Comparing Web and Touch Screen Transaction Log Files

    PubMed Central

    Huntington, Paul; Williams, Peter

    2001-01-01

    Background Digital health information is available on a wide variety of platforms including PC-access of the Internet, Wireless Application Protocol phones, CD-ROMs, and touch screen public kiosks. All these platforms record details of user sessions in transaction log files, and there is a growing body of research into the evaluation of this data. However, there is very little research that has examined the problems of comparing the transaction log files of kiosks and the Internet. Objectives To provide a first step towards examining the problems of comparing the transaction log files of kiosks and the Internet. Methods We studied two platforms: touch screen kiosks and a comparable Web site. For both of these platforms, we examined the menu structure (which affects transaction log file data), the log-file structure, and the metrics derived from log-file records. Results We found substantial differences between the generated metrics. Conclusions None of the metrics discussed can be regarded as an effective way of comparing the use of kiosks and Web sites. Two metrics stand out as potentially comparable and valuable: the number of user sessions per hour and user penetration of pages. PMID:11720960

  1. A construction scheme of web page comment information extraction system based on frequent subtree mining

    NASA Astrophysics Data System (ADS)

    Zhang, Xiaowen; Chen, Bingfeng

    2017-08-01

    Based on the frequent sub-tree mining algorithm, this paper proposes a construction scheme of web page comment information extraction system based on frequent subtree mining, referred to as FSM system. The entire system architecture and the various modules to do a brief introduction, and then the core of the system to do a detailed description, and finally give the system prototype.

  2. Kernel Methods for Mining Instance Data in Ontologies

    NASA Astrophysics Data System (ADS)

    Bloehdorn, Stephan; Sure, York

    The amount of ontologies and meta data available on the Web is constantly growing. The successful application of machine learning techniques for learning of ontologies from textual data, i.e. mining for the Semantic Web, contributes to this trend. However, no principal approaches exist so far for mining from the Semantic Web. We investigate how machine learning algorithms can be made amenable for directly taking advantage of the rich knowledge expressed in ontologies and associated instance data. Kernel methods have been successfully employed in various learning tasks and provide a clean framework for interfacing between non-vectorial data and machine learning algorithms. In this spirit, we express the problem of mining instances in ontologies as the problem of defining valid corresponding kernels. We present a principled framework for designing such kernels by means of decomposing the kernel computation into specialized kernels for selected characteristics of an ontology which can be flexibly assembled and tuned. Initial experiments on real world Semantic Web data enjoy promising results and show the usefulness of our approach.

  3. Abandoned Uranium Mines (AUM) Site Screening Map Service, 2016, US EPA Region 9

    EPA Pesticide Factsheets

    As described in detail in the Five-Year Report, US EPA completed on-the-ground screening of 521 abandoned uranium mine areas. US EPA and the Navajo EPA are using the Comprehensive Database and Atlas to determine which mines should be cleaned up first. US EPA continues to research and identify Potentially Responsible Parties (PRPs) under Superfund to contribute to the costs of cleanup efforts.This US EPA Region 9 web service contains the following map layers:Abandoned Uranium Mines, Priority Mines, Tronox Mines, Navajo Environmental Response Trust Mines, Mines with Enforcement Actions, Superfund AUM Regions, Navajo Nation Administrative Boundaries and Chapter Houses.Mine points have a maximum scale of 1:220,000, while Mine polygons have a minimum scale of 1:220,000. Chapter houses have a minimum scale of 1:200,000. BLM Land Status has a minimum scale of 1:150,000.Full FGDC metadata records for each layer can be found by clicking the layer name at the web service endpoint and viewing the layer description. Data used to create this web service are available for download at https://edg.epa.gov/metadata/catalog/data/data.page.Security Classification: Public. Access Constraints: None. Use Constraints: None. Please check sources, scale, accuracy, currentness and other available information. Please confirm that you are using the most recent copy of both data and metadata. Acknowledgement of the EPA would be appreciated.

  4. Using an improved association rules mining optimization algorithm in web-based mobile-learning system

    NASA Astrophysics Data System (ADS)

    Huang, Yin; Chen, Jianhua; Xiong, Shaojun

    2009-07-01

    Mobile-Learning (M-learning) makes many learners get the advantages of both traditional learning and E-learning. Currently, Web-based Mobile-Learning Systems have created many new ways and defined new relationships between educators and learners. Association rule mining is one of the most important fields in data mining and knowledge discovery in databases. Rules explosion is a serious problem which causes great concerns, as conventional mining algorithms often produce too many rules for decision makers to digest. Since Web-based Mobile-Learning System collects vast amounts of student profile data, data mining and knowledge discovery techniques can be applied to find interesting relationships between attributes of learners, assessments, the solution strategies adopted by learners and so on. Therefore ,this paper focus on a new data-mining algorithm, combined with the advantages of genetic algorithm and simulated annealing algorithm , called ARGSA(Association rules based on an improved Genetic Simulated Annealing Algorithm), to mine the association rules. This paper first takes advantage of the Parallel Genetic Algorithm and Simulated Algorithm designed specifically for discovering association rules. Moreover, the analysis and experiment are also made to show the proposed method is superior to the Apriori algorithm in this Mobile-Learning system.

  5. A New Essential Functions Installed DWH in Hospital Information System: Process Mining Techniques and Natural Language Processing.

    PubMed

    Honda, Masayuki; Matsumoto, Takehiro

    2017-01-01

    Several kinds of event log data produced in daily clinical activities have yet to be used for secure and efficient improvement of hospital activities. Data Warehouse systems in Hospital Information Systems used for the analysis of structured data such as disease, lab-tests, and medications, have also shown efficient outcomes. This article is focused on two kinds of essential functions: process mining using log data and non-structured data analysis via Natural Language Processing.

  6. Graph Mining Meets the Semantic Web

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Lee, Sangkeun; Sukumar, Sreenivas R; Lim, Seung-Hwan

    The Resource Description Framework (RDF) and SPARQL Protocol and RDF Query Language (SPARQL) were introduced about a decade ago to enable flexible schema-free data interchange on the Semantic Web. Today, data scientists use the framework as a scalable graph representation for integrating, querying, exploring and analyzing data sets hosted at different sources. With increasing adoption, the need for graph mining capabilities for the Semantic Web has emerged. We address that need through implementation of three popular iterative Graph Mining algorithms (Triangle count, Connected component analysis, and PageRank). We implement these algorithms as SPARQL queries, wrapped within Python scripts. We evaluatemore » the performance of our implementation on 6 real world data sets and show graph mining algorithms (that have a linear-algebra formulation) can indeed be unleashed on data represented as RDF graphs using the SPARQL query interface.« less

  7. Mining Longitudinal Web Queries: Trends and Patterns.

    ERIC Educational Resources Information Center

    Wang, Peiling; Berry, Michael W.; Yang, Yiheng

    2003-01-01

    Analyzed user queries submitted to an academic Web site during a four-year period, using a relational database, to examine users' query behavior, to identify problems they encounter, and to develop techniques for optimizing query analysis and mining. Linguistic analyses focus on query structures, lexicon, and word associations using statistical…

  8. An Expertise Recommender using Web Mining

    NASA Technical Reports Server (NTRS)

    Joshi, Anupam; Chandrasekaran, Purnima; ShuYang, Michelle; Ramakrishnan, Ramya

    2001-01-01

    This report explored techniques to mine web pages of scientists to extract information regarding their expertise, build expertise chains and referral webs, and semi automatically combine this information with directory information services to create a recommender system that permits query by expertise. The approach included experimenting with existing techniques that have been reported in research literature in recent past , and adapted them as needed. In addition, software tools were developed to capture and use this information.

  9. Application of data mining in science and technology management information system based on WebGIS

    NASA Astrophysics Data System (ADS)

    Wu, Xiaofang; Xu, Zhiyong; Bao, Shitai; Chen, Feixiang

    2009-10-01

    With the rapid development of science and technology and the quick increase of information, a great deal of data is accumulated in the management department of science and technology. Usually, many knowledge and rules are contained and concealed in the data. Therefore, how to excavate and use the knowledge fully is very important in the management of science and technology. It will help to examine and approve the project of science and technology more scientifically and make the achievement transformed as the realistic productive forces easier. Therefore, the data mine technology will be researched and applied to the science and technology management information system to find and excavate the knowledge in the paper. According to analyzing the disadvantages of traditional science and technology management information system, the database technology, data mining and web geographic information systems (WebGIS) technology will be introduced to develop and construct the science and technology management information system based on WebGIS. The key problems are researched in detail such as data mining and statistical analysis. What's more, the prototype system is developed and validated based on the project data of National Natural Science Foundation Committee. The spatial data mining is done from the axis of time, space and other factors. Then the variety of knowledge and rules will be excavated by using data mining technology, which helps to provide an effective support for decisionmaking.

  10. Feasibility of an 8-week African American Web-based Pilot Program Promoting Healthy Eating Behaviors: Family Eats

    USDA-ARS?s Scientific Manuscript database

    To assess log-on rates and change in mediating variables achieved from a web-based nutrition intervention for African American families, a parent and 9- to 12-year-old daughter (n=67 families) completed questionnaires measuring dietary change mediating variables. Overall log-on rate was 59%. Signifi...

  11. Ed Blogs: A Vygotsky Dialectivist Technology Portal

    ERIC Educational Resources Information Center

    Hargis, Jace; Schofield, Kathleen

    2006-01-01

    This article provides a brief history of Web logging and subsequent examples of Web log use specific to an educational arena. Although the blogging phenomenon began a few years ago, it is only recently that a critical mass of educators has realized its potential in and out of the classroom for teaching and learning content while acting as a…

  12. WebCIS: large scale deployment of a Web-based clinical information system.

    PubMed

    Hripcsak, G; Cimino, J J; Sengupta, S

    1999-01-01

    WebCIS is a Web-based clinical information system. It sits atop the existing Columbia University clinical information system architecture, which includes a clinical repository, the Medical Entities Dictionary, an HL7 interface engine, and an Arden Syntax based clinical event monitor. WebCIS security features include authentication with secure tokens, authorization maintained in an LDAP server, SSL encryption, permanent audit logs, and application time outs. WebCIS is currently used by 810 physicians at the Columbia-Presbyterian center of New York Presbyterian Healthcare to review and enter data into the electronic medical record. Current deployment challenges include maintaining adequate database performance despite complex queries, replacing large numbers of computers that cannot run modern Web browsers, and training users that have never logged onto the Web. Although the raised expectations and higher goals have increased deployment costs, the end result is a far more functional, far more available system.

  13. Binary Coded Web Access Pattern Tree in Education Domain

    ERIC Educational Resources Information Center

    Gomathi, C.; Moorthi, M.; Duraiswamy, K.

    2008-01-01

    Web Access Pattern (WAP), which is the sequence of accesses pursued by users frequently, is a kind of interesting and useful knowledge in practice. Sequential Pattern mining is the process of applying data mining techniques to a sequential database for the purposes of discovering the correlation relationships that exist among an ordered list of…

  14. Feasibility and preliminary efficacy of a web-based smoking cessation intervention for HIV-infected smokers: a randomized controlled trial.

    PubMed

    Shuter, Jonathan; Morales, Daniela A; Considine-Dunn, Shannon E; An, Lawrence C; Stanton, Cassandra A

    2014-09-01

    To evaluate the feasibility and preliminary efficacy of a Web-based tobacco treatment for persons living with HIV (PLWH). Prospective, randomized controlled trial. HIV-care center in the Bronx, New York. Eligibility criteria included HIV infection, current tobacco usage, interest in quitting, and access to a computer with internet. One hundred thirty-eight subjects enrolled, and 134 completed the study. Positively Smoke Free on the Web (PSFW), an 8-session, 7-week targeted tobacco treatment program for PLWH, was compared with standard care (brief advice to quit and self-help brochure). All subjects were offered nicotine patches. The main feasibility outcomes were number of sessions logged into, number of Web pages visited, number of interactive clicks, and total time logged in. The main efficacy outcome was biochemically verified, 7-day point prevalence abstinence 3 months after intervention. PSFW subjects logged into a mean of 5.5 of 8 sessions and 26.2 of 41 pages. They executed a mean of 10 interactive clicks during a mean total of 59.8 minutes logged in. Most required reminder phone calls to complete the sessions. Educational level, anxiety score, and home access of the Web site were associated with Web site usage. Ten percent of the PSFW group vs. 4.3% of controls achieved the abstinence end point. Among those who completed all 8 sessions, 17.9% were abstinent, and among women completers, 30.8% were abstinent. Web-based treatment is a feasible strategy for PLWH smokers, and preliminary findings suggest therapeutic efficacy.

  15. Three diameter-limit cuttings in West Virginia hardwoods a 5-year report

    Treesearch

    Russell J. Hutnik

    1958-01-01

    Mine timbers are a basic need of West Virginia's giant coal industry. The annual requirement of sawed mine timbers is roughly 250 million board feet. The mines also use a large volume of wood in rough form for props and lagging. Yet, compared to sawlogs and veneer logs, these mine timbers are low-value products. This means that they must be produced at low cost....

  16. Text and Structural Data Mining of Influenza Mentions in Web and Social Media

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Corley, Courtney D.; Cook, Diane; Mikler, Armin R.

    Text and structural data mining of Web and social media (WSM) provides a novel disease surveillance resource and can identify online communities for targeted public health communications (PHC) to assure wide dissemination of pertinent information. WSM that mention influenza are harvested over a 24-week period, 5-October-2008 to 21-March-2009. Link analysis reveals communities for targeted PHC. Text mining is shown to identify trends in flu posts that correlate to real-world influenza-like-illness patient report data. We also bring to bear a graph-based data mining technique to detect anomalies among flu blogs connected by publisher type, links, and user-tags.

  17. Optimizing the Information Presentation on Mining Potential by using Web Services Technology with Restful Protocol

    NASA Astrophysics Data System (ADS)

    Abdillah, T.; Dai, R.; Setiawan, E.

    2018-02-01

    This study aims to develop the application of Web Services technology with RestFul Protocol to optimize the information presentation on mining potential. This study used User Interface Design approach for the information accuracy and relevance as well as the Web Service for the reliability in presenting the information. The results show that: the information accuracy and relevance regarding mining potential can be seen from the achievement of User Interface implementation in the application that is based on the following rules: The consideration of the appropriate colours and objects, the easiness of using the navigation, and users’ interaction with the applications that employs symbols and languages understood by the users; the information accuracy and relevance related to mining potential can be observed by the information presented by using charts and Tool Tip Text to help the users understand the provided chart/figure; the reliability of the information presentation is evident by the results of Web Services testing in Figure 4.5.6. This study finds out that User Interface Design and Web Services approaches (for the access of different Platform apps) are able to optimize the presentation. The results of this study can be used as a reference for software developers and Provincial Government of Gorontalo.

  18. Motivation Mining: Prospecting the Web.

    ERIC Educational Resources Information Center

    Small, Ruth V.; Arnone, Marilyn P.

    1999-01-01

    Describes WebMAC instruments, which differ from other Web-evaluation instruments because they have a theoretical base, are user-centered, are designed for students in grades 7 through 12, and assess the motivational quality of Web sites. Examples are given of uses of WebMAC Middle and WebMAC Senior in activities to promote evaluation and…

  19. Analysis of mesenchymal stem cell differentiation in vitro using classification association rule mining.

    PubMed

    Wang, Weiqi; Wang, Yanbo Justin; Bañares-Alcántara, René; Coenen, Frans; Cui, Zhanfeng

    2009-12-01

    In this paper, data mining is used to analyze the data on the differentiation of mammalian Mesenchymal Stem Cells (MSCs), aiming at discovering known and hidden rules governing MSC differentiation, following the establishment of a web-based public database containing experimental data on the MSC proliferation and differentiation. To this effect, a web-based public interactive database comprising the key parameters which influence the fate and destiny of mammalian MSCs has been constructed and analyzed using Classification Association Rule Mining (CARM) as a data-mining technique. The results show that the proposed approach is technically feasible and performs well with respect to the accuracy of (classification) prediction. Key rules mined from the constructed MSC database are consistent with experimental observations, indicating the validity of the method developed and the first step in the application of data mining to the study of MSCs.

  20. Design of a Web-tool for diagnostic clinical trials handling medical imaging research.

    PubMed

    Baltasar Sánchez, Alicia; González-Sistal, Angel

    2011-04-01

    New clinical studies in medicine are based on patients and controls using different imaging diagnostic modalities. Medical information systems are not designed for clinical trials employing clinical imaging. Although commercial software and communication systems focus on storage of image data, they are not suitable for storage and mining of new types of quantitative data. We sought to design a Web-tool to support diagnostic clinical trials involving different experts and hospitals or research centres. The image analysis of this project is based on skeletal X-ray imaging. It involves a computerised image method using quantitative analysis of regions of interest in healthy bone and skeletal metastases. The database is implemented with ASP.NET 3.5 and C# technologies for our Web-based application. For data storage, we chose MySQL v.5.0, one of the most popular open source databases. User logins were necessary, and access to patient data was logged for auditing. For security, all data transmissions were carried over encrypted connections. This Web-tool is available to users scattered at different locations; it allows an efficient organisation and storage of data (case report form) and images and allows each user to know precisely what his task is. The advantages of our Web-tool are as follows: (1) sustainability is guaranteed; (2) network locations for collection of data are secured; (3) all clinical information is stored together with the original images and the results derived from processed images and statistical analysis that enable us to perform retrospective studies; (4) changes are easily incorporated because of the modular architecture; and (5) assessment of trial data collected at different sites is centralised to reduce statistical variance.

  1. Mining Student Data Captured from a Web-Based Tutoring Tool: Initial Exploration and Results

    ERIC Educational Resources Information Center

    Merceron, Agathe; Yacef, Kalina

    2004-01-01

    In this article we describe the initial investigations that we have conducted on student data collected from a web-based tutoring tool. We have used some data mining techniques such as association rule and symbolic data analysis, as well as traditional SQL queries to gain further insight on the students' learning and deduce information to improve…

  2. Beyond accuracy: creating interoperable and scalable text-mining web services.

    PubMed

    Wei, Chih-Hsuan; Leaman, Robert; Lu, Zhiyong

    2016-06-15

    The biomedical literature is a knowledge-rich resource and an important foundation for future research. With over 24 million articles in PubMed and an increasing growth rate, research in automated text processing is becoming increasingly important. We report here our recently developed web-based text mining services for biomedical concept recognition and normalization. Unlike most text-mining software tools, our web services integrate several state-of-the-art entity tagging systems (DNorm, GNormPlus, SR4GN, tmChem and tmVar) and offer a batch-processing mode able to process arbitrary text input (e.g. scholarly publications, patents and medical records) in multiple formats (e.g. BioC). We support multiple standards to make our service interoperable and allow simpler integration with other text-processing pipelines. To maximize scalability, we have preprocessed all PubMed articles, and use a computer cluster for processing large requests of arbitrary text. Our text-mining web service is freely available at http://www.ncbi.nlm.nih.gov/CBBresearch/Lu/Demo/tmTools/#curl : Zhiyong.Lu@nih.gov. Published by Oxford University Press 2016. This work is written by US Government employees and is in the public domain in the US.

  3. Stratification-Based Outlier Detection over the Deep Web.

    PubMed

    Xian, Xuefeng; Zhao, Pengpeng; Sheng, Victor S; Fang, Ligang; Gu, Caidong; Yang, Yuanfeng; Cui, Zhiming

    2016-01-01

    For many applications, finding rare instances or outliers can be more interesting than finding common patterns. Existing work in outlier detection never considers the context of deep web. In this paper, we argue that, for many scenarios, it is more meaningful to detect outliers over deep web. In the context of deep web, users must submit queries through a query interface to retrieve corresponding data. Therefore, traditional data mining methods cannot be directly applied. The primary contribution of this paper is to develop a new data mining method for outlier detection over deep web. In our approach, the query space of a deep web data source is stratified based on a pilot sample. Neighborhood sampling and uncertainty sampling are developed in this paper with the goal of improving recall and precision based on stratification. Finally, a careful performance evaluation of our algorithm confirms that our approach can effectively detect outliers in deep web.

  4. Stratification-Based Outlier Detection over the Deep Web

    PubMed Central

    Xian, Xuefeng; Zhao, Pengpeng; Sheng, Victor S.; Fang, Ligang; Gu, Caidong; Yang, Yuanfeng; Cui, Zhiming

    2016-01-01

    For many applications, finding rare instances or outliers can be more interesting than finding common patterns. Existing work in outlier detection never considers the context of deep web. In this paper, we argue that, for many scenarios, it is more meaningful to detect outliers over deep web. In the context of deep web, users must submit queries through a query interface to retrieve corresponding data. Therefore, traditional data mining methods cannot be directly applied. The primary contribution of this paper is to develop a new data mining method for outlier detection over deep web. In our approach, the query space of a deep web data source is stratified based on a pilot sample. Neighborhood sampling and uncertainty sampling are developed in this paper with the goal of improving recall and precision based on stratification. Finally, a careful performance evaluation of our algorithm confirms that our approach can effectively detect outliers in deep web. PMID:27313603

  5. Demonstration of the Web-based Interspecies Correlation Estimation (Web-ICE) modeling application

    EPA Science Inventory

    The Web-based Interspecies Correlation Estimation (Web-ICE) modeling application is available to the risk assessment community through a user-friendly internet platform (http://epa.gov/ceampubl/fchain/webice/). ICE models are log-linear least square regressions that predict acute...

  6. Patterns of usage for a Web-based clinical information system.

    PubMed

    Chen, Elizabeth S; Cimino, James J

    2004-01-01

    Understanding how clinicians are using clinical information systems to assist with their everyday tasks is valuable to the system design and development process. Developers of such systems are interested in monitoring usage in order to make enhancements. System log files are rich resources for gaining knowledge about how the system is being used. We have analyzed the log files of our Web-based clinical information system (WebCIS) to obtain various usage statistics including which WebCIS features are frequently being used. We have also identified usage patterns, which convey how the user is traversing the system. We present our method and these results as well as describe how the results can be used to customize menus, shortcut lists, and patient reports in WebCIS and similar systems.

  7. Analysis of web-related threats in ten years of logs from a scientific portal

    NASA Astrophysics Data System (ADS)

    Santos, Rafael D. C.; Grégio, André R. A.; Raddick, Jordan; Vattki, Vamsi; Szalay, Alex

    2012-06-01

    SkyServer is an Internet portal to data from the Sloan Digital Sky Survey, the largest online archive of astronomy data in the world. provides free access to hundreds of millions of celestial objects for science, education and outreach purposes. Logs of accesses to SkyServer comprise around 930 million hits, 140 million web services accesses and 170 million SQL submitted queries, collected over the past 10 years. These logs also contain indications of compromise attempts on the servers. In this paper, we show some threats that were detected in ten years of stored logs, and compare them with known threats in those years. Also, we present an analysis of the evolution of those threats over these years.

  8. Deploying and sharing U-Compare workflows as web services.

    PubMed

    Kontonatsios, Georgios; Korkontzelos, Ioannis; Kolluru, Balakrishna; Thompson, Paul; Ananiadou, Sophia

    2013-02-18

    U-Compare is a text mining platform that allows the construction, evaluation and comparison of text mining workflows. U-Compare contains a large library of components that are tuned to the biomedical domain. Users can rapidly develop biomedical text mining workflows by mixing and matching U-Compare's components. Workflows developed using U-Compare can be exported and sent to other users who, in turn, can import and re-use them. However, the resulting workflows are standalone applications, i.e., software tools that run and are accessible only via a local machine, and that can only be run with the U-Compare platform. We address the above issues by extending U-Compare to convert standalone workflows into web services automatically, via a two-click process. The resulting web services can be registered on a central server and made publicly available. Alternatively, users can make web services available on their own servers, after installing the web application framework, which is part of the extension to U-Compare. We have performed a user-oriented evaluation of the proposed extension, by asking users who have tested the enhanced functionality of U-Compare to complete questionnaires that assess its functionality, reliability, usability, efficiency and maintainability. The results obtained reveal that the new functionality is well received by users. The web services produced by U-Compare are built on top of open standards, i.e., REST and SOAP protocols, and therefore, they are decoupled from the underlying platform. Exported workflows can be integrated with any application that supports these open standards. We demonstrate how the newly extended U-Compare enhances the cross-platform interoperability of workflows, by seamlessly importing a number of text mining workflow web services exported from U-Compare into Taverna, i.e., a generic scientific workflow construction platform.

  9. Deploying and sharing U-Compare workflows as web services

    PubMed Central

    2013-01-01

    Background U-Compare is a text mining platform that allows the construction, evaluation and comparison of text mining workflows. U-Compare contains a large library of components that are tuned to the biomedical domain. Users can rapidly develop biomedical text mining workflows by mixing and matching U-Compare’s components. Workflows developed using U-Compare can be exported and sent to other users who, in turn, can import and re-use them. However, the resulting workflows are standalone applications, i.e., software tools that run and are accessible only via a local machine, and that can only be run with the U-Compare platform. Results We address the above issues by extending U-Compare to convert standalone workflows into web services automatically, via a two-click process. The resulting web services can be registered on a central server and made publicly available. Alternatively, users can make web services available on their own servers, after installing the web application framework, which is part of the extension to U-Compare. We have performed a user-oriented evaluation of the proposed extension, by asking users who have tested the enhanced functionality of U-Compare to complete questionnaires that assess its functionality, reliability, usability, efficiency and maintainability. The results obtained reveal that the new functionality is well received by users. Conclusions The web services produced by U-Compare are built on top of open standards, i.e., REST and SOAP protocols, and therefore, they are decoupled from the underlying platform. Exported workflows can be integrated with any application that supports these open standards. We demonstrate how the newly extended U-Compare enhances the cross-platform interoperability of workflows, by seamlessly importing a number of text mining workflow web services exported from U-Compare into Taverna, i.e., a generic scientific workflow construction platform. PMID:23419017

  10. MyWEST: my Web Extraction Software Tool for effective mining of annotations from web-based databanks.

    PubMed

    Masseroli, Marco; Stella, Andrea; Meani, Natalia; Alcalay, Myriam; Pinciroli, Francesco

    2004-12-12

    High-throughput technologies create the necessity to mine large amounts of gene annotations from diverse databanks, and to integrate the resulting data. Most databanks can be interrogated only via Web, for a single gene at a time, and query results are generally available only in the HTML format. Although some databanks provide batch retrieval of data via FTP, this requires expertise and resources for locally reimplementing the databank. We developed MyWEST, a tool aimed at researchers without extensive informatics skills or resources, which exploits user-defined templates to easily mine selected annotations from different Web-interfaced databanks, and aggregates and structures results in an automatically updated database. Using microarray results from a model system of retinoic acid-induced differentiation, MyWEST effectively gathered relevant annotations from various biomolecular databanks, highlighted significant biological characteristics and supported a global approach to the understanding of complex cellular mechanisms. MyWEST is freely available for non-profit use at http://www.medinfopoli.polimi.it/MyWEST/

  11. Impacts of Intensive Logging on the Trophic Organisation of Ant Communities in a Biodiversity Hotspot

    PubMed Central

    Woodcock, Paul; Edwards, David P.; Newton, Rob J.; Vun Khen, Chey; Bottrell, Simon H.; Hamer, Keith C.

    2013-01-01

    Trophic organisation defines the flow of energy through ecosystems and is a key component of community structure. Widespread and intensifying anthropogenic disturbance threatens to disrupt trophic organisation by altering species composition and relative abundances and by driving shifts in the trophic ecology of species that persist in disturbed ecosystems. We examined how intensive disturbance caused by selective logging affects trophic organisation in the biodiversity hotspot of Sabah, Borneo. Using stable nitrogen isotopes, we quantified the positions in the food web of 159 leaf-litter ant species in unlogged and logged rainforest and tested four predictions: (i) there is a negative relationship between the trophic position of a species in unlogged forest and its change in abundance following logging, (ii) the trophic positions of species are altered by logging, (iii) disturbance alters the frequency distribution of trophic positions within the ant assemblage, and (iv) disturbance reduces food chain length. We found that ant abundance was 30% lower in logged forest than in unlogged forest but changes in abundance of individual species were not related to trophic position, providing no support for prediction (i). However, trophic positions of individual species were significantly higher in logged forest, supporting prediction (ii). Consequently, the frequency distribution of trophic positions differed significantly between unlogged and logged forest, supporting prediction (iii), and food chains were 0.2 trophic levels longer in logged forest, the opposite of prediction (iv). Our results demonstrate that disturbance can alter trophic organisation even without trophically-biased changes in community composition. Nonetheless, the absence of any reduction in food chain length in logged forest suggests that species-rich arthropod food webs do not experience trophic downgrading or a related collapse in trophic organisation despite the disturbance caused by logging. These food webs appear able to bend without breaking in the face of some forms of anthropogenic disturbance. PMID:23593302

  12. Impacts of intensive logging on the trophic organisation of ant communities in a biodiversity hotspot.

    PubMed

    Woodcock, Paul; Edwards, David P; Newton, Rob J; Vun Khen, Chey; Bottrell, Simon H; Hamer, Keith C

    2013-01-01

    Trophic organisation defines the flow of energy through ecosystems and is a key component of community structure. Widespread and intensifying anthropogenic disturbance threatens to disrupt trophic organisation by altering species composition and relative abundances and by driving shifts in the trophic ecology of species that persist in disturbed ecosystems. We examined how intensive disturbance caused by selective logging affects trophic organisation in the biodiversity hotspot of Sabah, Borneo. Using stable nitrogen isotopes, we quantified the positions in the food web of 159 leaf-litter ant species in unlogged and logged rainforest and tested four predictions: (i) there is a negative relationship between the trophic position of a species in unlogged forest and its change in abundance following logging, (ii) the trophic positions of species are altered by logging, (iii) disturbance alters the frequency distribution of trophic positions within the ant assemblage, and (iv) disturbance reduces food chain length. We found that ant abundance was 30% lower in logged forest than in unlogged forest but changes in abundance of individual species were not related to trophic position, providing no support for prediction (i). However, trophic positions of individual species were significantly higher in logged forest, supporting prediction (ii). Consequently, the frequency distribution of trophic positions differed significantly between unlogged and logged forest, supporting prediction (iii), and food chains were 0.2 trophic levels longer in logged forest, the opposite of prediction (iv). Our results demonstrate that disturbance can alter trophic organisation even without trophically-biased changes in community composition. Nonetheless, the absence of any reduction in food chain length in logged forest suggests that species-rich arthropod food webs do not experience trophic downgrading or a related collapse in trophic organisation despite the disturbance caused by logging. These food webs appear able to bend without breaking in the face of some forms of anthropogenic disturbance.

  13. 76 FR 60474 - Intent To Prepare a Draft Environmental Impact Statement (DEIS) for the Haile Gold Mine in...

    Federal Register 2010, 2011, 2012, 2013, 2014

    2011-09-29

    ...--County on January 28, 2011. The public notice is available on Charleston District's public Web site at... eight open mining pits over a twelve-year period, with pit depths ranging from 110 to 840 feet deep. The... of January 28, 2011, and are available on Charleston District's public Web site at http://www.sac...

  14. Competency-based residency training and the web log: modeling practice-based learning and enhancing medical knowledge.

    PubMed

    Hollon, Matthew F

    2015-01-01

    By using web-based tools in medical education, there are opportunities to innovatively teach important principles from the general competencies of graduate medical education. Postulating that faculty transparency in learning from uncertainties in clinical work could help residents to incorporate the principles of practice-based learning and improvement (PBLI) in their professional development, faculty in this community-based residency program modeled the steps of PBLI on a weekly basis through the use of a web log. The program confidentially surveyed residents before and after this project about actions consistent with PBLI and knowledge acquired through reading the web log. The frequency that residents encountered clinical situations where they felt uncertain declined over the course of the 24 weeks of the project from a mean frequency of uncertainty of 36% to 28% (Wilcoxon signed rank test, p=0.008); however, the frequency with which residents sought answers when faced with uncertainty did not change (Wilcoxon signed rank test, p=0.39), remaining high at approximately 80%. Residents answered a mean of 52% of knowledge questions correct when tested prior to faculty posts to the blog, rising to a mean of 65% of questions correct when tested at the end of the project (paired t-test, p=0.001). Faculty role modeling of PBLI behaviors and posting clinical questions and answers to a web log led to modest improvements in medical knowledge but did not alter behavior that was already taking place frequently among residents.

  15. A Quantitative Cost Effectiveness Model for Web-Supported Academic Instruction

    ERIC Educational Resources Information Center

    Cohen, Anat; Nachmias, Rafi

    2006-01-01

    This paper describes a quantitative cost effectiveness model for Web-supported academic instruction. The model was designed for Web-supported instruction (rather than distance learning only) characterizing most of the traditional higher education institutions. It is based on empirical data (Web logs) of students' and instructors' usage…

  16. Effective Filtering of Query Results on Updated User Behavioral Profiles in Web Mining

    PubMed Central

    Sadesh, S.; Suganthe, R. C.

    2015-01-01

    Web with tremendous volume of information retrieves result for user related queries. With the rapid growth of web page recommendation, results retrieved based on data mining techniques did not offer higher performance filtering rate because relationships between user profile and queries were not analyzed in an extensive manner. At the same time, existing user profile based prediction in web data mining is not exhaustive in producing personalized result rate. To improve the query result rate on dynamics of user behavior over time, Hamilton Filtered Regime Switching User Query Probability (HFRS-UQP) framework is proposed. HFRS-UQP framework is split into two processes, where filtering and switching are carried out. The data mining based filtering in our research work uses the Hamilton Filtering framework to filter user result based on personalized information on automatic updated profiles through search engine. Maximized result is fetched, that is, filtered out with respect to user behavior profiles. The switching performs accurate filtering updated profiles using regime switching. The updating in profile change (i.e., switches) regime in HFRS-UQP framework identifies the second- and higher-order association of query result on the updated profiles. Experiment is conducted on factors such as personalized information search retrieval rate, filtering efficiency, and precision ratio. PMID:26221626

  17. DOE Office of Scientific and Technical Information (OSTI.GOV)

    Raymond, David W.; Gaither, Katherine N.; Polsky, Yarom

    Sandia National Laboratories (Sandia) has a long history in developing compact, mobile, very high-speed drilling systems and this technology could be applied to increasing the rate at which boreholes are drilled during a mine accident response. The present study reviews current technical approaches, primarily based on technology developed under other programs, analyzes mine rescue specific requirements to develop a conceptual mine rescue drilling approach, and finally, proposes development of a phased mine rescue drilling system (MRDS) that accomplishes (1) development of rapid drilling MRDS equipment; (2) structuring improved web communication through the Mine Safety & Health Administration (MSHA) web site;more » (3) development of an improved protocol for employment of existing drilling technology in emergencies; (4) deployment of advanced technologies to complement mine rescue drilling operations during emergency events; and (5) preliminary discussion of potential future technology development of specialized MRDS equipment. This phased approach allows for rapid fielding of a basic system for improved rescue drilling, with the ability to improve the system over time at a reasonable cost.« less

  18. A Two-Tiered Model for Analyzing Library Web Site Usage Statistics, Part 1: Web Server Logs.

    ERIC Educational Resources Information Center

    Cohen, Laura B.

    2003-01-01

    Proposes a two-tiered model for analyzing web site usage statistics for academic libraries: one tier for library administrators that analyzes measures indicating library use, and a second tier for web site managers that analyzes measures aiding in server maintenance and site design. Discusses the technology of web site usage statistics, and…

  19. Recommendations for Benchmarking Web Site Usage among Academic Libraries.

    ERIC Educational Resources Information Center

    Hightower, Christy; Sih, Julie; Tilghman, Adam

    1998-01-01

    To help library directors and Web developers create a benchmarking program to compare statistics of academic Web sites, the authors analyzed the Web server log files of 14 university science and engineering libraries. Recommends a centralized voluntary reporting structure coordinated by the Association of Research Libraries (ARL) and a method for…

  20. Anthropogenic and natural sources of acidity and metals and their influence on the structure of stream food webs.

    PubMed

    Hogsden, Kristy L; Harding, Jon S

    2012-03-01

    We compared food web structure in 20 streams with either anthropogenic or natural sources of acidity and metals or circumneutral water chemistry in New Zealand. Community and diet analysis indicated that mining streams receiving anthropogenic inputs of acidic and metal-rich drainage had much simpler food webs (fewer species, shorter food chains, less links) than those in naturally acidic, naturally high metal, and circumneutral streams. Food webs of naturally high metal streams were structurally similar to those in mining streams, lacking fish predators and having few species. Whereas, webs in naturally acidic streams differed very little from those in circumneutral streams due to strong similarities in community composition and diets of secondary and top consumers. The combined negative effects of acidity and metals on stream food webs are clear. However, elevated metal concentrations, regardless of source, appear to play a more important role than acidity in driving food web structure. Copyright © 2011 Elsevier Ltd. All rights reserved.

  1. Social Web mining and exploitation for serious applications: Technosocial Predictive Analytics and related technologies for public health, environmental and national security surveillance.

    PubMed

    Kamel Boulos, Maged N; Sanfilippo, Antonio P; Corley, Courtney D; Wheeler, Steve

    2010-10-01

    This paper explores Technosocial Predictive Analytics (TPA) and related methods for Web "data mining" where users' posts and queries are garnered from Social Web ("Web 2.0") tools such as blogs, micro-blogging and social networking sites to form coherent representations of real-time health events. The paper includes a brief introduction to commonly used Social Web tools such as mashups and aggregators, and maps their exponential growth as an open architecture of participation for the masses and an emerging way to gain insight about people's collective health status of whole populations. Several health related tool examples are described and demonstrated as practical means through which health professionals might create clear location specific pictures of epidemiological data such as flu outbreaks. Copyright 2010 Elsevier Ireland Ltd. All rights reserved.

  2. Users' information-seeking behavior on a medical library Website

    PubMed Central

    Rozic-Hristovski, Anamarija; Hristovski, Dimitar; Todorovski, Ljupco

    2002-01-01

    The Central Medical Library (CMK) at the Faculty of Medicine, University of Ljubljana, Slovenia, started to build a library Website that included a guide to library services and resources in 1997. The evaluation of Website usage plays an important role in its maintenance and development. Analyzing and exploring regularities in the visitors' behavior can be used to enhance the quality and facilitate delivery of information services, identify visitors' interests, and improve the server's performance. The analysis of the CMK Website users' navigational behavior was carried out by analyzing the Web server log files. These files contained information on all user accesses to the Website and provided a great opportunity to learn more about the behavior of visitors to the Website. The majority of the available tools for Web log file analysis provide a predefined set of reports showing the access count and the transferred bytes grouped along several dimensions. In addition to the reports mentioned above, the authors wanted to be able to perform interactive exploration and ad hoc analysis and discover trends in a user-friendly way. Because of that, we developed our own solution for exploring and analyzing the Web logs based on data warehousing and online analytical processing technologies. The analytical solution we developed proved successful, so it may find further application in the field of Web log file analysis. We will apply the findings of the analysis to restructuring the CMK Website. PMID:11999179

  3. Competency-based residency training and the web log: modeling practice-based learning and enhancing medical knowledge†

    PubMed Central

    Hollon, Matthew F.

    2015-01-01

    Background By using web-based tools in medical education, there are opportunities to innovatively teach important principles from the general competencies of graduate medical education. Objectives Postulating that faculty transparency in learning from uncertainties in clinical work could help residents to incorporate the principles of practice-based learning and improvement (PBLI) in their professional development, faculty in this community-based residency program modeled the steps of PBLI on a weekly basis through the use of a web log. Method The program confidentially surveyed residents before and after this project about actions consistent with PBLI and knowledge acquired through reading the web log. Results The frequency that residents encountered clinical situations where they felt uncertain declined over the course of the 24 weeks of the project from a mean frequency of uncertainty of 36% to 28% (Wilcoxon signed rank test, p=0.008); however, the frequency with which residents sought answers when faced with uncertainty did not change (Wilcoxon signed rank test, p=0.39), remaining high at approximately 80%. Residents answered a mean of 52% of knowledge questions correct when tested prior to faculty posts to the blog, rising to a mean of 65% of questions correct when tested at the end of the project (paired t-test, p=0.001). Conclusions Faculty role modeling of PBLI behaviors and posting clinical questions and answers to a web log led to modest improvements in medical knowledge but did not alter behavior that was already taking place frequently among residents. PMID:26653701

  4. Rule-based statistical data mining agents for an e-commerce application

    NASA Astrophysics Data System (ADS)

    Qin, Yi; Zhang, Yan-Qing; King, K. N.; Sunderraman, Rajshekhar

    2003-03-01

    Intelligent data mining techniques have useful e-Business applications. Because an e-Commerce application is related to multiple domains such as statistical analysis, market competition, price comparison, profit improvement and personal preferences, this paper presents a hybrid knowledge-based e-Commerce system fusing intelligent techniques, statistical data mining, and personal information to enhance QoS (Quality of Service) of e-Commerce. A Web-based e-Commerce application software system, eDVD Web Shopping Center, is successfully implemented uisng Java servlets and an Oracle81 database server. Simulation results have shown that the hybrid intelligent e-Commerce system is able to make smart decisions for different customers.

  5. Research on the optimization strategy of web search engine based on data mining

    NASA Astrophysics Data System (ADS)

    Chen, Ronghua

    2018-04-01

    With the wide application of search engines, web site information has become an important way for people to obtain information. People have found that they are growing in an increasingly explosive manner. Web site information is verydifficult to find the information they need, and now the search engine can not meet the need, so there is an urgent need for the network to provide website personalized information service, data mining technology for this new challenge is to find a breakthrough. In order to improve people's accuracy of finding information from websites, a website search engine optimization strategy based on data mining is proposed, and verified by website search engine optimization experiment. The results show that the proposed strategy improves the accuracy of the people to find information, and reduces the time for people to find information. It has an important practical value.

  6. Average growth rates in the spruce-fir region of New England, based on remeasured plots

    Treesearch

    C. Allen Bickford; Franklin R. Longwood; Robert Bain

    1961-01-01

    The fact that trees grow makes the extraction of logs and bolts from a forest different from the extraction of ore from a mine. Every mine has a limit, and sooner or later it must be abandoned, either because the vein has run out or because the mine has become too costly to operate. But a properly managed forest - one where cutting maintains production - never becomes...

  7. Informal Learning through Expertise Mining in the Social Web

    ERIC Educational Resources Information Center

    Valencia-Garcia, Rafael; Garcia-Sanchez, Francisco; Casado-Lumbreras, Cristina; Castellanos-Nieves, Dagoberto; Fernandez-Breis, Jesualdo Tomas

    2012-01-01

    The advent of Web 2.0, also called the Social Web, has changed the way people interact with the Web. Assisted by the technologies associated with this new trend, users now play a much more active role as content providers. This Web paradigm shift has also changed how companies operate and interact with their employees, partners and customers. The…

  8. Web-of-Objects (WoO)-Based Context Aware Emergency Fire Management Systems for the Internet of Things

    PubMed Central

    Shamszaman, Zia Ush; Ara, Safina Showkat; Chong, Ilyoung; Jeong, Youn Kwae

    2014-01-01

    Recent advancements in the Internet of Things (IoT) and the Web of Things (WoT) accompany a smart life where real world objects, including sensing devices, are interconnected with each other. The Web representation of smart objects empowers innovative applications and services for various domains. To accelerate this approach, Web of Objects (WoO) focuses on the implementation aspects of bringing the assorted real world objects to the Web applications. In this paper; we propose an emergency fire management system in the WoO infrastructure. Consequently, we integrate the formation and management of Virtual Objects (ViO) which are derived from real world physical objects and are virtually connected with each other into the semantic ontology model. The charm of using the semantic ontology is that it allows information reusability, extensibility and interoperability, which enable ViOs to uphold orchestration, federation, collaboration and harmonization. Our system is context aware, as it receives contextual environmental information from distributed sensors and detects emergency situations. To handle a fire emergency, we present a decision support tool for the emergency fire management team. The previous fire incident log is the basis of the decision support system. A log repository collects all the emergency fire incident logs from ViOs and stores them in a repository. PMID:24531299

  9. Web-of-Objects (WoO)-based context aware emergency fire management systems for the Internet of Things.

    PubMed

    Shamszaman, Zia Ush; Ara, Safina Showkat; Chong, Ilyoung; Jeong, Youn Kwae

    2014-02-13

    Recent advancements in the Internet of Things (IoT) and the Web of Things (WoT) accompany a smart life where real world objects, including sensing devices, are interconnected with each other. The Web representation of smart objects empowers innovative applications and services for various domains. To accelerate this approach, Web of Objects (WoO) focuses on the implementation aspects of bringing the assorted real world objects to the Web applications. In this paper; we propose an emergency fire management system in the WoO infrastructure. Consequently, we integrate the formation and management of Virtual Objects (ViO) which are derived from real world physical objects and are virtually connected with each other into the semantic ontology model. The charm of using the semantic ontology is that it allows information reusability, extensibility and interoperability, which enable ViOs to uphold orchestration, federation, collaboration and harmonization. Our system is context aware, as it receives contextual environmental information from distributed sensors and detects emergency situations. To handle a fire emergency, we present a decision support tool for the emergency fire management team. The previous fire incident log is the basis of the decision support system. A log repository collects all the emergency fire incident logs from ViOs and stores them in a repository.

  10. Do you see what I hear: experiments in multi-channel sound and 3D visualization for network monitoring?

    NASA Astrophysics Data System (ADS)

    Ballora, Mark; Hall, David L.

    2010-04-01

    Detection of intrusions is a continuing problem in network security. Due to the large volumes of data recorded in Web server logs, analysis is typically forensic, taking place only after a problem has occurred. This paper describes a novel method of representing Web log information through multi-channel sound, while simultaneously visualizing network activity using a 3-D immersive environment. We are exploring the detection of intrusion signatures and patterns, utilizing human aural and visual pattern recognition ability to detect intrusions as they occur. IP addresses and return codes are mapped to an informative and unobtrusive listening environment to act as a situational sound track of Web traffic. Web log data is parsed and formatted using Python, then read as a data array by the synthesis language SuperCollider [1], which renders it as a sonification. This can be done either for the study of pre-existing data sets or in monitoring Web traffic in real time. Components rendered aurally include IP address, geographical information, and server Return Codes. Users can interact with the data, speeding or slowing the speed of representation (for pre-existing data sets) or "mixing" sound components to optimize intelligibility for tracking suspicious activity.

  11. Predicting Student Actions in a Procedural Training Environment

    ERIC Educational Resources Information Center

    Riofrio-Luzcando, Diego; Ramirez, Jaime; Berrocal-Lobo, Marta

    2017-01-01

    Data mining is known to have a potential for predicting user performance. However, there are few studies that explore its potential for predicting student behavior in a procedural training environment. This paper presents a collective student model, which is built from past student logs. These logs are first grouped into clusters. Then, an…

  12. Reviews Equipment: Data logger Book: Imagined Worlds Equipment: Mini data loggers Equipment: PICAXE-18M2 data logger Books: Engineering: A Very Short Introduction and To Engineer Is Human Book: Soap, Science, & Flat-Screen TVs Equipment: uLog and SensorLab Web Watch

    NASA Astrophysics Data System (ADS)

    2012-07-01

    WE RECOMMEND Data logger Fourier NOVA LINK: data logging and analysis To Engineer is Human Engineering: essays and insights Soap, Science, & Flat-Screen TVs People, politics, business and science overlap uLog sensors and sensor adapter A new addition to the LogIT range offers simplicity and ease of use WORTH A LOOK Imagined Worlds Socio-scientific predictions for the future Mini light data logger and mini temperature data logger Small-scale equipment for schools SensorLab Plus LogIT's supporting software, with extra features HANDLE WITH CARE CAXE110P PICAXE-18M2 data logger Data logger 'on view' but disappoints Engineering: A Very Short Introduction A broad-brush treatment fails to satisfy WEB WATCH Two very different websites for students: advanced physics questions answered and a more general BBC science resource

  13. Electrofacies analysis for coal lithotype profiling based on high-resolution wireline log data

    NASA Astrophysics Data System (ADS)

    Roslin, A.; Esterle, J. S.

    2016-06-01

    The traditional approach to coal lithotype analysis is based on a visual characterisation of coal in core, mine or outcrop exposures. As not all wells are fully cored, the petroleum and coal mining industries increasingly use geophysical wireline logs for lithology interpretation.This study demonstrates a method for interpreting coal lithotypes from geophysical wireline logs, and in particular discriminating between bright or banded, and dull coal at similar densities to a decimetre level. The study explores the optimum combination of geophysical log suites for training the coal electrofacies interpretation, using neural network conception, and then propagating the results to wells with fewer wireline data. This approach is objective and has a recordable reproducibility and rule set.In addition to conventional gamma ray and density logs, laterolog resistivity, microresistivity and PEF data were used in the study. Array resistivity data from a compact micro imager (CMI tool) were processed into a single microresistivity curve and integrated with the conventional resistivity data in the cluster analysis. Microresistivity data were tested in the analysis to test the hypothesis that the improved vertical resolution of microresistivity curve can enhance the accuracy of the clustering analysis. The addition of PEF log allowed discrimination between low density bright to banded coal electrofacies and low density inertinite-rich dull electrofacies.The results of clustering analysis were validated statistically and the results of the electrofacies results were compared to manually derived coal lithotype logs.

  14. Student Consistency and Implications for Feedback in Online Assessment Systems

    ERIC Educational Resources Information Center

    Madhyastha, Tara M.; Tanimoto, Steven

    2009-01-01

    Most of the emphasis on mining online assessment logs has been to identify content-specific errors. However, the pattern of general "consistency" is domain independent, strongly related to performance, and can itself be a target of educational data mining. We demonstrate that simple consistency indicators are related to student outcomes,…

  15. 25 CFR 215.23 - Cooperation between superintendent and district mining supervisor.

    Code of Federal Regulations, 2011 CFR

    2011-04-01

    ... notices, reports, drill logs, maps, and records, and all other information relating to mining operations required by said regulations to be submitted by lessees, and shall maintain a file thereof for the superintendent. (b) The files of the Geological Survey supervisor relating to lead and zinc leases of Quapaw...

  16. Development of a Google-based search engine for data mining radiology reports.

    PubMed

    Erinjeri, Joseph P; Picus, Daniel; Prior, Fred W; Rubin, David A; Koppel, Paul

    2009-08-01

    The aim of this study is to develop a secure, Google-based data-mining tool for radiology reports using free and open source technologies and to explore its use within an academic radiology department. A Health Insurance Portability and Accountability Act (HIPAA)-compliant data repository, search engine and user interface were created to facilitate treatment, operations, and reviews preparatory to research. The Institutional Review Board waived review of the project, and informed consent was not required. Comprising 7.9 GB of disk space, 2.9 million text reports were downloaded from our radiology information system to a fileserver. Extensible markup language (XML) representations of the reports were indexed using Google Desktop Enterprise search engine software. A hypertext markup language (HTML) form allowed users to submit queries to Google Desktop, and Google's XML response was interpreted by a practical extraction and report language (PERL) script, presenting ranked results in a web browser window. The query, reason for search, results, and documents visited were logged to maintain HIPAA compliance. Indexing averaged approximately 25,000 reports per hour. Keyword search of a common term like "pneumothorax" yielded the first ten most relevant results of 705,550 total results in 1.36 s. Keyword search of a rare term like "hemangioendothelioma" yielded the first ten most relevant results of 167 total results in 0.23 s; retrieval of all 167 results took 0.26 s. Data mining tools for radiology reports will improve the productivity of academic radiologists in clinical, educational, research, and administrative tasks. By leveraging existing knowledge of Google's interface, radiologists can quickly perform useful searches.

  17. Students using visual thinking to learn science in a Web-based environment

    NASA Astrophysics Data System (ADS)

    Plough, Jean Margaret

    United States students' science test scores are low, especially in problem solving, and traditional science instruction could be improved. Consequently, visual thinking, constructing science structures, and problem solving in a web-based environment may be valuable strategies for improving science learning. This ethnographic study examined the science learning of fifteen fourth grade students in an after school computer club involving diverse students at an inner city school. The investigation was done from the perspective of the students, and it described the processes of visual thinking, web page construction, and problem solving in a web-based environment. The study utilized informal group interviews, field notes, Visual Learning Logs, and student web pages, and incorporated a Standards-Based Rubric which evaluated students' performance on eight science and technology standards. The Visual Learning Logs were drawings done on the computer to represent science concepts related to the Food Chain. Students used the internet to search for information on a plant or animal of their choice. Next, students used this internet information, with the information from their Visual Learning Logs, to make web pages on their plant or animal. Later, students linked their web pages to form Science Structures. Finally, students linked their Science Structures with the structures of other students, and used these linked structures as models for solving problems. Further, during informal group interviews, students answered questions about visual thinking, problem solving, and science concepts. The results of this study showed clearly that (1) making visual representations helped students understand science knowledge, (2) making links between web pages helped students construct Science Knowledge Structures, and (3) students themselves said that visual thinking helped them learn science. In addition, this study found that when using Visual Learning Logs, the main overall ideas of the science concepts were usually represented accurately. Further, looking for information on the internet may cause new problems in learning. Likewise, being absent, starting late, and/or dropping out all may negatively influence students' proficiency on the standards. Finally, the way Science Structures are constructed and linked may provide insights into the way individual students think and process information.

  18. Upper Animas Mining District

    EPA Pesticide Factsheets

    Web page provides narrative of What's New?, Site Description, Site Risk, Cleanup Progress, Community Involvement, Next Steps, Site Documents, FAQ, Contacts and LInks for the Upper Animas Mining District site in San Juan County, Colorado.

  19. Process mining techniques: an application to time management

    NASA Astrophysics Data System (ADS)

    Khowaja, Ali Raza

    2018-04-01

    In an environment people have to make sure that all of their work are completed within a given time in accordance with its quality. In order to achieve the real phenomenon of process mining one needs to understand all of these processes in a detailed manner. Personal Information and communication has always been a highlighting issue on internet but for now information and communication tools within factual life refers to their daily schedule, location analysis, environmental analysis and, more generally, social media applications support these systems which makes data available for data analysis generated through event logs, but also for process analysis which combines environmental and location analysis. Process mining can be used to exploit all these real live processes with the help of the event logs which are already available in those datasets through user censored data or may be user labeled data. These processes could be used to redesign a user's flow and understand all these processes in a bit more detailed manner. In order to increase the quality of each of the processes that we go through our daily lives is to give a closer look to each of the processes and after analyzing them, one should make changes to get better results. On the contrarily, we applied process mining techniques on seven different subjects combined in a single dataset collected from Korea. Above all, the following paper comments on the efficiency of processes in the event logs referring to time management's sphere of influence.

  20. Data Mining of Extremely Large Ad-Hoc Data Sets to Produce Reverse Web-Link Graphs

    DTIC Science & Technology

    2017-03-01

    in most of the MR cases. From these studies , we also learned that computing -optimized instances should be chosen for serialized/compressed input data...maximum 200 words) Data mining can be a valuable tool, particularly in the acquisition of military intelligence. As the second study within a larger Naval...open web crawler data set Common Crawl. Similar to previous studies , this research employs MapReduce (MR) for sorting and categorizing output value

  1. Mining Social Media and Web Searches For Disease Detection

    PubMed Central

    Yang, Y. Tony; Horneffer, Michael; DiLisio, Nicole

    2013-01-01

    Web-based social media is increasingly being used across different settings in the health care industry. The increased frequency in the use of the Internet via computer or mobile devices provides an opportunity for social media to be the medium through which people can be provided with valuable health information quickly and directly. While traditional methods of detection relied predominately on hierarchical or bureaucratic lines of communication, these often failed to yield timely and accurate epidemiological intelligence. New web-based platforms promise increased opportunities for a more timely and accurate spreading of information and analysis. This article aims to provide an overview and discussion of the availability of timely and accurate information. It is especially useful for the rapid identification of an outbreak of an infectious disease that is necessary to promptly and effectively develop public health responses. These web-based platforms include search queries, data mining of web and social media, process and analysis of blogs containing epidemic key words, text mining, and geographical information system data analyses. These new sources of analysis and information are intended to complement traditional sources of epidemic intelligence. Despite the attractiveness of these new approaches, further study is needed to determine the accuracy of blogger statements, as increases in public participation may not necessarily mean the information provided is more accurate. PMID:25170475

  2. Mining social media and web searches for disease detection.

    PubMed

    Yang, Y Tony; Horneffer, Michael; DiLisio, Nicole

    2013-04-28

    Web-based social media is increasingly being used across different settings in the health care industry. The increased frequency in the use of the Internet via computer or mobile devices provides an opportunity for social media to be the medium through which people can be provided with valuable health information quickly and directly. While traditional methods of detection relied predominately on hierarchical or bureaucratic lines of communication, these often failed to yield timely and accurate epidemiological intelligence. New web-based platforms promise increased opportunities for a more timely and accurate spreading of information and analysis. This article aims to provide an overview and discussion of the availability of timely and accurate information. It is especially useful for the rapid identification of an outbreak of an infectious disease that is necessary to promptly and effectively develop public health responses. These web-based platforms include search queries, data mining of web and social media, process and analysis of blogs containing epidemic key words, text mining, and geographical information system data analyses. These new sources of analysis and information are intended to complement traditional sources of epidemic intelligence. Despite the attractiveness of these new approaches, further study is needed to determine the accuracy of blogger statements, as increases in public participation may not necessarily mean the information provided is more accurate.

  3. GoWeb: a semantic search engine for the life science web.

    PubMed

    Dietze, Heiko; Schroeder, Michael

    2009-10-01

    Current search engines are keyword-based. Semantic technologies promise a next generation of semantic search engines, which will be able to answer questions. Current approaches either apply natural language processing to unstructured text or they assume the existence of structured statements over which they can reason. Here, we introduce a third approach, GoWeb, which combines classical keyword-based Web search with text-mining and ontologies to navigate large results sets and facilitate question answering. We evaluate GoWeb on three benchmarks of questions on genes and functions, on symptoms and diseases, and on proteins and diseases. The first benchmark is based on the BioCreAtivE 1 Task 2 and links 457 gene names with 1352 functions. GoWeb finds 58% of the functional GeneOntology annotations. The second benchmark is based on 26 case reports and links symptoms with diseases. GoWeb achieves 77% success rate improving an existing approach by nearly 20%. The third benchmark is based on 28 questions in the TREC genomics challenge and links proteins to diseases. GoWeb achieves a success rate of 79%. GoWeb's combination of classical Web search with text-mining and ontologies is a first step towards answering questions in the biomedical domain. GoWeb is online at: http://www.gopubmed.org/goweb.

  4. Beyond Google: The Invisible Web in the Academic Library

    ERIC Educational Resources Information Center

    Devine, Jane; Egger-Sider, Francine

    2004-01-01

    This article analyzes the concept of the Invisible Web and its implication for academic librarianship. It offers a guide to tools that can be used to mine the Invisible Web and discusses the benefits of using the Invisible Web to promote interest in library services. In addition, the article includes an expanded definition, a literature review,…

  5. Intelligent Information Retrieval and Web Mining Architecture Using SOA

    ERIC Educational Resources Information Center

    El-Bathy, Naser Ibrahim

    2010-01-01

    The study of this dissertation provides a solution to a very specific problem instance in the area of data mining, data warehousing, and service-oriented architecture in publishing and newspaper industries. The research question focuses on the integration of data mining and data warehousing. The research problem focuses on the development of…

  6. Introducing Text Analytics as a Graduate Business School Course

    ERIC Educational Resources Information Center

    Edgington, Theresa M.

    2011-01-01

    Text analytics refers to the process of analyzing unstructured data from documented sources, including open-ended surveys, blogs, and other types of web dialog. Text analytics has enveloped the concept of text mining, an analysis approach influenced heavily from data mining. While text mining has been covered extensively in various computer…

  7. [Distribution characteristics of soil nematodes in reclaimed land of copper-mine-tailings in different plant associations].

    PubMed

    Zhu, Yong-heng; Li, Ke-zhong; Zhang, Heng; Han, Fei; Zhou, Ju-hua; Gao, Ting-ting

    2015-02-01

    A survey was carried out to investigate soil nematode communities in the plant associations of gramineae (Arthraxon lanceolatus, AL; Imperata cylindrica, IC) and leguminosae (Glycine soja, GS) in reclaimed land of copper-mine-tailings and in the plant associations of gramineae (Digitaria chrysoblephara, DC-CK) of peripheral control in Fenghuang Mountain, Tongling City. A total of 1277 nematodes were extracted and sorted into 51 genera. The average individual density of the nematodes was 590 individuals · 100 g(-1) dry soil. In order to analyze the distribution character- istics of soil nematode communities in reclaimed land of copper-mine-tailings, Shannon community diversity index and soil food web structure indices were applied in the research. The results showed that the total number of nematode genus and the Shannon community diversity index of soil nematode in the three plant associations of AL, IC and GS were less than that in the plant associations of DC-CK. Compared with the ecological indices of soil nematode communities among the different plant associations in reclaimed land of copper-mine-tailings and peripheral natural habitat, we found that the structure of soil food web in the plant associations of GS was more mature, with bacterial decomposition being dominant in the soil organic matter decomposition, and that the soil ecosystem in the plant associations of GS was not stable with low interference. This indicated that the soil food web in the plant associations of leguminosae had a greater development potential to improve the ecological stability of the reclaimed land of copper-mine-tailings. On the other hand, the structure of soil food web in the plant associations of AL and IC were relatively stable in a structured state with fungal decomposition being dominant in the decomposition of soil organic matter. This indicated that the soil food web in the plant associations of gramineae was at a poor development level.

  8. Educational Data Mining and Problem-Based Learning

    ERIC Educational Resources Information Center

    Walldén, Sari; Mäkinen, Erkki

    2014-01-01

    This paper considers the use of log data provided by learning management systems when studying whether students obey the problem-based learning (PBL) method. Log analysis turns out to be a valuable tool in measuring the use of the learning material of interest. It gives reliable figures concerning not only the number of use sessions but also the…

  9. Statistically-Driven Visualizations of Student Interactions with a French Online Course Video

    ERIC Educational Resources Information Center

    Youngs, Bonnie L.; Prakash, Akhil; Nugent, Rebecca

    2018-01-01

    Logged tracking data for online courses are generally not available to instructors, students, and course designers and developers, and even if these data were available, most content-oriented instructors do not have the skill set to analyze them. Learning analytics, mined from logged course data and usually presented in the form of learning…

  10. Chapter 8: Mine reclamation practices to enhance forest development through natural succession

    Treesearch

    J. Groninger; J. Skousen; P. Angel; C. Barton; J. Burger; C. Zipper

    2017-01-01

    "Natural succession" is a term used to describe natural changes in plant community composition over time. In the forested Appalachian region, disturbances from storms, fire, logging, or mining can disrupt or destroy established forests. Natural processes that lead to restoration of the forest vegetation following such a disturbance usually begin quickly and...

  11. How does abundance scale with body size in coupled size-structured food webs?

    PubMed

    Blanchard, Julia L; Jennings, Simon; Law, Richard; Castle, Matthew D; McCloghrie, Paul; Rochet, Marie-Joëlle; Benoît, Eric

    2009-01-01

    1. Widely observed macro-ecological patterns in log abundance vs. log body mass of organisms can be explained by simple scaling theory based on food (energy) availability across a spectrum of body sizes. The theory predicts that when food availability falls with body size (as in most aquatic food webs where larger predators eat smaller prey), the scaling between log N vs. log m is steeper than when organisms of different sizes compete for a shared unstructured resource (e.g. autotrophs, herbivores and detritivores; hereafter dubbed 'detritivores'). 2. In real communities, the mix of feeding characteristics gives rise to complex food webs. Such complexities make empirical tests of scaling predictions prone to error if: (i) the data are not disaggregated in accordance with the assumptions of the theory being tested, or (ii) the theory does not account for all of the trophic interactions within and across the communities sampled. 3. We disaggregated whole community data collected in the North Sea into predator and detritivore components and report slopes of log abundance vs. log body mass relationships. Observed slopes for fish and epifaunal predator communities (-1.2 to -2.25) were significantly steeper than those for infaunal detritivore communities (-0.56 to -0.87). 4. We present a model describing the dynamics of coupled size spectra, to explain how coupling of predator and detritivore communities affects the scaling of log N vs. log m. The model captures the trophic interactions and recycling of material that occur in many aquatic ecosystems. 5. Our simulations demonstrate that the biological processes underlying growth and mortality in the two distinct size spectra lead to patterns consistent with data. Slopes of log N vs. log m were steeper and growth rates faster for predators compared to detritivores. Size spectra were truncated when primary production was too low for predators and when detritivores experienced predation pressure. 6. The approach also allows us to assess the effects of external sources of mortality (e.g. harvesting). Removal of large predators resulted in steeper predator spectra and increases in their prey (small fish and detritivores). The model predictions are remarkably consistent with observed patterns of exploited ecosystems.

  12. Mining Tasks from the Web Anchor Text Graph: MSR Notebook Paper for the TREC 2015 Tasks Track

    DTIC Science & Technology

    2015-11-20

    Mining Tasks from the Web Anchor Text Graph: MSR Notebook Paper for the TREC 2015 Tasks Track Paul N. Bennett Microsoft Research Redmond, USA pauben...anchor text graph has proven useful in the general realm of query reformulation [2], we sought to quantify the value of extracting key phrases from...anchor text in the broader setting of the task understanding track. Given a query, our approach considers a simple method for identifying a relevant

  13. A Tactical Framework for Cyberspace Situational Awareness

    DTIC Science & Technology

    2010-06-01

    Command & Control 1. VOIP Telephone 2. Internet Chat 3. Web App ( TBMCS ) 4. Email 5. Web App (PEX) 6. Database (CAMS) 7. Database (ARMS) 8...Database (LogMod) 9. Resource (WWW) 10. Application (PFPS) Mission Planning 1. Application (PFPS) 2. Email 3. Web App ( TBMCS ) 4. Internet Chat...1. Web App (PEX) 2. Database (ARMS) 3. Web App ( TBMCS ) 4. Email 5. Database (CAMS) 6. VOIP Telephone 7. Application (PFPS) 8. Internet Chat 9

  14. Web Camera Use of Mothers and Fathers When Viewing Their Hospitalized Neonate.

    PubMed

    Rhoads, Sarah J; Green, Angela; Gauss, C Heath; Mitchell, Anita; Pate, Barbara

    2015-12-01

    Mothers and fathers of neonates hospitalized in a neonatal intensive care unit (NICU) differ in their experiences related to NICU visitation. To describe the frequency and length of maternal and paternal viewing of their hospitalized neonates via a Web camera. A total of 219 mothers and 101 fathers used the Web camera that allows 24/7 NICU viewing from September 1, 2010, to December 31, 2012, which included 40 mother and father dyads. We conducted a review of the Web camera's Web site log-on records in this nonexperimental, descriptive study. Mothers and fathers had a significant difference in the mean number of log-ons to the Web camera system (P = .0293). Fathers virtually visited the NICU less often than mothers, but there was not a statistical difference between mothers and fathers in terms of the mean total number of minutes viewing the neonate (P = .0834) or in the maximum number of minutes of viewing in 1 session (P = .6924). Patterns of visitations over time were not measured. Web camera technology could be a potential intervention to aid fathers in visiting their neonates. Both parents should be offered virtual visits using the Web camera and oriented regarding how to use the Web camera. These findings are important to consider when installing Web cameras in a NICU. Future research should continue to explore Web camera use in NICUs.

  15. Using Web Metric Software to Drive: Mobile Website Development

    ERIC Educational Resources Information Center

    Tidal, Junior

    2011-01-01

    Many libraries have developed mobile versions of their websites. In order to understand their users, web developers have conducted both usability tests and focus groups, yet analytical software and web server logs can also be used to better understand users. Using data collected from these tools, the Ursula C. Schwerin Library has made informed…

  16. Signal Event Context: Trace Technologies of the habit@online

    ERIC Educational Resources Information Center

    Luke, Robert

    2003-01-01

    Web portals--those online environments that encourage users to trade personal information for the opportunity to personalise the information space--are experiencing a considerable resurgence in popularity. Web portals are web sites that allow users to log on with a username and password and create their very own datastructure. This datastructure…

  17. BAGEL4: a user-friendly web server to thoroughly mine RiPPs and bacteriocins.

    PubMed

    van Heel, Auke J; de Jong, Anne; Song, Chunxu; Viel, Jakob H; Kok, Jan; Kuipers, Oscar P

    2018-05-21

    Interest in secondary metabolites such as RiPPs (ribosomally synthesized and posttranslationally modified peptides) is increasing worldwide. To facilitate the research in this field we have updated our mining web server. BAGEL4 is faster than its predecessor and is now fully independent from ORF-calling. Gene clusters of interest are discovered using the core-peptide database and/or through HMM motifs that are present in associated context genes. The databases used for mining have been updated and extended with literature references and links to UniProt and NCBI. Additionally, we have included automated promoter and terminator prediction and the option to upload RNA expression data, which can be displayed along with the identified clusters. Further improvements include the annotation of the context genes, which is now based on a fast blast against the prokaryote part of the UniRef90 database, and the improved web-BLAST feature that dynamically loads structural data such as internal cross-linking from UniProt. Overall BAGEL4 provides the user with more information through a user-friendly web-interface which simplifies data evaluation. BAGEL4 is freely accessible at http://bagel4.molgenrug.nl.

  18. Asymmetric threat data mining and knowledge discovery

    NASA Astrophysics Data System (ADS)

    Gilmore, John F.; Pagels, Michael A.; Palk, Justin

    2001-03-01

    Asymmetric threats differ from the conventional force-on- force military encounters that the Defense Department has historically been trained to engage. Terrorism by its nature is now an operational activity that is neither easily detected or countered as its very existence depends on small covert attacks exploiting the element of surprise. But terrorism does have defined forms, motivations, tactics and organizational structure. Exploiting a terrorism taxonomy provides the opportunity to discover and assess knowledge of terrorist operations. This paper describes the Asymmetric Threat Terrorist Assessment, Countering, and Knowledge (ATTACK) system. ATTACK has been developed to (a) data mine open source intelligence (OSINT) information from web-based newspaper sources, video news web casts, and actual terrorist web sites, (b) evaluate this information against a terrorism taxonomy, (c) exploit country/region specific social, economic, political, and religious knowledge, and (d) discover and predict potential terrorist activities and association links. Details of the asymmetric threat structure and the ATTACK system architecture are presented with results of an actual terrorist data mining and knowledge discovery test case shown.

  19. Informing child welfare policy and practice: using knowledge discovery and data mining technology via a dynamic Web site.

    PubMed

    Duncan, Dean F; Kum, Hye-Chung; Weigensberg, Elizabeth Caplick; Flair, Kimberly A; Stewart, C Joy

    2008-11-01

    Proper management and implementation of an effective child welfare agency requires the constant use of information about the experiences and outcomes of children involved in the system, emphasizing the need for comprehensive, timely, and accurate data. In the past 20 years, there have been many advances in technology that can maximize the potential of administrative data to promote better evaluation and management in the field of child welfare. Specifically, this article discusses the use of knowledge discovery and data mining (KDD), which makes it possible to create longitudinal data files from administrative data sources, extract valuable knowledge, and make the information available via a user-friendly public Web site. This article demonstrates a successful project in North Carolina where knowledge discovery and data mining technology was used to develop a comprehensive set of child welfare outcomes available through a public Web site to facilitate information sharing of child welfare data to improve policy and practice.

  20. Chemotext: A Publicly Available Web Server for Mining Drug-Target-Disease Relationships in PubMed.

    PubMed

    Capuzzi, Stephen J; Thornton, Thomas E; Liu, Kammy; Baker, Nancy; Lam, Wai In; O'Banion, Colin P; Muratov, Eugene N; Pozefsky, Diane; Tropsha, Alexander

    2018-02-26

    Elucidation of the mechanistic relationships between drugs, their targets, and diseases is at the core of modern drug discovery research. Thousands of studies relevant to the drug-target-disease (DTD) triangle have been published and annotated in the Medline/PubMed database. Mining this database affords rapid identification of all published studies that confirm connections between vertices of this triangle or enable new inferences of such connections. To this end, we describe the development of Chemotext, a publicly available Web server that mines the entire compendium of published literature in PubMed annotated by Medline Subject Heading (MeSH) terms. The goal of Chemotext is to identify all known DTD relationships and infer missing links between vertices of the DTD triangle. As a proof-of-concept, we show that Chemotext could be instrumental in generating new drug repurposing hypotheses or annotating clinical outcomes pathways for known drugs. The Chemotext Web server is freely available at http://chemotext.mml.unc.edu .

  1. Scientists as Communicators: Inclusion of a Science/Education Liaison on Research Expeditions

    NASA Astrophysics Data System (ADS)

    Sautter, L. R.

    2004-12-01

    Communication of research and scientific results to an audience outside of one's field poses a challenge to many scientists. Many research scientists have a natural ability to address the challenge, while others may chose to seek assistance. Research cruise PIs maywish to consider including a Science/Education Liaison (SEL) on future grants. The SEL is a marine scientist whose job before, during and after the cruise is to work with the shipboard scientists to document the science conducted. The SEL's role is three-fold: (1) to communicate shipboard science activities near-real-time to the public via the web; (2) to develop a variety of web-based resources based on the scientific operations; and (3) to assist educators with the integration of these resources into classroom curricula. The first role involves at-sea writing and relaying from ship-to-shore (via email) a series of Daily Logs. NOAA Ocean Exploration (OE) has mastered the use of web-posted Daily Logs for their major expeditions (see their OceanExplorer website), introducing millions of users to deep sea exploration. Project Oceanica uses the OE daily log model to document research expeditions. In addition to writing daily logs and participating on OE expeditions, Oceanica's SEL also documents the cruise's scientific operations and preliminary findings using video and photos, so that web-based resources (photo galleries, video galleries, and PhotoDocumentaries) can be developed during and following the cruise, and posted on the expedition's home page within the Oceanica web site (see URL). We have created templates for constructing these science resources which allow the shipboard scientists to assist with web resource development. Bringing users to the site is achieved through email communications to a growing list of educators, scientists, and students, and through collaboration with the COSEE network. With a large research expedition-based inventory of web resources now available, Oceanica is training teachers and college faculty on the use and incorporation of these resources into middle school, high school and introductory college classrooms. Support for a SEL on shipboard expeditions serves to catalyze the dissemination of the scientific operations to a broad audience of users.

  2. Harvesting data from advanced technologies.

    DOT National Transportation Integrated Search

    2014-11-01

    Data streams are emerging everywhere such as Web logs, Web page click streams, sensor data streams, and credit card transaction flows. : Different from traditional data sets, data streams are sequentially generated and arrive one by one rather than b...

  3. Data mining for personal navigation

    NASA Astrophysics Data System (ADS)

    Hariharan, Gurushyam; Franti, Pasi; Mehta, Sandeep

    2002-03-01

    Relevance is the key in defining what data is to be extracted from the Internet. Traditionally, relevance has been defined mainly by keywords and user profiles. In this paper we discuss a fairly untouched dimension to relevance: location. Any navigational information sought by a user at large on earth is evidently governed by his location. We believe that task oriented data mining of the web amalgamated with location information is the key to providing relevant information for personal navigation. We explore the existential hurdles and propose novel approaches to tackle them. We also present naive, task-oriented data mining based approaches and their implementations in Java, to extract location based information. Ad-hoc pairing of data with coordinates (x, y) is very rare on the web. But if the same co-ordinates are converted to a logical address (state/city/street), a wide spectrum of location-based information base opens up. Hence, given the coordinates (x, y) on the earth, the scheme points to the logical address of the user. Location based information could either be picked up from fixed and known service providers (e.g. Yellow Pages) or from any arbitrary website on the Web. Once the web servers providing information relevant to the logical address are located, task oriented data mining is performed over these sites keeping in mind what information is interesting to the contemporary user. After all this, a simple data stream is provided to the user with information scaled to his convenience. The scheme has been implemented for cities of Finland.

  4. Large-Scale Overlays and Trends: Visually Mining, Panning and Zooming the Observable Universe.

    PubMed

    Luciani, Timothy Basil; Cherinka, Brian; Oliphant, Daniel; Myers, Sean; Wood-Vasey, W Michael; Labrinidis, Alexandros; Marai, G Elisabeta

    2014-07-01

    We introduce a web-based computing infrastructure to assist the visual integration, mining and interactive navigation of large-scale astronomy observations. Following an analysis of the application domain, we design a client-server architecture to fetch distributed image data and to partition local data into a spatial index structure that allows prefix-matching of spatial objects. In conjunction with hardware-accelerated pixel-based overlays and an online cross-registration pipeline, this approach allows the fetching, displaying, panning and zooming of gigabit panoramas of the sky in real time. To further facilitate the integration and mining of spatial and non-spatial data, we introduce interactive trend images-compact visual representations for identifying outlier objects and for studying trends within large collections of spatial objects of a given class. In a demonstration, images from three sky surveys (SDSS, FIRST and simulated LSST results) are cross-registered and integrated as overlays, allowing cross-spectrum analysis of astronomy observations. Trend images are interactively generated from catalog data and used to visually mine astronomy observations of similar type. The front-end of the infrastructure uses the web technologies WebGL and HTML5 to enable cross-platform, web-based functionality. Our approach attains interactive rendering framerates; its power and flexibility enables it to serve the needs of the astronomy community. Evaluation on three case studies, as well as feedback from domain experts emphasize the benefits of this visual approach to the observational astronomy field; and its potential benefits to large scale geospatial visualization in general.

  5. A First Step in Learning Analytics: Pre-Processing Low-Level Alice Logging Data of Middle School Students

    ERIC Educational Resources Information Center

    Werner, Linda; McDowell, Charlie; Denner, Jill

    2013-01-01

    Educational data mining can miss or misidentify key findings about student learning without a transparent process of analyzing the data. This paper describes the first steps in the process of using low-level logging data to understand how middle school students used Alice, an initial programming environment. We describe the steps that were…

  6. Geophysical examination of coal deposits

    NASA Astrophysics Data System (ADS)

    Jackson, L. J.

    1981-04-01

    Geophysical techniques for the solution of mining problems and as an aid to mine planning are reviewed. Techniques of geophysical borehole logging are discussed. The responses of the coal seams to logging tools are easily recognized on the logging records. Cores for laboratory analysis are cut from selected sections of the borehole. In addition, information about the density and chemical composition of the coal may be obtained. Surface seismic reflection surveys using two dimensional arrays of seismic sources and detectors detect faults with throws as small as 3 m depths of 800 m. In geologically disturbed areas, good results have been obtained from three dimensional surveys. Smaller faults as far as 500 m in advance of the working face may be detected using in seam seismic surveying conducted from a roadway or working face. Small disturbances are detected by pulse radar and continuous wave electromagnetic methods either from within boreholes or from underground. Other geophysical techniques which explicit the electrical, magnetic, gravitational, and geothermal properties of rocks are described.

  7. A Generic Framework for Extraction of Knowledge from Social Web Sources (Social Networking Websites) for an Online Recommendation System

    ERIC Educational Resources Information Center

    Sathick, Javubar; Venkat, Jaya

    2015-01-01

    Mining social web data is a challenging task and finding user interest for personalized and non-personalized recommendation systems is another important task. Knowledge sharing among web users has become crucial in determining usage of web data and personalizing content in various social websites as per the user's wish. This paper aims to design a…

  8. Ground-water data for the Hanna and Carbon basins, south-central Wyoming, through 1980

    USGS Publications Warehouse

    Daddow, P.B.

    1986-01-01

    Groundwater resources in the Hanna and Carbon Basins of Wyoming were assessed in a study from 1974 through 1980 because of the development of coal mining in the area. Data collected from 105 wells during that study, including well-completion records, lithologic logs, and water levels, are presented. The data are from stock wells, coal-test holes completed as observation wells by the U.S. Geological Survey. The data are mostly from mined coal-bearing formations: the Tertiary Hanna Formation and the Tertiary and Cretaceous Ferris Formation. Well-completion data and lithologic logs were collected on-site during drilling of the wells or from U.S. Geological Survey files, company records, Wyoming State Engineer well-permit files, and published reports. (USGS)

  9. Mining Hidden Gems Beneath the Surface: A Look At the Invisible Web.

    ERIC Educational Resources Information Center

    Carlson, Randal D.; Repman, Judi

    2002-01-01

    Describes resources for researchers called the Invisible Web that are hidden from the usual search engines and other tools and contrasts them with those resources available on the surface Web. Identifies specialized search tools, databases, and strategies that can be used to locate credible in-depth information. (Author/LRW)

  10. Opinion Integration and Summarization

    ERIC Educational Resources Information Center

    Lu, Yue

    2011-01-01

    As Web 2.0 applications become increasingly popular, more and more people express their opinions on the Web in various ways in real time. Such wide coverage of topics and abundance of users make the Web an extremely valuable source for mining people's opinions about all kinds of topics. However, since the opinions are usually expressed as…

  11. Utilization of services in a randomized trial testing phone- and web-based interventions for smoking cessation.

    PubMed

    Zbikowski, Susan M; Jack, Lisa M; McClure, Jennifer B; Deprey, Mona; Javitz, Harold S; McAfee, Timothy A; Catz, Sheryl L; Richards, Julie; Bush, Terry; Swan, Gary E

    2011-05-01

    Phone counseling has become standard for behavioral smoking cessation treatment. Newer options include Web and integrated phone-Web treatment. No prior research, to our knowledge, has systematically compared the effectiveness of these three treatment modalities in a randomized trial. Understanding how utilization varies by mode, the impact of utilization on outcomes, and predictors of utilization across each mode could lead to improved treatments. One thousand two hundred and two participants were randomized to phone, Web, or combined phone-Web cessation treatment. Services varied by modality and were tracked using automated systems. All participants received 12 weeks of varenicline, printed guides, an orientation call, and access to a phone supportline. Self-report data were collected at baseline and 6-month follow-up. Overall, participants utilized phone services more often than the Web-based services. Among treatment groups with Web access, a significant proportion logged in only once (37% phone-Web, 41% Web), and those in the phone-Web group logged in less often than those in the Web group (mean = 2.4 vs. 3.7, p = .0001). Use of the phone also was correlated with increased use of the Web. In multivariate analyses, greater use of the phone- or Web-based services was associated with higher cessation rates. Finally, older age and the belief that certain treatments could improve success were consistent predictors of greater utilization across groups. Other predictors varied by treatment group. Opportunities for enhancing treatment utilization exist, particularly for Web-based programs. Increasing utilization more broadly could result in better overall treatment effectiveness for all intervention modalities.

  12. Global Connections: Web Conferencing Tools Help Educators Collaborate Anytime, Anywhere

    ERIC Educational Resources Information Center

    Forrester, Dave

    2009-01-01

    Web conferencing tools help educators from around the world collaborate in real time. Teachers, school counselors, and administrators need only to put on their headsets, check the time zone, and log on to meet and learn from educators across the globe. In this article, the author discusses how educators can use Web conferencing at their schools.…

  13. Web-Based Learning Programs: Use by Learners with Various Cognitive Styles

    ERIC Educational Resources Information Center

    Chen, Ling-Hsiu

    2010-01-01

    To consider how Web-based learning program is utilized by learners with different cognitive styles, this study presents a Web-based learning system (WBLS) and analyzes learners' browsing data recorded in the log file to identify how learners' cognitive styles and learning behavior are related. In order to develop an adapted WBLS, this study also…

  14. 78 FR 77706 - Notice of Intent To Prepare an Environmental Impact Statement for the Proposed Gemfield Mine...

    Federal Register 2010, 2011, 2012, 2013, 2014

    2013-12-24

    ... gold mine and associated processing and ancillary facilities. The project would be located on public... media, newspapers and the BLM Web site at: http://www.blm.gov/nv/st/en/fo/battle_mountain_field.html... to construct, operate, reclaim, and close an open pit, heap leach, gold mining operation known as the...

  15. QuadBase2: web server for multiplexed guanine quadruplex mining and visualization

    PubMed Central

    Dhapola, Parashar; Chowdhury, Shantanu

    2016-01-01

    DNA guanine quadruplexes or G4s are non-canonical DNA secondary structures which affect genomic processes like replication, transcription and recombination. G4s are computationally identified by specific nucleotide motifs which are also called putative G4 (PG4) motifs. Despite the general relevance of these structures, there is currently no tool available that can allow batch queries and genome-wide analysis of these motifs in a user-friendly interface. QuadBase2 (quadbase.igib.res.in) presents a completely reinvented web server version of previously published QuadBase database. QuadBase2 enables users to mine PG4 motifs in up to 178 eukaryotes through the EuQuad module. This module interfaces with Ensembl Compara database, to allow users mine PG4 motifs in the orthologues of genes of interest across eukaryotes. PG4 motifs can be mined across genes and their promoter sequences in 1719 prokaryotes through ProQuad module. This module includes a feature that allows genome-wide mining of PG4 motifs and their visualization as circular histograms. TetraplexFinder, the module for mining PG4 motifs in user-provided sequences is now capable of handling up to 20 MB of data. QuadBase2 is a comprehensive PG4 motif mining tool that further expands the configurations and algorithms for mining PG4 motifs in a user-friendly way. PMID:27185890

  16. Using ant-behavior-based simulation model AntWeb to improve website organization

    NASA Astrophysics Data System (ADS)

    Li, Weigang; Pinheiro Dib, Marcos V.; Teles, Wesley M.; Morais de Andrade, Vlaudemir; Alves de Melo, Alba C. M.; Cariolano, Judas T.

    2002-03-01

    Some web usage mining algorithms showed the potential application to find the difference among the organizations expected by visitors to the website. However, there are still no efficient method and criterion for a web administrator to measure the performance of the modification. In this paper, we developed an AntWeb, a model inspired by ants' behavior to simulate the sequence of visiting the website, in order to measure the efficient of the web structure. We implemented a web usage mining algorithm using backtrack to the intranet website of the Politec Informatic Ltd., Brazil. We defined throughput (the number of visitors to reach their target pages per time unit relates to the total number of visitors) as an index to measure the website's performance. We also used the link in a web page to represent the effect of visitors' pheromone trails. For every modification in the website organization, for example, putting a link from the expected location to the target object, the simulation reported the value of throughput as a quick answer about this modification. The experiment showed the stability of our simulation model, and a positive modification to the intranet website of the Politec.

  17. Mining of the social network extraction

    NASA Astrophysics Data System (ADS)

    Nasution, M. K. M.; Hardi, M.; Syah, R.

    2017-01-01

    The use of Web as social media is steadily gaining ground in the study of social actor behaviour. However, information in Web can be interpreted in accordance with the ability of the method such as superficial methods for extracting social networks. Each method however has features and drawbacks: it cannot reveal the behaviour of social actors, but it has the hidden information about them. Therefore, this paper aims to reveal such information in the social networks mining. Social behaviour could be expressed through a set of words extracted from the list of snippets.

  18. Analysing Customer Opinions with Text Mining Algorithms

    NASA Astrophysics Data System (ADS)

    Consoli, Domenico

    2009-08-01

    Knowing what the customer thinks of a particular product/service helps top management to introduce improvements in processes and products, thus differentiating the company from their competitors and gain competitive advantages. The customers, with their preferences, determine the success or failure of a company. In order to know opinions of the customers we can use technologies available from the web 2.0 (blog, wiki, forums, chat, social networking, social commerce). From these web sites, useful information must be extracted, for strategic purposes, using techniques of sentiment analysis or opinion mining.

  19. Evaluation of the Feasibility of Screening Patients for Early Signs of Lung Carcinoma in Web Search Logs.

    PubMed

    White, Ryen W; Horvitz, Eric

    2017-03-01

    A statistical model that predicts the appearance of strong evidence of a lung carcinoma diagnosis via analysis of large-scale anonymized logs of web search queries from millions of people across the United States. To evaluate the feasibility of screening patients at risk of lung carcinoma via analysis of signals from online search activity. We identified people who issue special queries that provide strong evidence of a recent diagnosis of lung carcinoma. We then considered patterns of symptoms expressed as searches about concerning symptoms over several months prior to the appearance of the landmark web queries. We built statistical classifiers that predict the future appearance of landmark queries based on the search log signals. This was a retrospective log analysis of the online activity of millions of web searchers seeking health-related information online. Of web searchers who queried for symptoms related to lung carcinoma, some (n = 5443 of 4 813 985) later issued queries that provide strong evidence of recent clinical diagnosis of lung carcinoma and are regarded as positive cases in our analysis. Additional evidence on the reliability of these queries as representing clinical diagnoses is based on the significant increase in follow-on searches for treatments and medications for these searchers and on the correlation between lung carcinoma incidence rates and our log-based statistics. The remaining symptom searchers (n = 4 808 542) are regarded as negative cases. Performance of the statistical model for early detection from online search behavior, for different lead times, different sets of signals, and different cohorts of searchers stratified by potential risk. The statistical classifier predicting the future appearance of landmark web queries based on search log signals identified searchers who later input queries consistent with a lung carcinoma diagnosis, with a true-positive rate ranging from 3% to 57% for false-positive rates ranging from 0.00001 to 0.001, respectively. The methods can be used to identify people at highest risk up to a year in advance of the inferred diagnosis time. The 5 factors associated with the highest relative risk (RR) were evidence of family history (RR = 7.548; 95% CI, 3.937-14.470), age (RR = 3.558; 95% CI, 3.357-3.772), radon (RR = 2.529; 95% CI, 1.137-5.624), primary location (RR = 2.463; 95% CI, 1.364-4.446), and occupation (RR = 1.969; 95% CI, 1.143-3.391). Evidence of smoking (RR = 1.646; 95% CI, 1.032-2.260) was important but not top-ranked, which was due to the difficulty of identifying smoking history from search terms. Pattern recognition based on data drawn from large-scale web search queries holds opportunity for identifying risk factors and frames new directions with early detection of lung carcinoma.

  20. A General Purpose Connections type CTI Server Based on SIP Protocol and Its Implementation

    NASA Astrophysics Data System (ADS)

    Watanabe, Toru; Koizumi, Hisao

    In this paper, we propose a general purpose connections type CTI (Computer Telephony Integration) server that provides various CTI services such as voice logging where the CTI server communicates with IP-PBX using the SIP (Session Initiation Protocol), and accumulates voice packets of external line telephone call flowing between an IP telephone for extension and a VoIP gateway connected to outside line networks. The CTI server realizes CTI services such as voice logging, telephone conference, or IVR (interactive voice response) with accumulating and processing voice packets sampled. Furthermore, the CTI server incorporates a web server function which can provide various CTI services such as a Web telephone directory via a Web browser to PCs, cellular telephones or smart-phones in mobile environments.

  1. WWW Motivation Mining: Finding Treasures for Teaching Evaluation Skills, Grades 1-6. Professional Growth Series.

    ERIC Educational Resources Information Center

    Arnone, Marilyn P.; Small, Ruth V.

    Designed for elementary or middle school teachers and library media specialists, this book provides educators with practical, easy-to-use ways of applying motivation assessment techniques when selecting World Wide Web sites for inclusion in their lessons and offers concrete examples of how to use Web evaluation with young learners. WebMAC…

  2. Astrophysical data mining with GPU. A case study: Genetic classification of globular clusters

    NASA Astrophysics Data System (ADS)

    Cavuoti, S.; Garofalo, M.; Brescia, M.; Paolillo, M.; Pescape', A.; Longo, G.; Ventre, G.

    2014-01-01

    We present a multi-purpose genetic algorithm, designed and implemented with GPGPU/CUDA parallel computing technology. The model was derived from our CPU serial implementation, named GAME (Genetic Algorithm Model Experiment). It was successfully tested and validated on the detection of candidate Globular Clusters in deep, wide-field, single band HST images. The GPU version of GAME will be made available to the community by integrating it into the web application DAMEWARE (DAta Mining Web Application REsource, http://dame.dsf.unina.it/beta_info.html), a public data mining service specialized on massive astrophysical data. Since genetic algorithms are inherently parallel, the GPGPU computing paradigm leads to a speedup of a factor of 200× in the training phase with respect to the CPU based version.

  3. Inquiry and Aquifers.

    ERIC Educational Resources Information Center

    Leuenberger, Ted; Shepardson, Daniel; Harbor, Jon; Bell, Cheryl; Meyer, Jason; Klagges, Hope; Burgess, Willie

    2001-01-01

    Presents inquiry-oriented activities that acquaint students with groundwater sources, movement of water through aquifers, and contamination of groundwater by pollution. In one activity, students use well log data from web-based resources to explore groundwater systems. Provides sample well log data for those not having access to local information.…

  4. Web services-based text-mining demonstrates broad impacts for interoperability and process simplification.

    PubMed

    Wiegers, Thomas C; Davis, Allan Peter; Mattingly, Carolyn J

    2014-01-01

    The Critical Assessment of Information Extraction systems in Biology (BioCreAtIvE) challenge evaluation tasks collectively represent a community-wide effort to evaluate a variety of text-mining and information extraction systems applied to the biological domain. The BioCreative IV Workshop included five independent subject areas, including Track 3, which focused on named-entity recognition (NER) for the Comparative Toxicogenomics Database (CTD; http://ctdbase.org). Previously, CTD had organized document ranking and NER-related tasks for the BioCreative Workshop 2012; a key finding of that effort was that interoperability and integration complexity were major impediments to the direct application of the systems to CTD's text-mining pipeline. This underscored a prevailing problem with software integration efforts. Major interoperability-related issues included lack of process modularity, operating system incompatibility, tool configuration complexity and lack of standardization of high-level inter-process communications. One approach to potentially mitigate interoperability and general integration issues is the use of Web services to abstract implementation details; rather than integrating NER tools directly, HTTP-based calls from CTD's asynchronous, batch-oriented text-mining pipeline could be made to remote NER Web services for recognition of specific biological terms using BioC (an emerging family of XML formats) for inter-process communications. To test this concept, participating groups developed Representational State Transfer /BioC-compliant Web services tailored to CTD's NER requirements. Participants were provided with a comprehensive set of training materials. CTD evaluated results obtained from the remote Web service-based URLs against a test data set of 510 manually curated scientific articles. Twelve groups participated in the challenge. Recall, precision, balanced F-scores and response times were calculated. Top balanced F-scores for gene, chemical and disease NER were 61, 74 and 51%, respectively. Response times ranged from fractions-of-a-second to over a minute per article. We present a description of the challenge and summary of results, demonstrating how curation groups can effectively use interoperable NER technologies to simplify text-mining pipeline implementation. Database URL: http://ctdbase.org/ © The Author(s) 2014. Published by Oxford University Press.

  5. Web services-based text-mining demonstrates broad impacts for interoperability and process simplification

    PubMed Central

    Wiegers, Thomas C.; Davis, Allan Peter; Mattingly, Carolyn J.

    2014-01-01

    The Critical Assessment of Information Extraction systems in Biology (BioCreAtIvE) challenge evaluation tasks collectively represent a community-wide effort to evaluate a variety of text-mining and information extraction systems applied to the biological domain. The BioCreative IV Workshop included five independent subject areas, including Track 3, which focused on named-entity recognition (NER) for the Comparative Toxicogenomics Database (CTD; http://ctdbase.org). Previously, CTD had organized document ranking and NER-related tasks for the BioCreative Workshop 2012; a key finding of that effort was that interoperability and integration complexity were major impediments to the direct application of the systems to CTD's text-mining pipeline. This underscored a prevailing problem with software integration efforts. Major interoperability-related issues included lack of process modularity, operating system incompatibility, tool configuration complexity and lack of standardization of high-level inter-process communications. One approach to potentially mitigate interoperability and general integration issues is the use of Web services to abstract implementation details; rather than integrating NER tools directly, HTTP-based calls from CTD's asynchronous, batch-oriented text-mining pipeline could be made to remote NER Web services for recognition of specific biological terms using BioC (an emerging family of XML formats) for inter-process communications. To test this concept, participating groups developed Representational State Transfer /BioC-compliant Web services tailored to CTD's NER requirements. Participants were provided with a comprehensive set of training materials. CTD evaluated results obtained from the remote Web service-based URLs against a test data set of 510 manually curated scientific articles. Twelve groups participated in the challenge. Recall, precision, balanced F-scores and response times were calculated. Top balanced F-scores for gene, chemical and disease NER were 61, 74 and 51%, respectively. Response times ranged from fractions-of-a-second to over a minute per article. We present a description of the challenge and summary of results, demonstrating how curation groups can effectively use interoperable NER technologies to simplify text-mining pipeline implementation. Database URL: http://ctdbase.org/ PMID:24919658

  6. Operating System Support for Shared Hardware Data Structures

    DTIC Science & Technology

    2013-01-31

    Carbon [73] uses hardware queues to improve fine-grained multitasking for Recognition, Mining , and Synthesis. Compared to software ap- proaches...web transaction processing, data mining , and multimedia. Early work in database processors [114, 96, 79, 111] reduce the costs of relational database...assignment can be solved statically or dynamically. Static assignment deter- mines offline which data structures are assigned to use HWDS resources and at

  7. 77 FR 4360 - Notice of Availability of the Draft Environmental Impact Statement for the Hycroft Mine Expansion...

    Federal Register 2010, 2011, 2012, 2013, 2014

    2012-01-27

    ... comments related to the Hycroft Mine Expansion Draft EIS by any of the following methods: Web site: www.blm..., Nevada 89445, Attn. Kathleen Rehberg. Copies of the Hycroft Mine Expansion Draft EIS are available in the... hours. The FIRS is available 24 hours a day, 7 days a week, to leave a message or question with the...

  8. Computer Cache. Online Recess--Web Games for Play and Fun

    ERIC Educational Resources Information Center

    Byerly, Greg; Brodie, Carolyn S.

    2005-01-01

    There are many age-appropriate, free, and easy-to-use online games available on the Web. In this column the authors describe some of their favorites for use with and by elementary students. They have not included games that require children to log on and/or register with their names or play against someone else interactively over the Web. None of…

  9. Browser-Based Online Applications: Something for Everyone!

    ERIC Educational Resources Information Center

    Descy, Don E.

    2007-01-01

    Just as many people log onto a Web mail site (Gmail, Yahoo, MSN, etc.) to read, write and store their email, there are Web sites out there with word processing, database, and a myriad of other software applications that are not downloadable but used on the site through a Web browser. The user does not have to download the applications to a…

  10. Utilization of Services in a Randomized Trial Testing Phone- and Web-Based Interventions for Smoking Cessation

    PubMed Central

    Jack, Lisa M.; McClure, Jennifer B.; Deprey, Mona; Javitz, Harold S.; McAfee, Timothy A.; Catz, Sheryl L.; Richards, Julie; Bush, Terry; Swan, Gary E.

    2011-01-01

    Introduction: Phone counseling has become standard for behavioral smoking cessation treatment. Newer options include Web and integrated phone–Web treatment. No prior research, to our knowledge, has systematically compared the effectiveness of these three treatment modalities in a randomized trial. Understanding how utilization varies by mode, the impact of utilization on outcomes, and predictors of utilization across each mode could lead to improved treatments. Methods: One thousand two hundred and two participants were randomized to phone, Web, or combined phone–Web cessation treatment. Services varied by modality and were tracked using automated systems. All participants received 12 weeks of varenicline, printed guides, an orientation call, and access to a phone supportline. Self-report data were collected at baseline and 6-month follow-up. Results: Overall, participants utilized phone services more often than the Web-based services. Among treatment groups with Web access, a significant proportion logged in only once (37% phone–Web, 41% Web), and those in the phone–Web group logged in less often than those in the Web group (mean = 2.4 vs. 3.7, p = .0001). Use of the phone also was correlated with increased use of the Web. In multivariate analyses, greater use of the phone- or Web-based services was associated with higher cessation rates. Finally, older age and the belief that certain treatments could improve success were consistent predictors of greater utilization across groups. Other predictors varied by treatment group. Conclusions: Opportunities for enhancing treatment utilization exist, particularly for Web-based programs. Increasing utilization more broadly could result in better overall treatment effectiveness for all intervention modalities. PMID:21330267

  11. Alkemio: association of chemicals with biomedical topics by text and data mining

    PubMed Central

    Gijón-Correas, José A.; Andrade-Navarro, Miguel A.; Fontaine, Jean F.

    2014-01-01

    The PubMed® database of biomedical citations allows the retrieval of scientific articles studying the function of chemicals in biology and medicine. Mining millions of available citations to search reported associations between chemicals and topics of interest would require substantial human time. We have implemented the Alkemio text mining web tool and SOAP web service to help in this task. The tool uses biomedical articles discussing chemicals (including drugs), predicts their relatedness to the query topic with a naïve Bayesian classifier and ranks all chemicals by P-values computed from random simulations. Benchmarks on seven human pathways showed good retrieval performance (areas under the receiver operating characteristic curves ranged from 73.6 to 94.5%). Comparison with existing tools to retrieve chemicals associated to eight diseases showed the higher precision and recall of Alkemio when considering the top 10 candidate chemicals. Alkemio is a high performing web tool ranking chemicals for any biomedical topics and it is free to non-commercial users. Availability: http://cbdm.mdc-berlin.de/∼medlineranker/cms/alkemio. PMID:24838570

  12. Geovisualization of Local and Regional Migration Using Web-mined Demographics

    NASA Astrophysics Data System (ADS)

    Schuermann, R. T.; Chow, T. E.

    2014-11-01

    The intent of this research was to augment and facilitate analyses, which gauges the feasibility of web-mined demographics to study spatio-temporal dynamics of migration. As a case study, we explored the spatio-temporal dynamics of Vietnamese Americans (VA) in Texas through geovisualization of mined demographic microdata from the World Wide Web. Based on string matching across all demographic attributes, including full name, address, date of birth, age and phone number, multiple records of the same entity (i.e. person) over time were resolved and reconciled into a database. Migration trajectories were geovisualized through animated sprites by connecting the different addresses associated with the same person and segmenting the trajectory into small fragments. Intra-metropolitan migration patterns appeared at the local scale within many metropolitan areas. At the scale of metropolitan area, varying degrees of immigration and emigration manifest different types of migration clusters. This paper presents a methodology incorporating GIS methods and cartographic design to produce geovisualization animation, enabling the cognitive identification of migration patterns at multiple scales. Identification of spatio-temporal patterns often stimulates further research to better understand the phenomenon and enhance subsequent modeling.

  13. Logging on to Learn

    ERIC Educational Resources Information Center

    Butler, Kevin

    2010-01-01

    A classroom lecture at Capistrano Connections Academy in Southern California involves booting up the home computer, logging on to a Web site, and observing a teacher conducting a PowerPoint presentation of that day's lesson entirely online. Through microphone headsets, students can watch on their home computers, respond to the teacher's questions,…

  14. Advanced Cyber Attack Modeling Analysis and Visualization

    DTIC Science & Technology

    2010-03-01

    Graph Analysis Network Web Logs Netflow Data TCP Dump Data System Logs Detect Protect Security Management What-If Figure 8. TVA attack graphs for...Clustered Graphs,” in Proceedings of the Symposium on Graph Drawing, September 1996. [25] K. Lakkaraju, W. Yurcik, A. Lee, “NVisionIP: NetFlow

  15. Integration of Geographical Information Systems and Geophysical Applications with Distributed Computing Technologies.

    NASA Astrophysics Data System (ADS)

    Pierce, M. E.; Aktas, M. S.; Aydin, G.; Fox, G. C.; Gadgil, H.; Sayar, A.

    2005-12-01

    We examine the application of Web Service Architectures and Grid-based distributed computing technologies to geophysics and geo-informatics. We are particularly interested in the integration of Geographical Information System (GIS) services with distributed data mining applications. GIS services provide the general purpose framework for building archival data services, real time streaming data services, and map-based visualization services that may be integrated with data mining and other applications through the use of distributed messaging systems and Web Service orchestration tools. Building upon on our previous work in these areas, we present our current research efforts. These include fundamental investigations into increasing XML-based Web service performance, supporting real time data streams, and integrating GIS mapping tools with audio/video collaboration systems for shared display and annotation.

  16. Proceedings of GeoTech 85: Personal computers in geology conference

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Not Available

    1985-01-01

    This book presents the papers given at a conference which considered the use of microprocessors in the exploration of petroleum and natural gas deposits. Topics covered at the conference included seismic surveys, geochemistry, expert systems, artificial intelligence, data base management systems, a portable exploration work station, open pit planning on a microcomputer, well logging, fracture analysis, production scheduling of open pit mines, resistivity logging, and coal washability.

  17. Mass Storage Performance Information System

    NASA Technical Reports Server (NTRS)

    Scheuermann, Peter

    2000-01-01

    The purpose of this task is to develop a data warehouse to enable system administrators and their managers to gather information by querying the data logs of the MDSDS. Currently detailed logs capture the activity of the MDSDS internal to the different systems. The elements to be included in the data warehouse are requirements analysis, data cleansing, database design, database population, hardware/software acquisition, data transformation, query and report generation, and data mining.

  18. The equivalency between logic Petri workflow nets and workflow nets.

    PubMed

    Wang, Jing; Yu, ShuXia; Du, YuYue

    2015-01-01

    Logic Petri nets (LPNs) can describe and analyze batch processing functions and passing value indeterminacy in cooperative systems. Logic Petri workflow nets (LPWNs) are proposed based on LPNs in this paper. Process mining is regarded as an important bridge between modeling and analysis of data mining and business process. Workflow nets (WF-nets) are the extension to Petri nets (PNs), and have successfully been used to process mining. Some shortcomings cannot be avoided in process mining, such as duplicate tasks, invisible tasks, and the noise of logs. The online shop in electronic commerce in this paper is modeled to prove the equivalence between LPWNs and WF-nets, and advantages of LPWNs are presented.

  19. The Equivalency between Logic Petri Workflow Nets and Workflow Nets

    PubMed Central

    Wang, Jing; Yu, ShuXia; Du, YuYue

    2015-01-01

    Logic Petri nets (LPNs) can describe and analyze batch processing functions and passing value indeterminacy in cooperative systems. Logic Petri workflow nets (LPWNs) are proposed based on LPNs in this paper. Process mining is regarded as an important bridge between modeling and analysis of data mining and business process. Workflow nets (WF-nets) are the extension to Petri nets (PNs), and have successfully been used to process mining. Some shortcomings cannot be avoided in process mining, such as duplicate tasks, invisible tasks, and the noise of logs. The online shop in electronic commerce in this paper is modeled to prove the equivalence between LPWNs and WF-nets, and advantages of LPWNs are presented. PMID:25821845

  20. Development, installation, and testing services for an automatic, point type thermal sensor, fire protection system on a mining dozer. Final report

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Lease, W.D.

    1976-08-01

    Lease AFEX, Inc., modified its standard design of an automatic fire protection system used in the past on logging equipment, and long-term, in-mine tested system on a Fiat-Alli's HD-41B dozer at the Lemmons and Company coal mine, Boonville, Ind. The modification of the standard AFEX system involved improving the actuation device. The AFEX system is called a point-type thermal sensor, automatic fire protection system. The in-mine test took place in late 1975, and early 1976. The system was then tested by simulating a fire on the dozer. The system operated successfully after the 4 months of in-mine endurance testing. (Colormore » illustrations reproduced in black and white.)« less

  1. Completion reports, core logs, and hydrogeologic data from wells and piezometers in Prospect Gulch, San Juan County, Colorado

    USGS Publications Warehouse

    Johnson, Raymond H.; Yager, Douglas B.

    2006-01-01

    In the late nineteenth century, San Juan County, Colorado, was the center of a metal mining boom in the San Juan Mountains. Although most mining activity ceased by the 1990s, the effects of historical mining continue to contribute metals to ground water and surface water. Previous research by the U.S. Geological Survey identified ground-water discharge as a significant pathway for the loading of metals to surface water from both acid-mine drainage and acid-rock drainage. In an effort to understand the ground-water flow system in the upper Animas River watershed, Prospect Gulch was selected for further study because of the amount of previous data provided in and around that particular watershed. In support of this ground-water research effort, wells and piezometers were installed to allow for coring during installation, subsurface hydrologic testing, and the monitoring of ground-water hydraulic heads and geochemistry. This report summarizes the data that were collected during and after the installation of these wells and piezometers and includes (1) subsurface completion details, (2) locations and elevations, (3) geologic logs and elemental data, (4) slug test data for the estimation of subsurface hydraulic conductives, and (5) hydraulic head data.

  2. Literature Mining Methods for Toxicology and Construction of ...

    EPA Pesticide Factsheets

    Webinar Presentation on text-mining methodologies in use at NCCT and how they can be used to assist with the OECD Retinoid project. Presentation to 1st Workshop/Scientific Expert Group meeting on the OECD Retinoid Project - April 26, 2016 –Brussels, Presented remotely via web.

  3. Reference Architecture for MNE 5 Technical System

    DTIC Science & Technology

    2007-05-30

    of being available in most experiments. Core Services A core set of applications whi directories, web portal and collaboration applications etc. A...classifications Messages (xml, JMS, content level…) Meta data filtering, who can initiate services Web browsing Collaboration & messaging Border...Exchange Ref Architecture for MNE5 Tech System.doc 9 of 21 audit logging Person and machine Data lev objects, web services, messages rification el

  4. Monitoring food safety violation reports from internet forums.

    PubMed

    Kate, Kiran; Negi, Sumit; Kalagnanam, Jayant

    2014-01-01

    Food-borne illness is a growing public health concern in the world. Government bodies, which regulate and monitor the state of food safety, solicit citizen feedback about food hygiene practices followed by food establishments. They use traditional channels like call center, e-mail for such feedback collection. With the growing popularity of Web 2.0 and social media, citizens often post such feedback on internet forums, message boards etc. The system proposed in this paper applies text mining techniques to identify and mine such food safety complaints posted by citizens on web data sources thereby enabling the government agencies to gather more information about the state of food safety. In this paper, we discuss the architecture of our system and the text mining methods used. We also present results which demonstrate the effectiveness of this system in a real-world deployment.

  5. TCGA4U: A Web-Based Genomic Analysis Platform To Explore And Mine TCGA Genomic Data For Translational Research.

    PubMed

    Huang, Zhenzhen; Duan, Huilong; Li, Haomin

    2015-01-01

    Large-scale human cancer genomics projects, such as TCGA, generated large genomics data for further study. Exploring and mining these data to obtain meaningful analysis results can help researchers find potential genomics alterations that intervene the development and metastasis of tumors. We developed a web-based gene analysis platform, named TCGA4U, which used statistics methods and models to help translational investigators explore, mine and visualize human cancer genomic characteristic information from the TCGA datasets. Furthermore, through Gene Ontology (GO) annotation and clinical data integration, the genomic data were transformed into biological process, molecular function, cellular component and survival curves to help researchers identify potential driver genes. Clinical researchers without expertise in data analysis will benefit from such a user-friendly genomic analysis platform.

  6. Sexual information seeking on web search engines.

    PubMed

    Spink, Amanda; Koricich, Andrew; Jansen, B J; Cole, Charles

    2004-02-01

    Sexual information seeking is an important element within human information behavior. Seeking sexually related information on the Internet takes many forms and channels, including chat rooms discussions, accessing Websites or searching Web search engines for sexual materials. The study of sexual Web queries provides insight into sexually-related information-seeking behavior, of value to Web users and providers alike. We qualitatively analyzed queries from logs of 1,025,910 Alta Vista and AlltheWeb.com Web user queries from 2001. We compared the differences in sexually-related Web searching between Alta Vista and AlltheWeb.com users. Differences were found in session duration, query outcomes, and search term choices. Implications of the findings for sexual information seeking are discussed.

  7. DOE Office of Scientific and Technical Information (OSTI.GOV)

    Kargupta, H.; Stafford, B.; Hamzaoglu, I.

    This paper describes an experimental parallel/distributed data mining system PADMA (PArallel Data Mining Agents) that uses software agents for local data accessing and analysis and a web based interface for interactive data visualization. It also presents the results of applying PADMA for detecting patterns in unstructured texts of postmortem reports and laboratory test data for Hepatitis C patients.

  8. Data warehousing as a basis for web-based documentation of data mining and analysis.

    PubMed

    Karlsson, J; Eklund, P; Hallgren, C G; Sjödin, J G

    1999-01-01

    In this paper we present a case study for data warehousing intended to support data mining and analysis. We also describe a prototype for data retrieval. Further we discuss some technical issues related to a particular choice of a patient record environment.

  9. On-Board Mining in the Sensor Web

    NASA Astrophysics Data System (ADS)

    Tanner, S.; Conover, H.; Graves, S.; Ramachandran, R.; Rushing, J.

    2004-12-01

    On-board data mining can contribute to many research and engineering applications, including natural hazard detection and prediction, intelligent sensor control, and the generation of customized data products for direct distribution to users. The ability to mine sensor data in real time can also be a critical component of autonomous operations, supporting deep space missions, unmanned aerial and ground-based vehicles (UAVs, UGVs), and a wide range of sensor meshes, webs and grids. On-board processing is expected to play a significant role in the next generation of NASA, Homeland Security, Department of Defense and civilian programs, providing for greater flexibility and versatility in measurements of physical systems. In addition, the use of UAV and UGV systems is increasing in military, emergency response and industrial applications. As research into the autonomy of these vehicles progresses, especially in fleet or web configurations, the applicability of on-board data mining is expected to increase significantly. Data mining in real time on board sensor platforms presents unique challenges. Most notably, the data to be mined is a continuous stream, rather than a fixed store such as a database. This means that the data mining algorithms must be modified to make only a single pass through the data. In addition, the on-board environment requires real time processing with limited computing resources, thus the algorithms must use fixed and relatively small amounts of processing time and memory. The University of Alabama in Huntsville is developing an innovative processing framework for the on-board data and information environment. The Environment for On-Board Processing (EVE) and the Adaptive On-board Data Processing (AODP) projects serve as proofs-of-concept of advanced information systems for remote sensing platforms. The EVE real-time processing infrastructure will upload, schedule and control the execution of processing plans on board remote sensors. These plans provide capabilities for autonomous data mining, classification and feature extraction using both streaming and buffered data sources. A ground-based testbed provides a heterogeneous, embedded hardware and software environment representing both space-based and ground-based sensor platforms, including wireless sensor mesh architectures. The AODP project explores the EVE concepts in the world of sensor-networks, including ad-hoc networks of small sensor platforms.

  10. Designing and Managing Your Digital Library.

    ERIC Educational Resources Information Center

    Guenther, Kim

    2000-01-01

    Discusses digital libraries and Web site design issues. Highlights include accessibility issues, including standards, markup languages like HTML and XML, and metadata; building virtual communities; the use of Web portals for customized delivery of information; quality assurance tools, including data mining; and determining user needs, including…

  11. A Dynamic Recommender System for Improved Web Usage Mining and CRM Using Swarm Intelligence.

    PubMed

    Alphy, Anna; Prabakaran, S

    2015-01-01

    In modern days, to enrich e-business, the websites are personalized for each user by understanding their interests and behavior. The main challenges of online usage data are information overload and their dynamic nature. In this paper, to address these issues, a WebBluegillRecom-annealing dynamic recommender system that uses web usage mining techniques in tandem with software agents developed for providing dynamic recommendations to users that can be used for customizing a website is proposed. The proposed WebBluegillRecom-annealing dynamic recommender uses swarm intelligence from the foraging behavior of a bluegill fish. It overcomes the information overload by handling dynamic behaviors of users. Our dynamic recommender system was compared against traditional collaborative filtering systems. The results show that the proposed system has higher precision, coverage, F1 measure, and scalability than the traditional collaborative filtering systems. Moreover, the recommendations given by our system overcome the overspecialization problem by including variety in recommendations.

  12. A Dynamic Recommender System for Improved Web Usage Mining and CRM Using Swarm Intelligence

    PubMed Central

    Alphy, Anna; Prabakaran, S.

    2015-01-01

    In modern days, to enrich e-business, the websites are personalized for each user by understanding their interests and behavior. The main challenges of online usage data are information overload and their dynamic nature. In this paper, to address these issues, a WebBluegillRecom-annealing dynamic recommender system that uses web usage mining techniques in tandem with software agents developed for providing dynamic recommendations to users that can be used for customizing a website is proposed. The proposed WebBluegillRecom-annealing dynamic recommender uses swarm intelligence from the foraging behavior of a bluegill fish. It overcomes the information overload by handling dynamic behaviors of users. Our dynamic recommender system was compared against traditional collaborative filtering systems. The results show that the proposed system has higher precision, coverage, F1 measure, and scalability than the traditional collaborative filtering systems. Moreover, the recommendations given by our system overcome the overspecialization problem by including variety in recommendations. PMID:26229978

  13. The Islamic State Battle Plan: Press Release Natural Language Processing

    DTIC Science & Technology

    2016-06-01

    Processing, text mining , corpus, generalized linear model, cascade, R Shiny, leaflet, data visualization 15. NUMBER OF PAGES 83 16. PRICE CODE...Terrorism and Responses to Terrorism TDM Term Document Matrix TF Term Frequency TF-IDF Term Frequency-Inverse Document Frequency tm text mining (R...package=leaflet. Feinerer I, Hornik K (2015) Text Mining Package “tm,” Version 0.6-2. (Jul 3) https://cran.r-project.org/web/packages/tm/tm.pdf

  14. Numerical linear algebra in data mining

    NASA Astrophysics Data System (ADS)

    Eldén, Lars

    Ideas and algorithms from numerical linear algebra are important in several areas of data mining. We give an overview of linear algebra methods in text mining (information retrieval), pattern recognition (classification of handwritten digits), and PageRank computations for web search engines. The emphasis is on rank reduction as a method of extracting information from a data matrix, low-rank approximation of matrices using the singular value decomposition and clustering, and on eigenvalue methods for network analysis.

  15. ICCE/ICCAI 2000 Full & Short Papers (Methodologies).

    ERIC Educational Resources Information Center

    2000

    This document contains the full text of the following full and short papers on methodologies from ICCE/ICCAI 2000 (International Conference on Computers in Education/International Conference on Computer-Assisted Instruction): (1) "A Methodology for Learning Pattern Analysis from Web Logs by Interpreting Web Page Contents" (Chih-Kai Chang and…

  16. Teaching Geography in the Blogosphere

    ERIC Educational Resources Information Center

    Helmer, Joel W.; Bloch, Nick

    2010-01-01

    Over the last decade blog (a contraction of "web log") has crept into the national lexicon and culture, transforming the way people communicate. Although exact numbers are difficult to gather, the Web site "Technorati" is tracking more than 112.8 million blogs. Rarely far removed from technology; cell phones, iPods, Facebook…

  17. Library Web Proxy Use Survey Results.

    ERIC Educational Resources Information Center

    Murray, Peter E.

    2001-01-01

    Outlines the use of proxy Web servers by libraries and reports on a survey on their use in libraries. Highlights include proxy use for remote resource access, for filtering, for bandwidth conservation, and for gathering statistics; privacy policies regarding the use of proxy server log files; and a copy of the survey. (LRW)

  18. Participants, Usage, and Use Patterns of a Web-Based Intervention for the Prevention of Depression Within a Randomized Controlled Trial

    PubMed Central

    Bohlmeijer, Ernst T; Van Gemert-Pijnen, Julia EWC

    2013-01-01

    Background Although Web-based interventions have been shown to be effective, they are not widely implemented in regular care. Nonadherence (ie, participants not following the intervention protocol) is an issue. By studying the way Web-based interventions are used and whether there are differences between adherers (ie, participants that started all 9 lessons) and nonadherers, more insight can be gained into the process of adherence. Objective The aims of this study were to (1) describe the characteristics of participants and investigate their relationship with adherence, (2) investigate the utilization of the different features of the intervention and possible differences between adherers and nonadherers, and (3) identify what use patterns emerge and whether there are differences between adherers and nonadherers. Methods Data were used from 206 participants that used the Web-based intervention Living to the full, a Web-based intervention for the prevention of depression employing both a fully automated and human-supported format. Demographic and baseline characteristics of participants were collected by using an online survey. Log data were collected within the Web-based intervention itself. Both quantitative and qualitative analyses were performed. Results In all, 118 participants fully adhered to the intervention (ie, started all 9 lessons). Participants with an ethnicity other than Dutch were more often adherers (χ2 1=5.5, P=.02), and nonadherers used the Internet more hours per day on average (F1,203=3.918, P=.049). A logistic regression showed that being female (OR 2.02, 95% CI 1.01-4.04; P=.046) and having a higher need for cognition (OR 1.02; 95% CI 1.00-1.05; P=.02) increased the odds of adhering to the intervention. Overall, participants logged in an average of 4 times per lesson, but adherers logged in significantly more times per lesson than nonadherers (F1,204=20.710; P<.001). For use patterns, we saw that early nonadherers seemed to use fewer sessions and spend less time than late nonadherers and adherers, and fewer sessions to complete the lesson than adherers. Furthermore, late nonadherers seemed to have a shorter total duration of sessions than adherers. Conclusions By using log data combined with baseline characteristics of participants, we extracted valuable lessons for redesign of this intervention and the design of Web-based interventions in general. First, although characteristics of respondents can significantly predict adherence, their predictive value is small. Second, it is important to design Web-based interventions to foster adherence and usage of all features in an intervention. Trial Registration Dutch Trial Register Number: NTR3007; http://www.trialregister.nl/trialreg/admin/rctview.asp?TC=3007 (Archived by WebCite at http://www.webcitation.org/6ILhI3rd8). PMID:23963284

  19. Participants, usage, and use patterns of a web-based intervention for the prevention of depression within a randomized controlled trial.

    PubMed

    Kelders, Saskia M; Bohlmeijer, Ernst T; Van Gemert-Pijnen, Julia Ewc

    2013-08-20

    Although Web-based interventions have been shown to be effective, they are not widely implemented in regular care. Nonadherence (ie, participants not following the intervention protocol) is an issue. By studying the way Web-based interventions are used and whether there are differences between adherers (ie, participants that started all 9 lessons) and nonadherers, more insight can be gained into the process of adherence. The aims of this study were to (1) describe the characteristics of participants and investigate their relationship with adherence, (2) investigate the utilization of the different features of the intervention and possible differences between adherers and nonadherers, and (3) identify what use patterns emerge and whether there are differences between adherers and nonadherers. Data were used from 206 participants that used the Web-based intervention Living to the full, a Web-based intervention for the prevention of depression employing both a fully automated and human-supported format. Demographic and baseline characteristics of participants were collected by using an online survey. Log data were collected within the Web-based intervention itself. Both quantitative and qualitative analyses were performed. In all, 118 participants fully adhered to the intervention (ie, started all 9 lessons). Participants with an ethnicity other than Dutch were more often adherers (χ²₁=5.5, P=.02), and nonadherers used the Internet more hours per day on average (F₁,₂₀₃=3.918, P=.049). A logistic regression showed that being female (OR 2.02, 95% CI 1.01-4.04; P=.046) and having a higher need for cognition (OR 1.02; 95% CI 1.00-1.05; P=.02) increased the odds of adhering to the intervention. Overall, participants logged in an average of 4 times per lesson, but adherers logged in significantly more times per lesson than nonadherers (F₁,₂₀₄=20.710; P<.001). For use patterns, we saw that early nonadherers seemed to use fewer sessions and spend less time than late nonadherers and adherers, and fewer sessions to complete the lesson than adherers. Furthermore, late nonadherers seemed to have a shorter total duration of sessions than adherers. By using log data combined with baseline characteristics of participants, we extracted valuable lessons for redesign of this intervention and the design of Web-based interventions in general. First, although characteristics of respondents can significantly predict adherence, their predictive value is small. Second, it is important to design Web-based interventions to foster adherence and usage of all features in an intervention. Dutch Trial Register Number: NTR3007; http://www.trialregister.nl/trialreg/admin/rctview.asp?TC=3007 (Archived by WebCite at http://www.webcitation.org/6ILhI3rd8).

  20. DOE Office of Scientific and Technical Information (OSTI.GOV)

    Hodgson, H. E.

    The 1977 Symposium on the Geology of Rocky Mountain Coal was held May 9 and 10 on the campus of the Colorado School of Mines in Golden, Colorado. The 1977 Symposium was sponsored by the Colorado Geological Survey and the US Geological Survey. The 1977 Symposium consisted of four technical sessions: Depositional Models for Coal Exploration in the Rocky Mountain Cretaceous; Stratigraphy and Depositional Environments of Rocky Mountain Tertiary Coal Deposits; Depositional Models for Coal Exploration in non-Rocky Mountain Regions; and Application of Geology to Coal Mining and Coal Mine Planning. Several papers discuss geophysical survey and well logging techniquesmore » applied to the exploration of coal deposits and for mine planning. Fouteen papers have been entered individually into EDB and ERA. (LTN)« less

  1. Web-based versus traditional paper questionnaires: a mixed-mode survey with a Nordic perspective.

    PubMed

    Hohwü, Lena; Lyshol, Heidi; Gissler, Mika; Jonsson, Stefan Hrafn; Petzold, Max; Obel, Carsten

    2013-08-26

    Survey response rates have been declining over the past decade. The more widespread use of the Internet and Web-based technologies among potential health survey participants suggests that Web-based questionnaires may be an alternative to paper questionnaires in future epidemiological studies. To compare response rates in a population of parents by using 4 different modes of data collection for a questionnaire survey of which 1 involved a nonmonetary incentive. A random sample of 3148 parents of Danish children aged 2-17 years were invited to participate in the Danish part of the NordChild 2011 survey on their children's health and welfare. NordChild was conducted in 1984 and 1996 in collaboration with Finland, Iceland, Norway, and Sweden using mailed paper questionnaires only. In 2011, all countries used conventional paper versions only except Denmark where the parents were randomized into 4 groups: (1) 789 received a paper questionnaire only (paper), (2) 786 received the paper questionnaire and a log-in code to the Web-based questionnaire (paper/Web), (3) 787 received a log-in code to the Web-based questionnaire (Web), and (4) 786 received log-in details to the Web-based questionnaire and were given an incentive consisting of a chance to win a tablet computer (Web/tablet). In connection with the first reminder, the nonresponders in the paper, paper/Web, and Web groups were also present with the opportunity to win a tablet computer as a means of motivation. Descriptive analysis was performed using chi-square tests. Odds ratios were used to estimate differences in response rates between the 4 modes. In 2011, 1704 of 3148 (54.13%) respondents answered the Danish questionnaire. The highest response rate was with the paper mode (n=443, 56.2%). The other groups had similar response rates: paper/Web (n=422, 53.7%), Web (n=420, 53.4%), and Web/tablet (n=419, 53.3%) modes. Compared to the paper mode, the odds for response rate in the paper/Web decreased by 9% (OR 0.91, 95% CI 0.74-1.10) and by 11% (OR 0.89, 95% CI 0.73-1.09) in the Web and Web/tablet modes. The total number of responders for NordChild declined from 10,291 of 15,339 (67.09%) in 1984 and 10,667 of 15,254 (69.93%) in 1996 to 7805 of 15,945 (48.95%) in 2011 with similar declines in all 5 Nordic countries. Web-based questionnaires could replace traditional paper questionnaires with minor effects on response rates and lower costs. The increasing effect on the response rate on participants replying for a nonmonetary incentive could only be estimated within the 2 Web-based questionnaire modes before the first reminder. Alternative platforms to reach higher participation rates in population surveys should reflect the development of electronic devices and the ways in which the population primarily accesses the Internet.

  2. Web-Based Versus Traditional Paper Questionnaires: A Mixed-Mode Survey With a Nordic Perspective

    PubMed Central

    Lyshol, Heidi; Gissler, Mika; Jonsson, Stefan Hrafn; Petzold, Max; Obel, Carsten

    2013-01-01

    Background Survey response rates have been declining over the past decade. The more widespread use of the Internet and Web-based technologies among potential health survey participants suggests that Web-based questionnaires may be an alternative to paper questionnaires in future epidemiological studies. Objective To compare response rates in a population of parents by using 4 different modes of data collection for a questionnaire survey of which 1 involved a nonmonetary incentive. Methods A random sample of 3148 parents of Danish children aged 2-17 years were invited to participate in the Danish part of the NordChild 2011 survey on their children’s health and welfare. NordChild was conducted in 1984 and 1996 in collaboration with Finland, Iceland, Norway, and Sweden using mailed paper questionnaires only. In 2011, all countries used conventional paper versions only except Denmark where the parents were randomized into 4 groups: (1) 789 received a paper questionnaire only (paper), (2) 786 received the paper questionnaire and a log-in code to the Web-based questionnaire (paper/Web), (3) 787 received a log-in code to the Web-based questionnaire (Web), and (4) 786 received log-in details to the Web-based questionnaire and were given an incentive consisting of a chance to win a tablet computer (Web/tablet). In connection with the first reminder, the nonresponders in the paper, paper/Web, and Web groups were also present with the opportunity to win a tablet computer as a means of motivation. Descriptive analysis was performed using chi-square tests. Odds ratios were used to estimate differences in response rates between the 4 modes. Results In 2011, 1704 of 3148 (54.13%) respondents answered the Danish questionnaire. The highest response rate was with the paper mode (n=443, 56.2%). The other groups had similar response rates: paper/Web (n=422, 53.7%), Web (n=420, 53.4%), and Web/tablet (n=419, 53.3%) modes. Compared to the paper mode, the odds for response rate in the paper/Web decreased by 9% (OR 0.91, 95% CI 0.74-1.10) and by 11% (OR 0.89, 95% CI 0.73-1.09) in the Web and Web/tablet modes. The total number of responders for NordChild declined from 10,291 of 15,339 (67.09%) in 1984 and 10,667 of 15,254 (69.93%) in 1996 to 7805 of 15,945 (48.95%) in 2011 with similar declines in all 5 Nordic countries. Conclusions Web-based questionnaires could replace traditional paper questionnaires with minor effects on response rates and lower costs. The increasing effect on the response rate on participants replying for a nonmonetary incentive could only be estimated within the 2 Web-based questionnaire modes before the first reminder. Alternative platforms to reach higher participation rates in population surveys should reflect the development of electronic devices and the ways in which the population primarily accesses the Internet. PMID:23978658

  3. HC StratoMineR: A Web-Based Tool for the Rapid Analysis of High-Content Datasets.

    PubMed

    Omta, Wienand A; van Heesbeen, Roy G; Pagliero, Romina J; van der Velden, Lieke M; Lelieveld, Daphne; Nellen, Mehdi; Kramer, Maik; Yeong, Marley; Saeidi, Amir M; Medema, Rene H; Spruit, Marco; Brinkkemper, Sjaak; Klumperman, Judith; Egan, David A

    2016-10-01

    High-content screening (HCS) can generate large multidimensional datasets and when aligned with the appropriate data mining tools, it can yield valuable insights into the mechanism of action of bioactive molecules. However, easy-to-use data mining tools are not widely available, with the result that these datasets are frequently underutilized. Here, we present HC StratoMineR, a web-based tool for high-content data analysis. It is a decision-supportive platform that guides even non-expert users through a high-content data analysis workflow. HC StratoMineR is built by using My Structured Query Language for storage and querying, PHP: Hypertext Preprocessor as the main programming language, and jQuery for additional user interface functionality. R is used for statistical calculations, logic and data visualizations. Furthermore, C++ and graphical processor unit power is diffusely embedded in R by using the rcpp and rpud libraries for operations that are computationally highly intensive. We show that we can use HC StratoMineR for the analysis of multivariate data from a high-content siRNA knock-down screen and a small-molecule screen. It can be used to rapidly filter out undesirable data; to select relevant data; and to perform quality control, data reduction, data exploration, morphological hit picking, and data clustering. Our results demonstrate that HC StratoMineR can be used to functionally categorize HCS hits and, thus, provide valuable information for hit prioritization.

  4. Soil food web changes during spontaneous succession at post mining sites: a possible ecosystem engineering effect on food web organization?

    PubMed

    Frouz, Jan; Thébault, Elisa; Pižl, Václav; Adl, Sina; Cajthaml, Tomáš; Baldrián, Petr; Háněl, Ladislav; Starý, Josef; Tajovský, Karel; Materna, Jan; Nováková, Alena; de Ruiter, Peter C

    2013-01-01

    Parameters characterizing the structure of the decomposer food web, biomass of the soil microflora (bacteria and fungi) and soil micro-, meso- and macrofauna were studied at 14 non-reclaimed 1- 41-year-old post-mining sites near the town of Sokolov (Czech Republic). These observations on the decomposer food webs were compared with knowledge of vegetation and soil microstructure development from previous studies. The amount of carbon entering the food web increased with succession age in a similar way as the total amount of C in food web biomass and the number of functional groups in the food web. Connectance did not show any significant changes with succession age, however. In early stages of the succession, the bacterial channel dominated the food web. Later on, in shrub-dominated stands, the fungal channel took over. Even later, in the forest stage, the bacterial channel prevailed again. The best predictor of fungal bacterial ratio is thickness of fermentation layer. We argue that these changes correspond with changes in topsoil microstructure driven by a combination of plant organic matter input and engineering effects of earthworms. In early stages, soil is alkaline, and a discontinuous litter layer on the soil surface promotes bacterial biomass growth, so the bacterial food web channel can dominate. Litter accumulation on the soil surface supports the development of the fungal channel. In older stages, earthworms arrive, mix litter into the mineral soil and form an organo-mineral topsoil, which is beneficial for bacteria and enhances the bacterial food web channel.

  5. Soil Food Web Changes during Spontaneous Succession at Post Mining Sites: A Possible Ecosystem Engineering Effect on Food Web Organization?

    PubMed Central

    Frouz, Jan; Thébault, Elisa; Pižl, Václav; Adl, Sina; Cajthaml, Tomáš; Baldrián, Petr; Háněl, Ladislav; Starý, Josef; Tajovský, Karel; Materna, Jan; Nováková, Alena; de Ruiter, Peter C.

    2013-01-01

    Parameters characterizing the structure of the decomposer food web, biomass of the soil microflora (bacteria and fungi) and soil micro-, meso- and macrofauna were studied at 14 non-reclaimed 1– 41-year-old post-mining sites near the town of Sokolov (Czech Republic). These observations on the decomposer food webs were compared with knowledge of vegetation and soil microstructure development from previous studies. The amount of carbon entering the food web increased with succession age in a similar way as the total amount of C in food web biomass and the number of functional groups in the food web. Connectance did not show any significant changes with succession age, however. In early stages of the succession, the bacterial channel dominated the food web. Later on, in shrub-dominated stands, the fungal channel took over. Even later, in the forest stage, the bacterial channel prevailed again. The best predictor of fungal bacterial ratio is thickness of fermentation layer. We argue that these changes correspond with changes in topsoil microstructure driven by a combination of plant organic matter input and engineering effects of earthworms. In early stages, soil is alkaline, and a discontinuous litter layer on the soil surface promotes bacterial biomass growth, so the bacterial food web channel can dominate. Litter accumulation on the soil surface supports the development of the fungal channel. In older stages, earthworms arrive, mix litter into the mineral soil and form an organo-mineral topsoil, which is beneficial for bacteria and enhances the bacterial food web channel. PMID:24260281

  6. Alkemio: association of chemicals with biomedical topics by text and data mining.

    PubMed

    Gijón-Correas, José A; Andrade-Navarro, Miguel A; Fontaine, Jean F

    2014-07-01

    The PubMed® database of biomedical citations allows the retrieval of scientific articles studying the function of chemicals in biology and medicine. Mining millions of available citations to search reported associations between chemicals and topics of interest would require substantial human time. We have implemented the Alkemio text mining web tool and SOAP web service to help in this task. The tool uses biomedical articles discussing chemicals (including drugs), predicts their relatedness to the query topic with a naïve Bayesian classifier and ranks all chemicals by P-values computed from random simulations. Benchmarks on seven human pathways showed good retrieval performance (areas under the receiver operating characteristic curves ranged from 73.6 to 94.5%). Comparison with existing tools to retrieve chemicals associated to eight diseases showed the higher precision and recall of Alkemio when considering the top 10 candidate chemicals. Alkemio is a high performing web tool ranking chemicals for any biomedical topics and it is free to non-commercial users. http://cbdm.mdc-berlin.de/∼medlineranker/cms/alkemio. © The Author(s) 2014. Published by Oxford University Press on behalf of Nucleic Acids Research.

  7. EAGLE: 'EAGLE'Is an' Algorithmic Graph Library for Exploration

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    2015-01-16

    The Resource Description Framework (RDF) and SPARQL Protocol and RDF Query Language (SPARQL) were introduced about a decade ago to enable flexible schema-free data interchange on the Semantic Web. Today data scientists use the framework as a scalable graph representation for integrating, querying, exploring and analyzing data sets hosted at different sources. With increasing adoption, the need for graph mining capabilities for the Semantic Web has emerged. Today there is no tools to conduct "graph mining" on RDF standard data sets. We address that need through implementation of popular iterative Graph Mining algorithms (Triangle count, Connected component analysis, degree distribution,more » diversity degree, PageRank, etc.). We implement these algorithms as SPARQL queries, wrapped within Python scripts and call our software tool as EAGLE. In RDF style, EAGLE stands for "EAGLE 'Is an' algorithmic graph library for exploration. EAGLE is like 'MATLAB' for 'Linked Data.'« less

  8. 25 CFR 214.13 - Diligence; annual expenditures; mining records.

    Code of Federal Regulations, 2011 CFR

    2011-04-01

    ... within 90 days after an ore body of sufficient quantity is discovered, and shown by the logs or records.... Lessee shall, before commencing operations, file with the superintendent a plat and preliminary statement...

  9. Using Syntactic Patterns to Enhance Text Analytics

    ERIC Educational Resources Information Center

    Meyer, Bradley B.

    2017-01-01

    Large scale product and service reviews proliferate and are commonly found across the web. The ability to harvest, digest and analyze a large corpus of reviews from online websites is still however a difficult problem. This problem is referred to as "opinion mining." Opinion mining is an important area of research as advances in the…

  10. Web Usage Mining: Application to an Online Educational Digital Library Service

    ERIC Educational Resources Information Center

    Palmer, Bart C.

    2012-01-01

    This dissertation was situated in the crossroads of educational data mining (EDM), educational digital libraries (such as the National Science Digital Library; http://nsdl.org), and examination of teacher behaviors while creating online learning resources in an end-user authoring system, the Instructional Architect (IA; http://ia.usu.edu). The…

  11. 30 CFR 74.16 - Material required for record.

    Code of Federal Regulations, 2010 CFR

    2010-07-01

    ... 30 Mineral Resources 1 2010-07-01 2010-07-01 false Material required for record. 74.16 Section 74.16 Mineral Resources MINE SAFETY AND HEALTH ADMINISTRATION, DEPARTMENT OF LABOR COAL MINE SAFETY AND... deliver a complete sampling device free of charge to NIOSH at the address specified on the NIOSH Web page...

  12. Transport Traffic Analysis for Abusive Infrastructure Characterization

    DTIC Science & Technology

    2012-09-01

    3 month sample of spam directed toward the Hotmail web-mail service. Their false positive rate was between 0.0011 and 0.0014 [11]. Unlike autoRE, our...they used 240 machines to analyze a 220 GB Hotmail log in 1.5 hours. In another experiment on 2 months of Hotmail logs (450 GB), BotGraph was able to

  13. Types of Online Hierarchical Repository Structures

    ERIC Educational Resources Information Center

    Hershkovitz, Arnon; Azran, Ronit; Hardof-Jaffe, Sharon; Nachmias, Rafi

    2011-01-01

    This study presents an empirical investigation of online hierarchical repositories of items presented to university students in Web-supported course websites, using Web mining methods. To this end, data from 1747 courses were collected, and the use of online repositories of content items in these courses was examined. At a later stage, courses…

  14. Process modeling and bottleneck mining in online peer-review systems.

    PubMed

    Premchaiswadi, Wichian; Porouhan, Parham

    2015-01-01

    This paper is divided into three main parts. In the first part of the study, we captured, collected and formatted an event log describing the handling of reviews for proceedings of an international conference in Thailand. In the second part, we used several process mining techniques in order to discover process models, social, organizational, and hierarchical structures from the proceeding's event log. In the third part, we detected the deviations and bottlenecks of the peer review process by comparing the observed events (i.e., authentic dataset) with a pre-defined model (i.e., master map). Finally, we investigated the performance information as well as the total waiting time in order to improve the effectiveness and efficiency of the online submission and peer review system for the prospective conferences and seminars. Consequently, the main goals of the study were as follows: (1) to convert the collected event log into the appropriate format supported by process mining analysis tools, (2) to discover process models and to construct social networks based on the collected event log, and (3) to find deviations, discrepancies and bottlenecks between the collected event log and the master pre-defined model. The results showed that although each paper was initially sent to three different reviewers; it was not always possible to make a decision after the first round of reviewing; therefore, additional reviewers were invited. In total, all the accepted and rejected manuscripts were reviewed by an average of 3.9 and 3.2 expert reviewers, respectively. Moreover, obvious violations of the rules and regulations relating to careless or inappropriate peer review of a manuscript-committed by the editorial board and other staff-were identified. Nine blocks of activity in the authentic dataset were not completely compatible with the activities defined in the master model. Also, five of the activity traces were not correctly enabled, and seven activities were missed within the online submission system. On the other hand, dealing with the feedback (comments) received from the first and the third reviewers; the conference committee members and the organizers did not attend to those feedback/comments in a timely manner.

  15. Development of a user-friendly system for image processing of electron microscopy by integrating a web browser and PIONE with Eos.

    PubMed

    Tsukamoto, Takafumi; Yasunaga, Takuo

    2014-11-01

    Eos (Extensible object-oriented system) is one of the powerful applications for image processing of electron micrographs. In usual cases, Eos works with only character user interfaces (CUI) under the operating systems (OS) such as OS-X or Linux, not user-friendly. Thus, users of Eos need to be expert at image processing of electron micrographs, and have a little knowledge of computer science, as well. However, all the persons who require Eos does not an expert for CUI. Thus we extended Eos to a web system independent of OS with graphical user interfaces (GUI) by integrating web browser.Advantage to use web browser is not only to extend Eos with GUI, but also extend Eos to work under distributed computational environment. Using Ajax (Asynchronous JavaScript and XML) technology, we implemented more comfortable user-interface on web browser. Eos has more than 400 commands related to image processing for electron microscopy, and the usage of each command is different from each other. Since the beginning of development, Eos has managed their user-interface by using the interface definition file of "OptionControlFile" written in CSV (Comma-Separated Value) format, i.e., Each command has "OptionControlFile", which notes information for interface and its usage generation. Developed GUI system called "Zephyr" (Zone for Easy Processing of HYpermedia Resources) also accessed "OptionControlFIle" and produced a web user-interface automatically, because its mechanism is mature and convenient,The basic actions of client side system was implemented properly and can supply auto-generation of web-form, which has functions of execution, image preview, file-uploading to a web server. Thus the system can execute Eos commands with unique options for each commands, and process image analysis. There remain problems of image file format for visualization and workspace for analysis: The image file format information is useful to check whether the input/output file is correct and we also need to provide common workspace for analysis because the client is physically separated from a server. We solved the file format problem by extension of rules of OptionControlFile of Eos. Furthermore, to solve workspace problems, we have developed two type of system. The first system is to use only local environments. The user runs a web server provided by Eos, access to a web client through a web browser, and manipulate the local files with GUI on the web browser. The second system is employing PIONE (Process-rule for Input/Output Negotiation Environment), which is our developing platform that works under heterogenic distributed environment. The users can put their resources, such as microscopic images, text files and so on, into the server-side environment supported by PIONE, and so experts can write PIONE rule definition, which defines a workflow of image processing. PIONE run each image processing on suitable computers, following the defined rule. PIONE has the ability of interactive manipulation, and user is able to try a command with various setting values. In this situation, we contribute to auto-generation of GUI for a PIONE workflow.As advanced functions, we have developed a module to log user actions. The logs include information such as setting values in image processing, procedure of commands and so on. If we use the logs effectively, we can get a lot of advantages. For example, when an expert may discover some know-how of image processing, other users can also share logs including his know-hows and so we may obtain recommendation workflow of image analysis, if we analyze logs. To implement social platform of image processing for electron microscopists, we have developed system infrastructure, as well. © The Author 2014. Published by Oxford University Press on behalf of The Japanese Society of Microscopy. All rights reserved. For permissions, please e-mail: journals.permissions@oup.com.

  16. The Impacts of a Web-Aided Instructional Simulation on Science Learning.

    ERIC Educational Resources Information Center

    Hsu, Ying-Shao; Thomas, Rex A.

    2002-01-01

    Investigates the effects of selected characteristics of a web-aided instructional simulation on students' conceptual change, problem solving, and transfer abilities. Conducts a two-pronged research study with (n=117) students enrolled in a beginning meteorology course at Iowa State University. Compares three groups--with-log group, without-log…

  17. 46 CFR 310.58 - Service obligation for students executing or reexecuting contracts.

    Code of Federal Regulations, 2011 CFR

    2011-10-01

    ... repair; municipal and State port authorities; and port development, marine engineering, and tug and barge... MARAD using the web-based Internet system at https://mscs.marad.dot.gov. Reports may also be mailed to... and Plans by logging into the service obligation contract compliance Web site at http://mscs.marad.dot...

  18. 46 CFR 310.58 - Service obligation for students executing or reexecuting contracts.

    Code of Federal Regulations, 2010 CFR

    2010-10-01

    ... repair; municipal and State port authorities; and port development, marine engineering, and tug and barge... MARAD using the web-based Internet system at https://mscs.marad.dot.gov. Reports may also be mailed to... and Plans by logging into the service obligation contract compliance Web site at http://mscs.marad.dot...

  19. SciLinks

    Science.gov Websites

    SciLinks Forgot your login? Sign up for FREE access Log In I'm a ... Teacher Student User Name questions and satisfy their curiosity Learn More Sign up for Free Access Sites in the SciLinks program . SciLinks-Targeted, Grade-Specific Web Content for your Books Free web content to extend and expand student

  20. Use of an Academic Library Web Site Search Engine.

    ERIC Educational Resources Information Center

    Fagan, Jody Condit

    2002-01-01

    Describes an analysis of the search engine logs of Southern Illinois University, Carbondale's library to determine how patrons used the site search. Discusses results that showed patrons did not understand the function of the search and explains improvements that were made in the Web site and in online reference services. (Author/LRW)

  1. Simple, Scalable, Script-based, Science Processor for Measurements - Data Mining Edition (S4PM-DME)

    NASA Astrophysics Data System (ADS)

    Pham, L. B.; Eng, E. K.; Lynnes, C. S.; Berrick, S. W.; Vollmer, B. E.

    2005-12-01

    The S4PM-DME is the Goddard Earth Sciences Distributed Active Archive Center's (GES DAAC) web-based data mining environment. The S4PM-DME replaces the Near-line Archive Data Mining (NADM) system with a better web environment and a richer set of production rules. S4PM-DME enables registered users to submit and execute custom data mining algorithms. The S4PM-DME system uses the GES DAAC developed Simple Scalable Script-based Science Processor for Measurements (S4PM) to automate tasks and perform the actual data processing. A web interface allows the user to access the S4PM-DME system. The user first develops personalized data mining algorithm on his/her home platform and then uploads them to the S4PM-DME system. Algorithms in C and FORTRAN languages are currently supported. The user developed algorithm is automatically audited for any potential security problems before it is installed within the S4PM-DME system and made available to the user. Once the algorithm has been installed the user can promote the algorithm to the "operational" environment. From here the user can search and order the data available in the GES DAAC archive for his/her science algorithm. The user can also set up a processing subscription. The subscription will automatically process new data as it becomes available in the GES DAAC archive. The generated mined data products are then made available for FTP pickup. The benefits of using S4PM-DME are 1) to decrease the downloading time it typically takes a user to transfer the GES DAAC data to his/her system thus off-load the heavy network traffic, 2) to free-up the load on their system, and last 3) to utilize the rich and abundance ocean, atmosphere data from the MODIS and AIRS instruments available from the GES DAAC.

  2. Engagement and Nonusage Attrition With a Free Physical Activity Promotion Program: The Case of 10,000 Steps Australia.

    PubMed

    Guertler, Diana; Vandelanotte, Corneel; Kirwan, Morwenna; Duncan, Mitch J

    2015-07-15

    Data from controlled trials indicate that Web-based interventions generally suffer from low engagement and high attrition. This is important because the level of exposure to intervention content is linked to intervention effectiveness. However, data from real-life Web-based behavior change interventions are scarce, especially when looking at physical activity promotion. The aims of this study were to (1) examine the engagement with the freely available physical activity promotion program 10,000 Steps, (2) examine how the use of a smartphone app may be helpful in increasing engagement with the intervention and in decreasing nonusage attrition, and (3) identify sociodemographic- and engagement-related determinants of nonusage attrition. Users (N=16,948) were grouped based on which platform (website, app) they logged their physical activity: Web only, app only, or Web and app. Groups were compared on sociodemographics and engagement parameters (duration of usage, number of individual and workplace challenges started, and number of physical activity log days) using ANOVA and chi-square tests. For a subsample of users that had been members for at least 3 months (n=11,651), Kaplan-Meier survival curves were estimated to plot attrition over the first 3 months after registration. A Cox regression model was used to determine predictors of nonusage attrition. In the overall sample, user groups differed significantly in all sociodemographics and engagement parameters. Engagement with the program was highest for Web-and-app users. In the subsample, 50.00% (5826/11,651) of users stopped logging physical activity through the program after 30 days. Cox regression showed that user group predicted nonusage attrition: Web-and-app users (hazard ratio=0.86, 95% CI 0.81-0.93, P<.001) and app-only users (hazard ratio=0.63, 95% CI 0.58-0.68, P<.001) showed a reduced attrition risk compared to Web-only users. Further, having a higher number of individual challenges (hazard ratio=0.62, 95% CI 0.59-0.66, P<.001), workplace challenges (hazard ratio=0.94, 95% CI 0.90-0.97, P<.001), physical activity logging days (hazard ratio=0.921, 95% CI 0.919-0.922, P<.001), and steps logged per day (hazard ratio=0.99999, 95% CI 0.99998-0.99999, P<.001) were associated with reduced nonusage attrition risk as well as older age (hazard ratio=0.992, 95% CI 0.991-0.994, P<.001), being male (hazard ratio=0.85, 95% CI 0.82-0.89, P<.001), and being non-Australian (hazard ratio=0.87, 95% CI 0.82-0.91, P<.001). Compared to other freely accessible Web-based health behavior interventions, the 10,000 Steps program showed high engagement. The use of an app alone or in addition to the website can enhance program engagement and reduce risk of attrition. Better understanding of participant reasons for reducing engagement can assist in clarifying how to best address this issue to maximize behavior change.

  3. ESTminer: a Web interface for mining EST contig and cluster databases.

    PubMed

    Huang, Yecheng; Pumphrey, Janie; Gingle, Alan R

    2005-03-01

    ESTminer is a Web application and database schema for interactive mining of expressed sequence tag (EST) contig and cluster datasets. The Web interface contains a query frame that allows the selection of contigs/clusters with specific cDNA library makeup or a threshold number of members. The results are displayed as color-coded tree nodes, where the color indicates the fractional size of each cDNA library component. The nodes are expandable, revealing library statistics as well as EST or contig members, with links to sequence data, GenBank records or user configurable links. Also, the interface allows 'queries within queries' where the result set of a query is further filtered by the subsequent query. ESTminer is implemented in Java/JSP and the package, including MySQL and Oracle schema creation scripts, is available from http://cggc.agtec.uga.edu/Data/download.asp agingle@uga.edu.

  4. Socio-contextual Network Mining for User Assistance in Web-based Knowledge Gathering Tasks

    NASA Astrophysics Data System (ADS)

    Rajendran, Balaji; Kombiah, Iyakutti

    Web-based Knowledge Gathering (WKG) is a specialized and complex information seeking task carried out by many users on the web, for their various learning, and decision-making requirements. We construct a contextual semantic structure by observing the actions of the users involved in WKG task, in order to gain an understanding of their task and requirement. We also build a knowledge warehouse in the form of a master Semantic Link Network (SLX) that accommodates and assimilates all the contextual semantic structures. This master SLX, which is a socio-contextual network, is then mined to provide contextual inputs to the current users through their agents. We validated our approach through experiments and analyzed the benefits to the users in terms of resource explorations and the time saved. The results are positive enough to motivate us to implement in a larger scale.

  5. Final Report on Video Log Data Mining Project

    DOT National Transportation Integrated Search

    2012-06-01

    This report describes the development of an automated computer vision system that identities and inventories road signs : from imagery acquired from the Kansas Department of Transportations road profiling system that takes images every 26.4 : feet...

  6. Workplace Safety and Health Topics: Safety & Prevention

    MedlinePlus

    ... Health Records (EHRs) and Patient Work Information Engineering Controls Equipment Design in Mining Falls in the Workplace Green, Safe, and Healthy Jobs – Prevention through Design Hierarchy of Controls Industry and Occupation Coding and Support Logging Safety ...

  7. PDBj Mine: design and implementation of relational database interface for Protein Data Bank Japan

    PubMed Central

    Kinjo, Akira R.; Yamashita, Reiko; Nakamura, Haruki

    2010-01-01

    This article is a tutorial for PDBj Mine, a new database and its interface for Protein Data Bank Japan (PDBj). In PDBj Mine, data are loaded from files in the PDBMLplus format (an extension of PDBML, PDB's canonical XML format, enriched with annotations), which are then served for the user of PDBj via the worldwide web (WWW). We describe the basic design of the relational database (RDB) and web interfaces of PDBj Mine. The contents of PDBMLplus files are first broken into XPath entities, and these paths and data are indexed in the way that reflects the hierarchical structure of the XML files. The data for each XPath type are saved into the corresponding relational table that is named as the XPath itself. The generation of table definitions from the PDBMLplus XML schema is fully automated. For efficient search, frequently queried terms are compiled into a brief summary table. Casual users can perform simple keyword search, and 'Advanced Search' which can specify various conditions on the entries. More experienced users can query the database using SQL statements which can be constructed in a uniform manner. Thus, PDBj Mine achieves a combination of the flexibility of XML documents and the robustness of the RDB. Database URL: http://www.pdbj.org/ PMID:20798081

  8. PDBj Mine: design and implementation of relational database interface for Protein Data Bank Japan.

    PubMed

    Kinjo, Akira R; Yamashita, Reiko; Nakamura, Haruki

    2010-08-25

    This article is a tutorial for PDBj Mine, a new database and its interface for Protein Data Bank Japan (PDBj). In PDBj Mine, data are loaded from files in the PDBMLplus format (an extension of PDBML, PDB's canonical XML format, enriched with annotations), which are then served for the user of PDBj via the worldwide web (WWW). We describe the basic design of the relational database (RDB) and web interfaces of PDBj Mine. The contents of PDBMLplus files are first broken into XPath entities, and these paths and data are indexed in the way that reflects the hierarchical structure of the XML files. The data for each XPath type are saved into the corresponding relational table that is named as the XPath itself. The generation of table definitions from the PDBMLplus XML schema is fully automated. For efficient search, frequently queried terms are compiled into a brief summary table. Casual users can perform simple keyword search, and 'Advanced Search' which can specify various conditions on the entries. More experienced users can query the database using SQL statements which can be constructed in a uniform manner. Thus, PDBj Mine achieves a combination of the flexibility of XML documents and the robustness of the RDB. Database URL: http://www.pdbj.org/

  9. Near-line Archive Data Mining at the Goddard Distributed Active Archive Center

    NASA Astrophysics Data System (ADS)

    Pham, L.; Mack, R.; Eng, E.; Lynnes, C.

    2002-12-01

    NASA's Earth Observing System (EOS) is generating immense volumes of data, in some cases too much to provide to users with data-intensive needs. As an alternative to moving the data to the user and his/her research algorithms, we are providing a means to move the algorithms to the data. The Near-line Archive Data Mining (NADM) system is the Goddard Earth Sciences Distributed Active Archive Center's (GES DAAC) web data mining portal to the EOS Data and Information System (EOSDIS) data pool, a 50-TB online disk cache. The NADM web portal enables registered users to submit and execute data mining algorithm codes on the data in the EOSDIS data pool. A web interface allows the user to access the NADM system. The users first develops personalized data mining code on their home platform and then uploads them to the NADM system. The C, FORTRAN and IDL languages are currently supported. The user developed code is automatically audited for any potential security problems before it is installed within the NADM system and made available to the user. Once the code has been installed the user is provided a test environment where he/she can test the execution of the software against data sets of the user's choosing. When the user is satisfied with the results, he/she can promote their code to the "operational" environment. From here the user can interactively run his/her code on the data available in the EOSDIS data pool. The user can also set up a processing subscription. The subscription will automatically process new data as it becomes available in the EOSDIS data pool. The generated mined data products are then made available for FTP pickup. The NADM system uses the GES DAAC-developed Simple Scalable Script-based Science Processor (S4P) to automate tasks and perform the actual data processing. Users will also have the option of selecting a DAAC-provided data mining algorithm and using it to process the data of their choice.

  10. Clustering and Dimensionality Reduction to Discover Interesting Patterns in Binary Data

    NASA Astrophysics Data System (ADS)

    Palumbo, Francesco; D'Enza, Alfonso Iodice

    The attention towards binary data coding increased consistently in the last decade due to several reasons. The analysis of binary data characterizes several fields of application, such as market basket analysis, DNA microarray data, image mining, text mining and web-clickstream mining. The paper illustrates two different approaches exploiting a profitable combination of clustering and dimensionality reduction for the identification of non-trivial association structures in binary data. An application in the Association Rules framework supports the theory with the empirical evidence.

  11. Mining the SDSS SkyServer SQL queries log

    NASA Astrophysics Data System (ADS)

    Hirota, Vitor M.; Santos, Rafael; Raddick, Jordan; Thakar, Ani

    2016-05-01

    SkyServer, the Internet portal for the Sloan Digital Sky Survey (SDSS) astronomic catalog, provides a set of tools that allows data access for astronomers and scientific education. One of SkyServer data access interfaces allows users to enter ad-hoc SQL statements to query the catalog. SkyServer also presents some template queries that can be used as basis for more complex queries. This interface has logged over 330 million queries submitted since 2001. It is expected that analysis of this data can be used to investigate usage patterns, identify potential new classes of queries, find similar queries, etc. and to shed some light on how users interact with the Sloan Digital Sky Survey data and how scientists have adopted the new paradigm of e-Science, which could in turn lead to enhancements on the user interfaces and experience in general. In this paper we review some approaches to SQL query mining, apply the traditional techniques used in the literature and present lessons learned, namely, that the general text mining approach for feature extraction and clustering does not seem to be adequate for this type of data, and, most importantly, we find that this type of analysis can result in very different queries being clustered together.

  12. Web Search Studies: Multidisciplinary Perspectives on Web Search Engines

    NASA Astrophysics Data System (ADS)

    Zimmer, Michael

    Perhaps the most significant tool of our internet age is the web search engine, providing a powerful interface for accessing the vast amount of information available on the world wide web and beyond. While still in its infancy compared to the knowledge tools that precede it - such as the dictionary or encyclopedia - the impact of web search engines on society and culture has already received considerable attention from a variety of academic disciplines and perspectives. This article aims to organize a meta-discipline of “web search studies,” centered around a nucleus of major research on web search engines from five key perspectives: technical foundations and evaluations; transaction log analyses; user studies; political, ethical, and cultural critiques; and legal and policy analyses.

  13. TREC Microblog 2012 Track: Real-Time Algorithm for Microblog Ranking Systems

    DTIC Science & Technology

    2012-11-01

    such as information about the tweet and the user profile. We collected those tweets by means of web crawler and extract several features from the raw...Mining Text Data. 2012. [5] D. Feltoni. Twittersa: un sistema per l’analisi del sentimento nelle reti sociali. Master’s thesis, Roma Tre University...Morris. Twittersearch: a comparison of microblog search and web search. Proceedings of the fourth ACM international conference on Web search, 2011

  14. Identifying Engineering Students' English Sentence Reading Comprehension Errors: Applying a Data Mining Technique

    ERIC Educational Resources Information Center

    Tsai, Yea-Ru; Ouyang, Chen-Sen; Chang, Yukon

    2016-01-01

    The purpose of this study is to propose a diagnostic approach to identify engineering students' English reading comprehension errors. Student data were collected during the process of reading texts of English for science and technology on a web-based cumulative sentence analysis system. For the analysis, the association-rule, data mining technique…

  15. Application of Learning Analytics Using Clustering Data Mining for Students' Disposition Analysis

    ERIC Educational Resources Information Center

    Bharara, Sanyam; Sabitha, Sai; Bansal, Abhay

    2018-01-01

    Learning Analytics (LA) is an emerging field in which sophisticated analytic tools are used to improve learning and education. It draws from, and is closely tied to, a series of other fields of study like business intelligence, web analytics, academic analytics, educational data mining, and action analytics. The main objective of this research…

  16. Rare disease diagnosis: A review of web search, social media and large-scale data-mining approaches.

    PubMed

    Svenstrup, Dan; Jørgensen, Henrik L; Winther, Ole

    2015-01-01

    Physicians and the general public are increasingly using web-based tools to find answers to medical questions. The field of rare diseases is especially challenging and important as shown by the long delay and many mistakes associated with diagnoses. In this paper we review recent initiatives on the use of web search, social media and data mining in data repositories for medical diagnosis. We compare the retrieval accuracy on 56 rare disease cases with known diagnosis for the web search tools google.com, pubmed.gov, omim.org and our own search tool findzebra.com. We give a detailed description of IBM's Watson system and make a rough comparison between findzebra.com and Watson on subsets of the Doctor's dilemma dataset. The recall@10 and recall@20 (fraction of cases where the correct result appears in top 10 and top 20) for the 56 cases are found to be be 29%, 16%, 27% and 59% and 32%, 18%, 34% and 64%, respectively. Thus, FindZebra has a significantly (p < 0.01) higher recall than the other 3 search engines. When tested under the same conditions, Watson and FindZebra showed similar recall@10 accuracy. However, the tests were performed on different subsets of Doctors dilemma questions. Advances in technology and access to high quality data have opened new possibilities for aiding the diagnostic process. Specialized search engines, data mining tools and social media are some of the areas that hold promise.

  17. Rare disease diagnosis: A review of web search, social media and large-scale data-mining approaches

    PubMed Central

    Svenstrup, Dan; Jørgensen, Henrik L; Winther, Ole

    2015-01-01

    Physicians and the general public are increasingly using web-based tools to find answers to medical questions. The field of rare diseases is especially challenging and important as shown by the long delay and many mistakes associated with diagnoses. In this paper we review recent initiatives on the use of web search, social media and data mining in data repositories for medical diagnosis. We compare the retrieval accuracy on 56 rare disease cases with known diagnosis for the web search tools google.com, pubmed.gov, omim.org and our own search tool findzebra.com. We give a detailed description of IBM's Watson system and make a rough comparison between findzebra.com and Watson on subsets of the Doctor's dilemma dataset. The recall@10 and recall@20 (fraction of cases where the correct result appears in top 10 and top 20) for the 56 cases are found to be be 29%, 16%, 27% and 59% and 32%, 18%, 34% and 64%, respectively. Thus, FindZebra has a significantly (p < 0.01) higher recall than the other 3 search engines. When tested under the same conditions, Watson and FindZebra showed similar recall@10 accuracy. However, the tests were performed on different subsets of Doctors dilemma questions. Advances in technology and access to high quality data have opened new possibilities for aiding the diagnostic process. Specialized search engines, data mining tools and social media are some of the areas that hold promise. PMID:26442199

  18. Study of Command and Control (C&C) Structures on Integrating Unmanned Autonomous Systems (UAS) into Manned Environments

    DTIC Science & Technology

    2012-09-01

    and traveled all the way around Lake Tahoe. The self - driving cars have logged over 140,000 miles since October 9, 2010 (Google 2010) pictured here...UNDERWATER VEHICLES (AUV) STARFISH is the name given to a small team of autonomous robotic fish - a project carried out by the Acoustic Research...www.scribd.com/doc/42245301/Manual-Mine- Clearance-Book1. Accessed July 23, 2012. Google. The Self - Driving Car Logs more Miles on New Wheels. August 7

  19. A web-based genomic sequence database for the Streptomycetaceae: a tool for systematics and genome mining

    USDA-ARS?s Scientific Manuscript database

    The ARS Microbial Genome Sequence Database (http://199.133.98.43), a web-based database server, was established utilizing the BIGSdb (Bacterial Isolate Genomics Sequence Database) software package, developed at Oxford University, as a tool to manage multi-locus sequence data for the family Streptomy...

  20. Mining the Human Phenome using Semantic Web Technologies: A Case Study for Type 2 Diabetes

    PubMed Central

    Pathak, Jyotishman; Kiefer, Richard C.; Bielinski, Suzette J.; Chute, Christopher G.

    2012-01-01

    The ability to conduct genome-wide association studies (GWAS) has enabled new exploration of how genetic variations contribute to health and disease etiology. However, historically GWAS have been limited by inadequate sample size due to associated costs for genotyping and phenotyping of study subjects. This has prompted several academic medical centers to form “biobanks” where biospecimens linked to personal health information, typically in electronic health records (EHRs), are collected and stored on large number of subjects. This provides tremendous opportunities to discover novel genotype-phenotype associations and foster hypothesis generation. In this work, we study how emerging Semantic Web technologies can be applied in conjunction with clinical and genotype data stored at the Mayo Clinic Biobank to mine the phenotype data for genetic associations. In particular, we demonstrate the role of using Resource Description Framework (RDF) for representing EHR diagnoses and procedure data, and enable federated querying via standardized Web protocols to identify subjects genotyped with Type 2 Diabetes for discovering gene-disease associations. Our study highlights the potential of Web-scale data federation techniques to execute complex queries. PMID:23304343

  1. Mining the human phenome using semantic web technologies: a case study for Type 2 Diabetes.

    PubMed

    Pathak, Jyotishman; Kiefer, Richard C; Bielinski, Suzette J; Chute, Christopher G

    2012-01-01

    The ability to conduct genome-wide association studies (GWAS) has enabled new exploration of how genetic variations contribute to health and disease etiology. However, historically GWAS have been limited by inadequate sample size due to associated costs for genotyping and phenotyping of study subjects. This has prompted several academic medical centers to form "biobanks" where biospecimens linked to personal health information, typically in electronic health records (EHRs), are collected and stored on large number of subjects. This provides tremendous opportunities to discover novel genotype-phenotype associations and foster hypothesis generation. In this work, we study how emerging Semantic Web technologies can be applied in conjunction with clinical and genotype data stored at the Mayo Clinic Biobank to mine the phenotype data for genetic associations. In particular, we demonstrate the role of using Resource Description Framework (RDF) for representing EHR diagnoses and procedure data, and enable federated querying via standardized Web protocols to identify subjects genotyped with Type 2 Diabetes for discovering gene-disease associations. Our study highlights the potential of Web-scale data federation techniques to execute complex queries.

  2. Integration of Text- and Data-Mining Technologies for Use in Banking Applications

    NASA Astrophysics Data System (ADS)

    Maslankowski, Jacek

    Unstructured data, most of it in the form of text files, typically accounts for 85% of an organization's knowledge stores, but it's not always easy to find, access, analyze or use (Robb 2004). That is why it is important to use solutions based on text and data mining. This solution is known as duo mining. This leads to improve management based on knowledge owned in organization. The results are interesting. Data mining provides to lead with structuralized data, usually powered from data warehouses. Text mining, sometimes called web mining, looks for patterns in unstructured data — memos, document and www. Integrating text-based information with structured data enriches predictive modeling capabilities and provides new stores of insightful and valuable information for driving business and research initiatives forward.

  3. Clickstream data yields high-resolution maps of science.

    PubMed

    Bollen, Johan; Van de Sompel, Herbert; Hagberg, Aric; Bettencourt, Luis; Chute, Ryan; Rodriguez, Marko A; Balakireva, Lyudmila

    2009-01-01

    Intricate maps of science have been created from citation data to visualize the structure of scientific activity. However, most scientific publications are now accessed online. Scholarly web portals record detailed log data at a scale that exceeds the number of all existing citations combined. Such log data is recorded immediately upon publication and keeps track of the sequences of user requests (clickstreams) that are issued by a variety of users across many different domains. Given these advantages of log datasets over citation data, we investigate whether they can produce high-resolution, more current maps of science. Over the course of 2007 and 2008, we collected nearly 1 billion user interactions recorded by the scholarly web portals of some of the most significant publishers, aggregators and institutional consortia. The resulting reference data set covers a significant part of world-wide use of scholarly web portals in 2006, and provides a balanced coverage of the humanities, social sciences, and natural sciences. A journal clickstream model, i.e. a first-order Markov chain, was extracted from the sequences of user interactions in the logs. The clickstream model was validated by comparing it to the Getty Research Institute's Architecture and Art Thesaurus. The resulting model was visualized as a journal network that outlines the relationships between various scientific domains and clarifies the connection of the social sciences and humanities to the natural sciences. Maps of science resulting from large-scale clickstream data provide a detailed, contemporary view of scientific activity and correct the underrepresentation of the social sciences and humanities that is commonly found in citation data.

  4. Clickstream Data Yields High-Resolution Maps of Science

    PubMed Central

    Bollen, Johan; Van de Sompel, Herbert; Rodriguez, Marko A.; Balakireva, Lyudmila

    2009-01-01

    Background Intricate maps of science have been created from citation data to visualize the structure of scientific activity. However, most scientific publications are now accessed online. Scholarly web portals record detailed log data at a scale that exceeds the number of all existing citations combined. Such log data is recorded immediately upon publication and keeps track of the sequences of user requests (clickstreams) that are issued by a variety of users across many different domains. Given these advantages of log datasets over citation data, we investigate whether they can produce high-resolution, more current maps of science. Methodology Over the course of 2007 and 2008, we collected nearly 1 billion user interactions recorded by the scholarly web portals of some of the most significant publishers, aggregators and institutional consortia. The resulting reference data set covers a significant part of world-wide use of scholarly web portals in 2006, and provides a balanced coverage of the humanities, social sciences, and natural sciences. A journal clickstream model, i.e. a first-order Markov chain, was extracted from the sequences of user interactions in the logs. The clickstream model was validated by comparing it to the Getty Research Institute's Architecture and Art Thesaurus. The resulting model was visualized as a journal network that outlines the relationships between various scientific domains and clarifies the connection of the social sciences and humanities to the natural sciences. Conclusions Maps of science resulting from large-scale clickstream data provide a detailed, contemporary view of scientific activity and correct the underrepresentation of the social sciences and humanities that is commonly found in citation data. PMID:19277205

  5. Publications - GMC 373 | Alaska Division of Geological & Geophysical

    Science.gov Websites

    DGGS GMC 373 Publication Details Title: 1964 Bear Creek Mining Company drill logs and assay records for and assay records for the Orange Hill Property, Nabesna Quadrangle, Alaska: Drill holes OH #1 and OH

  6. Privileges, Privacy, and Protection of Youth Bloggers in the Social Studies Classroom

    ERIC Educational Resources Information Center

    Berson, Ilene R.; Berson, Michael J.

    2006-01-01

    Internet users continue to develop new ways of communicating online and disseminating information; one of these methods, the blog, also known as web log, has become a significant cultural phenomenon. Blogs offer an interactive medium for internet users to create and contribute content to the web. In some social studies classrooms, teachers are…

  7. Exploring the Use of Blogs as Learning Spaces in the Higher Education Sector

    ERIC Educational Resources Information Center

    Williams, Jeremy B.; Jacobs, Joanne

    2004-01-01

    "Blogging"--a contraction of the term "web logging"--is perhaps best described as a form of micro-publishing. Easy to use, from any Internet connection point, blogging has become firmly established as a web based communications tool. The blogging phenomenon has evolved from its early origin as a medium for the publication of…

  8. 76 FR 54835 - Child Labor Regulations, Orders and Statements of Interpretation; Child Labor Violations-Civil...

    Federal Register 2010, 2011, 2012, 2013, 2014

    2011-09-02

    ....m. in your local time zone, or log onto the Wage and Hour Division's Web site for a nationwide... INFORMATION: I. Electronic Access and Filing Comments Public Participation: This notice of proposed rulemaking is available through the Federal Register and the http://www.regulations.gov Web site. You may also...

  9. At Their Service

    ERIC Educational Resources Information Center

    Villano, Matt

    2006-01-01

    For years, doing laundry at Columbia University (New York) was just as labor-intensive as it is at most universities. Fortunately, as of last spring, laundry life at Columbia has changed dramatically. Today, with the help of a real-time Web-based service called LaundryView, students can log on to the system via the LaundryView Web site from a link…

  10. A Systematic Understanding of Successful Web Searches in Information-Based Tasks

    ERIC Educational Resources Information Center

    Zhou, Mingming

    2013-01-01

    The purpose of this study is to research how Chinese university students solve information-based problems. With the Search Performance Index as the measure of search success, participants were divided into high, medium and low-performing groups. Based on their web search logs, these three groups were compared along five dimensions of the search…

  11. Developing an Efficient Computational Method that Estimates the Ability of Students in a Web-Based Learning Environment

    ERIC Educational Resources Information Center

    Lee, Young-Jin

    2012-01-01

    This paper presents a computational method that can efficiently estimate the ability of students from the log files of a Web-based learning environment capturing their problem solving processes. The computational method developed in this study approximates the posterior distribution of the student's ability obtained from the conventional Bayes…

  12. Bridging the Distance: The Use of Blogs as Reflective Learning Tools for Placement Students

    ERIC Educational Resources Information Center

    Wolf, Katharina

    2010-01-01

    This paper reviews the effectiveness of web logs ("blogs"), or online journals, within the context of a compulsory final-year placement unit for public relations students. The key goal behind the use of Web2.0 technology was to encourage ongoing, reflective practice via an exchange between students thereby limiting feelings of isolation…

  13. Graphical viewer for displaying locations and logs of selected wells and test holes in Putnam County, New York

    USGS Publications Warehouse

    Wolcott, Stephen W.

    2005-01-01

    Aquifers (water bearing geologic units) are the primary source of drinking water in most of Putnam County, N.Y. The principal sources of data used to define the geometry and hydraulic characteristics of aquifers are the logs of wells and test holes within the county. This report explains how to use a graphical viewer, available on the World Wide Web (http://ny.water.usgs.gov/pubs/of/of051198), to locate selected wells and test holes in Putnam County and display their logs.

  14. Randomized controlled trial of a web-based computer-tailored smoking cessation program as a supplement to nicotine patch therapy.

    PubMed

    Strecher, Victor J; Shiffman, Saul; West, Robert

    2005-05-01

    To assess the efficacy of World Wide Web-based tailored behavioral smoking cessation materials among nicotine patch users. Two-group randomized controlled trial. World Wide Web in England and Republic of Ireland. A total of 3971 subjects who purchased a particular brand of nicotine patch and logged-on to use a free web-based behavioral support program. Web-based tailored behavioral smoking cessation materials or web-based non-tailored materials. Twenty-eight-day continuous abstinence rates were assessed by internet-based survey at 6-week follow-up and 10-week continuous rates at 12-week follow-up. Using three approaches to the analyses of 6- and 12-week outcomes, participants in the tailored condition reported clinically and statistically significantly higher continuous abstinence rates than participants in the non-tailored condition. In our primary analyses using as a denominator all subjects who logged-on to the treatment site at least once, continuous abstinence rates at 6 weeks were 29.0% in the tailored condition versus 23.9% in the non-tailored condition (OR = 1.30; P = 0.0006); at 12 weeks continuous abstinence rates were 22.8% versus 18.1%, respectively (OR = 1.34; P = 0.0006). Moreover, satisfaction with the program was significantly higher in the tailored than in the non-tailored condition. The results of this study demonstrate a benefit of the web-based tailored behavioral support materials used in conjunction with nicotine replacement therapy. A web-based program that collects relevant information from users and tailors the intervention to their specific needs had significant advantages over a web-based non-tailored cessation program.

  15. Changes in host-parasitoid food web structure with elevation.

    PubMed

    Maunsell, Sarah C; Kitching, Roger L; Burwell, Chris J; Morris, Rebecca J

    2015-03-01

    Gradients in elevation are increasingly used to investigate how species respond to changes in local climatic conditions. Whilst many studies have shown elevational patterns in species richness and turnover, little is known about how food web structure is affected by elevation. Contrasting responses of predator and prey species to elevation may lead to changes in food web structure. We investigated how the quantitative structure of a herbivore-parasitoid food web changes with elevation in an Australian subtropical rain forest. On four occasions, spread over 1 year, we hand-collected leaf miners at twelve sites, along three elevational gradients (between 493 m and 1159 m a.s.l). A total of 5030 insects, including 603 parasitoids, were reared, and summary food webs were created for each site. We also carried out a replicated manipulative experiment by translocating an abundant leaf-mining weevil Platynotocis sp., which largely escaped parasitism at high elevations (≥ 900 m a.s.l.), to lower, warmer elevations, to test if it would experience higher parasitism pressure. We found strong evidence that the environmental change that occurs with increasing elevation affects food web structure. Quantitative measures of generality, vulnerability and interaction evenness decreased significantly with increasing elevation (and decreasing temperature), whilst elevation did not have a significant effect on connectance. Mined plant composition also had a significant effect on generality and vulnerability, but not on interaction evenness. Several relatively abundant species of leaf miner appeared to escape parasitism at higher elevations, but contrary to our prediction, Platynotocis sp. did not experience greater levels of parasitism when translocated to lower elevations. Our study indicates that leaf-mining herbivores and their parasitoids respond differently to environmental conditions imposed by elevation, thus producing structural changes in their food webs. Increasing temperatures and changes in vegetation communities that are likely to result from climate change may have a restructuring effect on host-parasitoid food webs. Our translocation experiment, however, indicated that leaf miners currently escaping parasitism at high elevations may not automatically experience higher parasitism under warmer conditions and future changes in food web structure may depend on the ability of parasitoids to adapt to novel hosts. © 2014 The Authors. Journal of Animal Ecology © 2014 British Ecological Society.

  16. Impact of Predicting Health Care Utilization Via Web Search Behavior: A Data-Driven Analysis.

    PubMed

    Agarwal, Vibhu; Zhang, Liangliang; Zhu, Josh; Fang, Shiyuan; Cheng, Tim; Hong, Chloe; Shah, Nigam H

    2016-09-21

    By recent estimates, the steady rise in health care costs has deprived more than 45 million Americans of health care services and has encouraged health care providers to better understand the key drivers of health care utilization from a population health management perspective. Prior studies suggest the feasibility of mining population-level patterns of health care resource utilization from observational analysis of Internet search logs; however, the utility of the endeavor to the various stakeholders in a health ecosystem remains unclear. The aim was to carry out a closed-loop evaluation of the utility of health care use predictions using the conversion rates of advertisements that were displayed to the predicted future utilizers as a surrogate. The statistical models to predict the probability of user's future visit to a medical facility were built using effective predictors of health care resource utilization, extracted from a deidentified dataset of geotagged mobile Internet search logs representing searches made by users of the Baidu search engine between March 2015 and May 2015. We inferred presence within the geofence of a medical facility from location and duration information from users' search logs and putatively assigned medical facility visit labels to qualifying search logs. We constructed a matrix of general, semantic, and location-based features from search logs of users that had 42 or more search days preceding a medical facility visit as well as from search logs of users that had no medical visits and trained statistical learners for predicting future medical visits. We then carried out a closed-loop evaluation of the utility of health care use predictions using the show conversion rates of advertisements displayed to the predicted future utilizers. In the context of behaviorally targeted advertising, wherein health care providers are interested in minimizing their cost per conversion, the association between show conversion rate and predicted utilization score, served as a surrogate measure of the model's utility. We obtained the highest area under the curve (0.796) in medical visit prediction with our random forests model and daywise features. Ablating feature categories one at a time showed that the model performance worsened the most when location features were dropped. An online evaluation in which advertisements were served to users who had a high predicted probability of a future medical visit showed a 3.96% increase in the show conversion rate. Results from our experiments done in a research setting suggest that it is possible to accurately predict future patient visits from geotagged mobile search logs. Results from the offline and online experiments on the utility of health utilization predictions suggest that such prediction can have utility for health care providers.

  17. Impact of Predicting Health Care Utilization Via Web Search Behavior: A Data-Driven Analysis

    PubMed Central

    Zhang, Liangliang; Zhu, Josh; Fang, Shiyuan; Cheng, Tim; Hong, Chloe; Shah, Nigam H

    2016-01-01

    Background By recent estimates, the steady rise in health care costs has deprived more than 45 million Americans of health care services and has encouraged health care providers to better understand the key drivers of health care utilization from a population health management perspective. Prior studies suggest the feasibility of mining population-level patterns of health care resource utilization from observational analysis of Internet search logs; however, the utility of the endeavor to the various stakeholders in a health ecosystem remains unclear. Objective The aim was to carry out a closed-loop evaluation of the utility of health care use predictions using the conversion rates of advertisements that were displayed to the predicted future utilizers as a surrogate. The statistical models to predict the probability of user’s future visit to a medical facility were built using effective predictors of health care resource utilization, extracted from a deidentified dataset of geotagged mobile Internet search logs representing searches made by users of the Baidu search engine between March 2015 and May 2015. Methods We inferred presence within the geofence of a medical facility from location and duration information from users’ search logs and putatively assigned medical facility visit labels to qualifying search logs. We constructed a matrix of general, semantic, and location-based features from search logs of users that had 42 or more search days preceding a medical facility visit as well as from search logs of users that had no medical visits and trained statistical learners for predicting future medical visits. We then carried out a closed-loop evaluation of the utility of health care use predictions using the show conversion rates of advertisements displayed to the predicted future utilizers. In the context of behaviorally targeted advertising, wherein health care providers are interested in minimizing their cost per conversion, the association between show conversion rate and predicted utilization score, served as a surrogate measure of the model’s utility. Results We obtained the highest area under the curve (0.796) in medical visit prediction with our random forests model and daywise features. Ablating feature categories one at a time showed that the model performance worsened the most when location features were dropped. An online evaluation in which advertisements were served to users who had a high predicted probability of a future medical visit showed a 3.96% increase in the show conversion rate. Conclusions Results from our experiments done in a research setting suggest that it is possible to accurately predict future patient visits from geotagged mobile search logs. Results from the offline and online experiments on the utility of health utilization predictions suggest that such prediction can have utility for health care providers. PMID:27655225

  18. Influence of plankton mercury dynamics and trophic pathways on mercury concentrations of top predator fish of a mining-impacted reservoir

    USGS Publications Warehouse

    Stewart, A.R.; Saiki, M.K.; Kuwabara, J.S.; Alpers, Charles N.; Marvin-DiPasquale, M.; Krabbenhoft, D.P.

    2008-01-01

    Physical and biogeochemical characteristics of the aquatic environment that affect growth dynamics of phytoplankton and the zooplankton communities that depend on them may also affect uptake of methylmercury (MeHg) into the pelagic food web of oligotrophic reservoirs. We evaluated changes in the quality and quantity of suspended particulate material, zooplankton taxonomy, and MeHg concentrations coincident with seasonal changes in water storage of a mining-impacted reservoir in northern California, USA. MeHg concentrations in bulk zooplankton increased from 4 ng??g-1 at low water to 77 ?? 6.1 ng??g-1 at high water and were positively correlated with cladoceran biomass (r = 0.66) and negatively correlated with rotifer biomass (r = -0.65). Stable isotope analysis revealed overall higher MeHg concentrations in the pelagic-based food web relative to the benthic-based food web. Statistically similar patterns of trophic enrichment of MeHg (slopes) for the pelagic and benthic food webs and slightly higher MeHg concentrations in zooplankton than in benthic invertebrates suggest that the difference in MeHg bioaccumulation among trophic pathways is set at the base of the food webs. These results suggest an important role for plankton dynamics in driving the MeHg content of zooplankton and ultimately MeHg bioaccumulation in top predators in pelagic-based food webs. ?? 2008 NRC.

  19. Mining and Risk of Tuberculosis in Sub-Saharan Africa

    PubMed Central

    Basu, Sanjay; McKee, Martin; Lurie, Mark

    2011-01-01

    Objectives. We estimated the relationship between mining and tuberculosis (TB) among countries in sub-Saharan Africa. Methods. We used multivariate regression to estimate the contribution of mining activity to TB incidence, prevalence, and mortality, as well as rates of TB among people living with HIV, with control for economic, health system, and population confounders. Results. Mining production was associated with higher population TB incidence rates (adjusted b = 0.093; 95% confidence interval [CI] = 0.067, 0.120; with an increase of mining production of 1 SD corresponding to about 33% higher TB incidence or 760 000 more incident cases), after adjustment for economic and population controls. Similar results were observed for TB prevalence and mortality, as well as with alternative measures of mining activity. Independent of HIV, there were significant associations between mining production and TB incidence in countries with high HIV prevalence (≥ 4% antenatal HIV prevalence; HIV-adjusted B = 0.066; 95% CI = 0.050, 0.082) and between log gold mining production and TB incidence in all studied countries (HIV-adjusted B = 0.053; 95% CI = 0.032, 0.073). Conclusions. Mining is a significant determinant of countrywide variation in TB among sub-Saharan African nations. Comprehensive TB control strategies should explicitly address the role of mining activity and environments in the epidemic. PMID:20516372

  20. Mining and risk of tuberculosis in sub-Saharan Africa.

    PubMed

    Stuckler, David; Basu, Sanjay; McKee, Martin; Lurie, Mark

    2011-03-01

    We estimated the relationship between mining and tuberculosis (TB) among countries in sub-Saharan Africa. We used multivariate regression to estimate the contribution of mining activity to TB incidence, prevalence, and mortality, as well as rates of TB among people living with HIV, with control for economic, health system, and population confounders. Mining production was associated with higher population TB incidence rates (adjusted b = 0.093; 95% confidence interval [CI] = 0.067, 0.120; with an increase of mining production of 1 SD corresponding to about 33% higher TB incidence or 760,000 more incident cases), after adjustment for economic and population controls. Similar results were observed for TB prevalence and mortality, as well as with alternative measures of mining activity. Independent of HIV, there were significant associations between mining production and TB incidence in countries with high HIV prevalence (≥ 4% antenatal HIV prevalence; HIV-adjusted B = 0.066; 95% CI = 0.050, 0.082) and between log gold mining production and TB incidence in all studied countries (HIV-adjusted B = 0.053; 95% CI = 0.032, 0.073). Mining is a significant determinant of countrywide variation in TB among sub-Saharan African nations. Comprehensive TB control strategies should explicitly address the role of mining activity and environments in the epidemic.

  1. Measuring the Effects of Cumulative Influence: Using NCW to Prevent or Minimize Civilian Casualties

    DTIC Science & Technology

    2007-06-01

    can be targeted.. How many computers, how many web site visits, how many TV channels, how many text capable cell phones , the number of email...Total number of cell phone text messages transmitted e. Total number of cell phone text messages replied to f. Build a location map of the city’s...determine how many of the identified people have email accounts, web pages, web logs or cell phone text capabilities. The campaign goes something like

  2. Text Mining for Adverse Drug Events: the Promise, Challenges, and State of the Art

    PubMed Central

    Harpaz, Rave; Callahan, Alison; Tamang, Suzanne; Low, Yen; Odgers, David; Finlayson, Sam; Jung, Kenneth; LePendu, Paea; Shah, Nigam H.

    2014-01-01

    Text mining is the computational process of extracting meaningful information from large amounts of unstructured text. Text mining is emerging as a tool to leverage underutilized data sources that can improve pharmacovigilance, including the objective of adverse drug event detection and assessment. This article provides an overview of recent advances in pharmacovigilance driven by the application of text mining, and discusses several data sources—such as biomedical literature, clinical narratives, product labeling, social media, and Web search logs—that are amenable to text-mining for pharmacovigilance. Given the state of the art, it appears text mining can be applied to extract useful ADE-related information from multiple textual sources. Nonetheless, further research is required to address remaining technical challenges associated with the text mining methodologies, and to conclusively determine the relative contribution of each textual source to improving pharmacovigilance. PMID:25151493

  3. Use of Internet audience measurement data to gauge market share for online health information services.

    PubMed

    Wood, Fred B; Benson, Dennis; LaCroix, Eve-Marie; Siegel, Elliot R; Fariss, Susan

    2005-07-01

    The transition to a largely Internet and Web-based environment for dissemination of health information has changed the health information landscape and the framework for evaluation of such activities. A multidimensional evaluative approach is needed. This paper discusses one important dimension of Web evaluation-usage data. In particular, we discuss the collection and analysis of external data on website usage in order to develop a better understanding of the health information (and related US government information) market space, and to estimate the market share or relative levels of usage for National Library of Medicine (NLM) and National Institutes of Health (NIH) websites compared to other health information providers. The primary method presented is Internet audience measurement based on Web usage by external panels of users and assembled by private vendors-in this case, comScore. A secondary method discussed is Web usage based on Web log software data. The principle metrics for both methods are unique visitors and total pages downloaded per month. NLM websites (primarily MedlinePlus and PubMed) account for 55% to 80% of total NIH website usage depending on the metric used. In turn, NIH.gov top-level domain usage (inclusive of NLM) ranks second only behind WebMD in the US domestic home health information market and ranks first on a global basis. NIH.gov consistently ranks among the top three or four US government top-level domains based on global Web usage. On a site-specific basis, the top health information websites in terms of global usage appear to be WebMD, MSN Health, PubMed, Yahoo! Health, AOL Health, and MedlinePlus. Based on MedlinePlus Web log data and external Internet audience measurement data, the three most heavily used cancer-centric websites appear to be www.cancer.gov (National Cancer Institute), www.cancer.org (American Cancer Society), and www.breastcancer.org (non-profit organization). Internet audience measurement has proven useful to NLM, with significant advantages compared to sole reliance on usage data from Web log software. Internet audience data has helped NLM better understand the relative usage of NLM and NIH websites in the intersection of the health information and US government information market sectors, which is the primary market intersector for NLM and NIH. However important, Web usage is only one dimension of a complete Web evaluation framework, and other primary research methods, such as online user surveys, usability tests, and focus groups, are also important for comprehensive evaluation that includes qualitative elements, such as user satisfaction and user friendliness, as well as quantitative indicators of website usage.

  4. Use of Internet Audience Measurement Data to Gauge Market Share for Online Health Information Services

    PubMed Central

    Benson, Dennis; LaCroix, Eve-Marie; Siegel, Elliot R; Fariss, Susan

    2005-01-01

    Background The transition to a largely Internet and Web-based environment for dissemination of health information has changed the health information landscape and the framework for evaluation of such activities. A multidimensional evaluative approach is needed. Objective This paper discusses one important dimension of Web evaluation—usage data. In particular, we discuss the collection and analysis of external data on website usage in order to develop a better understanding of the health information (and related US government information) market space, and to estimate the market share or relative levels of usage for National Library of Medicine (NLM) and National Institutes of Health (NIH) websites compared to other health information providers. Methods The primary method presented is Internet audience measurement based on Web usage by external panels of users and assembled by private vendors—in this case, comScore. A secondary method discussed is Web usage based on Web log software data. The principle metrics for both methods are unique visitors and total pages downloaded per month. Results NLM websites (primarily MedlinePlus and PubMed) account for 55% to 80% of total NIH website usage depending on the metric used. In turn, NIH.gov top-level domain usage (inclusive of NLM) ranks second only behind WebMD in the US domestic home health information market and ranks first on a global basis. NIH.gov consistently ranks among the top three or four US government top-level domains based on global Web usage. On a site-specific basis, the top health information websites in terms of global usage appear to be WebMD, MSN Health, PubMed, Yahoo! Health, AOL Health, and MedlinePlus. Based on MedlinePlus Web log data and external Internet audience measurement data, the three most heavily used cancer-centric websites appear to be www.cancer.gov (National Cancer Institute), www.cancer.org (American Cancer Society), and www.breastcancer.org (non-profit organization). Conclusions Internet audience measurement has proven useful to NLM, with significant advantages compared to sole reliance on usage data from Web log software. Internet audience data has helped NLM better understand the relative usage of NLM and NIH websites in the intersection of the health information and US government information market sectors, which is the primary market intersector for NLM and NIH. However important, Web usage is only one dimension of a complete Web evaluation framework, and other primary research methods, such as online user surveys, usability tests, and focus groups, are also important for comprehensive evaluation that includes qualitative elements, such as user satisfaction and user friendliness, as well as quantitative indicators of website usage. PMID:15998622

  5. LAKE FORK

    EPA Science Inventory

    The Lake Fork of the Arkansas River Watershed has been adversely affected through mining, water diversion and storage projects, grazing, logging, and other human influences over the past 120 years. It is the goals of the LFWWG to improve the health of Lake fork by addressing th...

  6. Understanding Academic Information Seeking Habits through Analysis of Web Server Log Files: The Case of the Teachers College Library Website

    ERIC Educational Resources Information Center

    Asunka, Stephen; Chae, Hui Soo; Hughes, Brian; Natriello, Gary

    2009-01-01

    Transaction logs of user activity on an academic library website were analyzed to determine general usage patterns on the website. This paper reports on insights gained from the analysis, and identifies and discusses issues relating to content access, interface design and general functionality of the website. (Contains 13 figures and 8 tables.)

  7. Learning System of Web Navigation Patterns through Hypertext Probabilistic Grammars

    ERIC Educational Resources Information Center

    Cortes Vasquez, Augusto

    2015-01-01

    One issue of real interest in the area of web data mining is to capture users' activities during connection and extract behavior patterns that help define their preferences in order to improve the design of future pages adapting websites interfaces to individual users. This research is intended to provide, first of all, a presentation of the…

  8. Mining Learning Social Networks for Cooperative Learning with Appropriate Learning Partners in a Problem-Based Learning Environment

    ERIC Educational Resources Information Center

    Chen, Chih-Ming; Chang, Chia-Cheng

    2014-01-01

    Many studies have identified web-based cooperative learning as an increasingly popular educational paradigm with potential to increase learner satisfaction and interactions. However, peer-to-peer interaction often suffers barriers owing to a failure to explore useful social interaction information in web-based cooperative learning environments.…

  9. InterMine Webservices for Phytozome (Rev2)

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Carlson, Joseph; Goodstein, David; Rokhsar, Dan

    2014-07-10

    A datawarehousing framework for information provides a useful infrastructure for providers and users of genomic data. For providers, the infrastructure give them a consistent mechanism for extracting raw data. While for the users, the web services supported by the software allows them to make complex, and often unique, queries of the data. Previously, phytozome.net used BioMart to provide the infrastructure. As the complexity, scale and diversity of the dataset as grown, we decided to implement an InterMine web service on our servers. This change was largely motivated by the ability to have a more complex table structure and richer webmore » reporting mechanism than BioMart. For InterMine to achieve its more complex database schema it requires an XML description of the data and an appropriate loader. Unlimited one-to-many and many-to-many relationship between the tables can be enabled in the schema. We have implemented support for:1.) Genomes and annotations for the data in Phytozome. This set is the 48 organisms currently stored in a back end CHADO datastore. The data loaders are modified versions of the CHADO data adapters from FlyMine. 2.) Interproscan results from all proteins in the Phytozome database. 3.) Clusters of proteins into a grouped heirarchically by similarity. 4.) Cufflinks results from tissue-specific RNA-Seq data of Phytozome organisms. 5.) Diversity data (GATK and SnpEFF results) from a set of individual organism. The last two datatypes are new in this implementation of our web services. We anticipate that the scale of these data will increase considerably in the near future.« less

  10. minepath.org: a free interactive pathway analysis web server.

    PubMed

    Koumakis, Lefteris; Roussos, Panos; Potamias, George

    2017-07-03

    ( www.minepath.org ) is a web-based platform that elaborates on, and radically extends the identification of differentially expressed sub-paths in molecular pathways. Besides the network topology, the underlying MinePath algorithmic processes exploit exact gene-gene molecular relationships (e.g. activation, inhibition) and are able to identify differentially expressed pathway parts. Each pathway is decomposed into all its constituent sub-paths, which in turn are matched with corresponding gene expression profiles. The highly ranked, and phenotype inclined sub-paths are kept. Apart from the pathway analysis algorithm, the fundamental innovation of the MinePath web-server concerns its advanced visualization and interactive capabilities. To our knowledge, this is the first pathway analysis server that introduces and offers visualization of the underlying and active pathway regulatory mechanisms instead of genes. Other features include live interaction, immediate visualization of functional sub-paths per phenotype and dynamic linked annotations for the engaged genes and molecular relations. The user can download not only the results but also the corresponding web viewer framework of the performed analysis. This feature provides the flexibility to immediately publish results without publishing source/expression data, and get all the functionality of a web based pathway analysis viewer. © The Author(s) 2017. Published by Oxford University Press on behalf of Nucleic Acids Research.

  11. PLAN2L: a web tool for integrated text mining and literature-based bioentity relation extraction.

    PubMed

    Krallinger, Martin; Rodriguez-Penagos, Carlos; Tendulkar, Ashish; Valencia, Alfonso

    2009-07-01

    There is an increasing interest in using literature mining techniques to complement information extracted from annotation databases or generated by bioinformatics applications. Here we present PLAN2L, a web-based online search system that integrates text mining and information extraction techniques to access systematically information useful for analyzing genetic, cellular and molecular aspects of the plant model organism Arabidopsis thaliana. Our system facilitates a more efficient retrieval of information relevant to heterogeneous biological topics, from implications in biological relationships at the level of protein interactions and gene regulation, to sub-cellular locations of gene products and associations to cellular and developmental processes, i.e. cell cycle, flowering, root, leaf and seed development. Beyond single entities, also predefined pairs of entities can be provided as queries for which literature-derived relations together with textual evidences are returned. PLAN2L does not require registration and is freely accessible at http://zope.bioinfo.cnio.es/plan2l.

  12. HS.Register - An Audit-Trail Tool to Respond to the General Data Protection Regulation (GDPR).

    PubMed

    Gonçalves-Ferreira, Duarte; Leite, Mariana; Santos-Pereira, Cátia; Correia, Manuel E; Antunes, Luis; Cruz-Correia, Ricardo

    2018-01-01

    Introduction The new General Data Protection Regulation (GDPR) compels health care institutions and their software providers to properly document all personal data processing and provide clear evidence that their systems are inline with the GDPR. All applications involved in personal data processing should therefore produce meaningful event logs that can later be used for the effective auditing of complex processes. Aim This paper aims to describe and evaluate HS.Register, a system created to collect and securely manage at scale audit logs and data produced by a large number of systems. Methods HS.Register creates a single audit log by collecting and aggregating all kinds of meaningful event logs and data (e.g. ActiveDirectory, syslog, log4j, web server logs, REST, SOAP and HL7 messages). It also includes specially built dashboards for easy auditing and monitoring of complex processes, crossing different systems in an integrated way, as well as providing tools for helping on the auditing and on the diagnostics of difficult problems, using a simple web application. HS.Register is currently installed at five large Portuguese Hospitals and is composed of the following open-source components: HAproxy, RabbitMQ, Elasticsearch, Logstash and Kibana. Results HS.Register currently collects and analyses an average of 93 million events per week and it is being used to document and audit HL7 communications. Discussion Auditing tools like HS.Register are likely to become mandatory in the near future to allow for traceability and detailed auditing for GDPR compliance.

  13. Morbidity and Health Risk Factors Among New Mexico Miners: A Comparison Across Mining Sectors.

    PubMed

    Shumate, Alice M; Yeoman, Kristin; Victoroff, Tristan; Evans, Kandace; Karr, Roger; Sanchez, Tami; Sood, Akshay; Laney, Anthony Scott

    2017-08-01

    This study examines differences in chronic health outcomes between coal, uranium, metal, and nonmetal miners. In a cross-sectional study using data from a health screening program for current and former New Mexico miners, log-binomial logistic regression models were used to estimate relative risks of respiratory and heart disease, cancer, osteoarthritis, and back pain associated with mining in each sector as compared with coal, adjusting for other relevant risk factors. Differential risks in angina, pulmonary symptoms, asthma, cancer, osteoarthritis, and back pain between mining sectors were found. New Mexico miners experience different chronic health challenges across sectors. These results demonstrate the importance of using comparable data to understand how health risks differ across mining sectors. Further investigation among a broader geographic population of miners will help identify the health priorities and needs in each sector.

  14. Applicability of a Web-Based, Individualized Exercise Intervention in Patients With Liver Disease, Cystic Fibrosis, Esophageal Cancer, and Psychiatric Disorders: Process Evaluation of 4 Ongoing Clinical Trials.

    PubMed

    Pfirrmann, Daniel; Haller, Nils; Huber, Yvonne; Jung, Patrick; Lieb, Klaus; Gockel, Ines; Poplawska, Krystyna; Schattenberg, Jörn Markus; Simon, Perikles

    2018-05-22

    In the primary and secondary prevention of civilization diseases, regular physical activity is recommended in international guidelines to improve disease-related symptoms, delay the progression of the disease, or to enhance postoperative outcomes. In the preoperative context, there has been a paradigm shift in favor of using preconditioning concepts before surgery. Web-based interventions seem an innovative and effective tool for delivering general information, individualized exercise recommendations, and peer support. Our first objective was to assess feasibility of our Web-based interventional concept and analyze similarities and differences in a sustained exercise implementation in different diseases. The second objective was to investigate the overall participants' satisfaction with our Web-based concept. A total of 4 clinical trials are still being carried out, including patients with esophageal carcinoma scheduled for oncologic esophagectomy (internet-based perioperative exercise program, iPEP, study), nonalcoholic fatty liver disease (hepatic inflammation and physical performance in patients with nonalcoholic steatohepatitis, HELP, study), depression (exercise for depression, EXDEP, study), and cystic fibrosis (cystic fibrosis online mentoring for microbiome, exercise, and diet, COMMED, study). During the intervention period, the study population had access to the website with disease-specific content and a disease-specific discussion forum. All participants received weekly, individual tailored exercise recommendations from the sports therapist. The main outcome was the using behavior, which was obtained by investigating the log-in rate and duration. A total of 20 participants (5 from each trial) were analyzed. During the intervention period, a regular contact and a consequent implementation of exercise prescription were easily achieved in all substudies. Across the 4 substudies, there was a significant decrease in log-in rates (P<.001) and log-in durations (P<.001) over time. A detailed view of the different studies shows a significant decrease in log-in rates and log-in durations in the HELP study (P=.004; P=.002) and iPEP study (P=.02; P=.001), whereas the EXDEP study (P=.58; P=.38) and COMMED study (P=.87; P=.56) showed no significant change over the 8-week intervention period. There was no significant change in physical activity within all studies (P=.31). Only in the HELP study, the physical activity level increased steadily over the period analyzed (P=.045). Overall, 17 participants (85%, 17/20) felt secure and were not scared of injury, with no major differences in the subtrials. The universal use of the Web-based intervention appears to be applicable across the heterogonous collectives of our study patients with regard to age and disease. Although the development of physical activity shows only moderate improvements, flexible communication and tailored support could be easily integrated into patients' daily routine. iPEP study: ClinicalTrials.gov NCT02478996; https://clinicaltrials.gov/ct2/show/NCT02478996 (Archived by WebCite at http://www.webcitation.org/6zL1UmHaW); HELP study: ClinicalTrials.gov NCT02526732; http://www.webcitation.org/6zJjX7d6K (Archived by WebCite at http://www.webcitation.org/6Nch4ldcL); EXDEP study: ClinicalTrials.gov NCT02874833; https://clinicaltrials.gov/ct2/show/NCT02874833 (Archived by WebCite at http://www.webcitation.org/6zJjj7FuA). ©Daniel Pfirrmann, Nils Haller, Yvonne Huber, Patrick Jung, Klaus Lieb, Ines Gockel, Krystyna Poplawska, Jörn Markus Schattenberg, Perikles Simon. Originally published in JMIR Research Protocols (http://www.researchprotocols.org), 22.05.2018.

  15. Blog Revolution: Expanding Classroom Horizons with Web Logs

    ERIC Educational Resources Information Center

    Richardson, Will

    2005-01-01

    Blogs are not a passing fad as a new blog is created every second. There are more than 900,000 blog posts a day. Blogs are one of many new disruptive technologies that are transforming the world. They are creating a richer, more dynamic, more interactive Web where participation is the rule rather than the exception. Classrooms and schools are…

  16. Some Features of "Alt" Texts Associated with Images in Web Pages

    ERIC Educational Resources Information Center

    Craven, Timothy C.

    2006-01-01

    Introduction: This paper extends a series on summaries of Web objects, in this case, the alt attribute of image files. Method: Data were logged from 1894 pages from Yahoo!'s random page service and 4703 pages from the Google directory; an img tag was extracted randomly from each where present; its alt attribute, if any, was recorded; and the…

  17. A Password-Protected Web Site for Mothers Expressing Milk for Their Preterm Infants.

    PubMed

    Blatz, MaryAnn; Dowling, Donna; Underwood, Patricia W; Bieda, Amy; Graham, Gregory

    2017-06-01

    Research has demonstrated that breast milk significantly decreases morbidities that impact length of stay for preterm infants, but there is a need to test interventions to improve breastfeeding outcomes. Since many Americans are using technologies such as the Intranet and smartphones to find health information and manage health, a Web site was developed for mothers who provide breast milk for their preterm hospitalized infants. This study examined the efficacy of a Web site for mothers to educate them about breast milk expression and assist them in monitoring their breast milk supply. Quantitative and qualitative data were collected from mothers whose preterm infants were hospitalized in a level IV neonatal intensive care unit (NICU) or transitional care unit (TCU) in an urban academic medical center in the Midwest. Eighteen mothers participated in evaluation of the Web site. Thirteen mothers consistently logged on to the password-protected Web site (mean [standard deviation] = 13.3 [11.7]) times. Most participants, (69.2%), reported they used the breast milk educational information. Most mothers indicated that using the Web site log helped in tracking their pumping. These findings can be used to direct the design and development of web-based resources for mothers of preterm infants IMPLICATIONS FOR PRACTICE:: NICU and TCU staffs need to examine and establish approaches to actively involve mothers in monitoring the establishment and maintenance of an adequate supply of breast milk to improve neonatal health outcomes. An electronic health application that incorporates the features identified in this study should be developed and tested.

  18. 30 CFR 955.14 - Examination.

    Code of Federal Regulations, 2010 CFR

    2010-07-01

    ... CONDUCT OF SURFACE MINING OPERATIONS WITHIN EACH STATE CERTIFICATION OF BLASTERS IN FEDERAL PROGRAM STATES...(b) of this chapter, and shall include: (i) Objective questions; (ii) Blasting log problems; and (iii) Initiation system and delay sequence problems. (c) Reexamination. (1) Any person who fails the examination...

  19. ASSESSING STREAM BED STABILITY AND EXCESS SEDIMENTATION IN MOUNTAIN STREAMS

    EPA Science Inventory

    Land use and resource exploitation in headwaters catchments?such as logging, mining, and road building?often increase sediment supply to streams, potentially causing excess sedimentation. Decreases in mean substrate size and increases in fine stream bed sediments can lead to inc...

  20. Handling Dynamic Weights in Weighted Frequent Pattern Mining

    NASA Astrophysics Data System (ADS)

    Ahmed, Chowdhury Farhan; Tanbeer, Syed Khairuzzaman; Jeong, Byeong-Soo; Lee, Young-Koo

    Even though weighted frequent pattern (WFP) mining is more effective than traditional frequent pattern mining because it can consider different semantic significances (weights) of items, existing WFP algorithms assume that each item has a fixed weight. But in real world scenarios, the weight (price or significance) of an item can vary with time. Reflecting these changes in item weight is necessary in several mining applications, such as retail market data analysis and web click stream analysis. In this paper, we introduce the concept of a dynamic weight for each item, and propose an algorithm, DWFPM (dynamic weighted frequent pattern mining), that makes use of this concept. Our algorithm can address situations where the weight (price or significance) of an item varies dynamically. It exploits a pattern growth mining technique to avoid the level-wise candidate set generation-and-test methodology. Furthermore, it requires only one database scan, so it is eligible for use in stream data mining. An extensive performance analysis shows that our algorithm is efficient and scalable for WFP mining using dynamic weights.

  1. Selenium in ecosystems within the mountaintop coal mining and valley-fill region of southern West Virginia-assessment and ecosystem-scale modeling

    USGS Publications Warehouse

    Presser, Theresa S.

    2013-01-01

    Investigating the presence and variability of prey and predator species in demographically open systems such as streams also is key to model outcomes given the overall environmental stressors (for example, general landscape change, food-web disruption, recolonization potential) imposed on the composition of biological communities in coal mining and valley-fill affected watersheds

  2. Publications - GMC 273 | Alaska Division of Geological & Geophysical

    Science.gov Websites

    holes received at the GMC (1 box, holes N1 through N8) of the INEXCO Mining Company Nikolai Project , holes N1 through N8) of the INEXCO Mining Company Nikolai Project, McCarthy, Alaska that consist of core Alaska's Mineral Industry Reports AKGeology.info Rare Earth Elements WebGeochem Engineering Geology Alaska

  3. An open-source data storage and visualization back end for experimental data.

    PubMed

    Nielsen, Kenneth; Andersen, Thomas; Jensen, Robert; Nielsen, Jane H; Chorkendorff, Ib

    2014-04-01

    In this article, a flexible free and open-source software system for data logging and presentation will be described. The system is highly modular and adaptable and can be used in any laboratory in which continuous and/or ad hoc measurements require centralized storage. A presentation component for the data back end has furthermore been written that enables live visualization of data on any device capable of displaying Web pages. The system consists of three parts: data-logging clients, a data server, and a data presentation Web site. The logging of data from independent clients leads to high resilience to equipment failure, whereas the central storage of data dramatically eases backup and data exchange. The visualization front end allows direct monitoring of acquired data to see live progress of long-duration experiments. This enables the user to alter experimental conditions based on these data and to interfere with the experiment if needed. The data stored consist both of specific measurements and of continuously logged system parameters. The latter is crucial to a variety of automation and surveillance features, and three cases of such features are described: monitoring system health, getting status of long-duration experiments, and implementation of instant alarms in the event of failure.

  4. Clickstream data yields high-resolution maps of science

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Bollen, Johan; Van De Sompel, Herbert; Hagberg, Aric

    2009-01-01

    Intricate maps of science have been created from citation data to visualize the structure of scientific activity. However, most scientific publications are now accessed online. Scholarly web portals record detailed log data at a scale that exceeds the number of all existing citations combined. Such log data is recorded immediately upon publication and keeps track of the sequences of user requests (clickstreams) that are issued by a variety of users across many different domains. Given these advantagees of log datasets over citation data, we investigate whether they can produce high-resolution, more current maps of science.

  5. WWW Motivation Mining: Finding Treasures for Teaching Evaluation Skills, Grades 7-12. Professional Growth Series.

    ERIC Educational Resources Information Center

    Small, Ruth V.; Arnone, Marilyn P.

    Intended for use by middle or high school teachers and library media specialists, this book describes a World Wide Web evaluation tool developed specifically for use by high school students and designed to provide hands-on experience in critically evaluating the strengths and weaknesses of Web sites. The book uses a workbook format and is…

  6. Simplified process model discovery based on role-oriented genetic mining.

    PubMed

    Zhao, Weidong; Liu, Xi; Dai, Weihui

    2014-01-01

    Process mining is automated acquisition of process models from event logs. Although many process mining techniques have been developed, most of them are based on control flow. Meanwhile, the existing role-oriented process mining methods focus on correctness and integrity of roles while ignoring role complexity of the process model, which directly impacts understandability and quality of the model. To address these problems, we propose a genetic programming approach to mine the simplified process model. Using a new metric of process complexity in terms of roles as the fitness function, we can find simpler process models. The new role complexity metric of process models is designed from role cohesion and coupling, and applied to discover roles in process models. Moreover, the higher fitness derived from role complexity metric also provides a guideline for redesigning process models. Finally, we conduct case study and experiments to show that the proposed method is more effective for streamlining the process by comparing with related studies.

  7. Increased capture of pediatric surgical complications utilizing a novel case-log web application to enhance quality improvement.

    PubMed

    Fisher, Jason C; Kuenzler, Keith A; Tomita, Sandra S; Sinha, Prashant; Shah, Paresh; Ginsburg, Howard B

    2017-01-01

    Documenting surgical complications is limited by multiple barriers and is not fostered in the electronic health record. Tracking complications is essential for quality improvement (QI) and required for board certification. Current registry platforms do not facilitate meaningful complication reporting. We developed a novel web application that improves accuracy and reduces barriers to documenting complications. We deployed a custom web application that allows pediatric surgeons to maintain case logs. The program includes a module for entering complication data in real time. Reminders to enter outcome data occur at key postoperative intervals to optimize recall of events. Between October 1, 2014, and March 31, 2015, frequencies of surgical complications captured by the existing hospital reporting system were compared with data aggregated by our application. 780 cases were captured by the web application, compared with 276 cases registered by the hospital system. We observed an increase in the capture of major complications when compared to the hospital dataset (14 events vs. 4 events). This web application improved real-time reporting of surgical complications, exceeding the accuracy of administrative datasets. Custom informatics solutions may help reduce barriers to self-reporting of adverse events and improve the data that presently inform pediatric surgical QI. Diagnostic study/Retrospective study. Level III - case control study. Copyright © 2017 Elsevier Inc. All rights reserved.

  8. Web mining for topics defined by complex and precise predicates

    NASA Astrophysics Data System (ADS)

    Lee, Ching-Cheng; Sampathkumar, Sushma

    2004-04-01

    The enormous growth of the World Wide Web has made it important to perform resource discovery efficiently for any given topic. Several new techniques have been proposed in the recent years for this kind of topic specific web-mining, and among them a key new technique called focused crawling which is able to crawl topic-specific portions of the web without having to explore all pages. Most existing research on focused crawling considers a simple topic definition that typically consists of one or more keywords connected by an OR operator. However this kind of simple topic definition may result in too many irrelevant pages in which the same keyword appears in a wrong context. In this research we explore new strategies for crawling topic specific portions of the web using complex and precise predicates. A complex predicate will allow the user to precisely specify a topic using Boolean operators such as "AND", "OR" and "NOT". Our work will concentrate on defining a format to specify this kind of a complex topic definition and secondly on devising a crawl strategy to crawl the topic specific portions of the web defined by the complex predicate, efficiently and with minimal overhead. Our new crawl strategy will improve the performance of topic-specific web crawling by reducing the number of irrelevant pages crawled. In order to demonstrate the effectiveness of the above approach, we have built a complete focused crawler called "Eureka" with complex predicate support, and a search engine that indexes and supports end-user searches on the crawled pages.

  9. Disentangling road network impacts: The need for a holistic approach

    USDA-ARS?s Scientific Manuscript database

    Traditional and alternative energy development, logging and mining activities, together with off-highway vehicles (OHV) and exurban development, have increased the density of linear disturbances on public and private lands throughout the world. We argue that the dramatic increase in linear disturba...

  10. Use of a Case-Based Hypermedia Resource in an Early Literacy Coaching Intervention with Pre-Kindergarten Teachers

    ERIC Educational Resources Information Center

    Powell, Douglas R.; Diamond, Karen E.; Koehler, Matthew J.

    2010-01-01

    Use of a case-based hypermedia resource (HR) was examined in a Web-based early literacy coaching intervention with pre-kindergarten teachers of at-risk children. Web usage logs, written records of coach feedback to teachers on their instruction, and a teacher questionnaire were the primary data sources. Visits to the HR content pages were unevenly…

  11. 76 FR 7838 - Claverack Creek, LLC; Notice of Preliminary Permit Application Accepted for Filing and Soliciting...

    Federal Register 2010, 2011, 2012, 2013, 2014

    2011-02-11

    ...-deep intake canal; (5) new trash racks, head gates, and stop log structure; (6) an existing 6-foot... Internet. See 18 CFR 385.2001(a)(1)(iii) and the instructions on the Commission's Web site http://www.ferc... copy of the application, can be viewed or printed on the ``eLibrary'' link of the Commission's Web site...

  12. Patient Adoption and Utilization of a Web-Based and Mobile-Based Portal for Collecting Outcomes After Elective Orthopedic Surgery.

    PubMed

    Bell, Kerri; Warnick, Eugene; Nicholson, Kristen; Ulcoq, Sarah; Kim, Seong Jin; Schroeder, Gregory D; Vaccaro, Alexander

    2018-03-01

    Health care increasingly collects patient-reported outcomes (PROs) via web-based platforms. The purpose of this study was to evaluate how patient age influences portal engagement. Patients undergoing elective surgery at a single multispecialty orthopedic practice from September 2014 to February 2017 had access to an online portal to complete PROs, message the clinic, and view physical therapy instructions. A mobile app was optionally available. Age, sex, log-in frequency, PRO completion rates, and number of messages sent were reviewed retrospectively. Message frequency, log-in rates, and PRO compliance were highest for patients aged 41 to 50, 51 to 60, and 61 to 70, respectively. Mobile app use decreased with age ( P = .002); yet, at all ages, the mobile app group was more engaged. In particular, for patients aged 18 to 30 years, log-in frequency increased 2.5-fold and PRO compliance improved 44% ( P < .001) in the mobile app group. This study demonstrates that portal interaction varies by age and that data capture is highest in patients who choose the mobile app.

  13. Size distribution of radon daughter particles in uranium mine atmospheres.

    PubMed

    George, A C; Hinchliffe, L; Sladowski, R

    1975-06-01

    The size distribution of radon daughters was measured in several uranium mines using four compact diffusion batteries and a round jet cascade impactor. Simultaneously, measurements were made of uncombined fractions of radon daughters, radon concentration, working level and particle concentration. The size distributions found for radon daughters were log normal. The activity median diameters ranged from 0.09 mum to 0.3 mum with a mean value of 0.17 mum. Geometric standard deviations were in the range from 1.3 to 4 with a mean value of 2.7. Uncombined fractions expressed in accordance with the ICRP definition ranged from 0.004 to 0.16 with a mean value of 0.04. The radon daughter sizes in these mines are greater than the sizes assumed by various authors in calculating respiratory tract dose. The disparity may reflect the widening use of diesel-powered equipment in large uranium mines.

  14. Stress monitoring versus microseismic ruptures in an active deep mine

    NASA Astrophysics Data System (ADS)

    Tonnellier, Alice; Bouffier, Christian; Bigarré, Pascal; Nyström, Anders; Österberg, Anders; Fjellström, Peter

    2015-04-01

    Nowadays, underground mining industry has developed high-technology mass mining methods to optimise the productivity at deep levels. Such massive extraction induces high-level stress redistribution generating seismic events around the mining works, threatening safety and economics. For this reason mining irregular deep ore bodies calls for steadily enhanced scientific practises and technologies to guarantee the mine environment to be safer and stable for the miners and the infrastructures. INERIS, within the framework of the FP7 European project I2Mine and in partnership with the Swedish mining company Boliden, has developed new methodologies in order to monitor both quasi-static stress changes and ruptures in a seismic prone area. To this purpose, a unique local permanent microseismic and stress monitoring network has been installed into the deep-working Garpenberg mine situated to the north of Uppsala (Sweden). In this mine, ore is extracted using sublevel stoping with paste fill production/distribution system and long-hole drilling method. This monitoring network has been deployed between about 1100 and 1250 meter depth. It consists in six 1-component and five 3-component microseismic probes (14-Hz geophones) deployed in the Lappberget area, in addition to three 3D stress monitoring cells that focus on a very local exploited area. Objective is three-fold: to quantify accurately quasi-static stress changes and freshly-induced stress gradients with drift development in the orebody, to study quantitatively those stress changes versus induced detected and located microseismic ruptures, and possibly to identify quasi-static stress transfer from those seismic ruptures. Geophysical and geotechnical data are acquired continuously and automatically transferred to INERIS datacenter through the web. They are made available on a secured web cloud monitoring infrastructure called e.cenaris and completed with mine data. Such interface enables the visualisation of the monitoring data coming from the mine in quasi-real time and facilitates information exchanges and decision making for experts and stakeholders. On the basis of these data acquisition and sharing, preliminary analysis has been started to highlight whether stress variations and seismic sources behaviour might be directly bound with mine working evolution and could improve the knowledge on the equilibrium states inside the mine. Knowing such parameters indeed will be a potential solution to understand better the response of deep mining activities to the exploitation solicitations and to develop, if possible, methods to prevent from major hazards such as rock bursts and other ground failure phenomena.

  15. Data Mining of Web-Based Documents on Social Networking Sites That Included Suicide-Related Words Among Korean Adolescents.

    PubMed

    Song, Juyoung; Song, Tae Min; Seo, Dong-Chul; Jin, Jae Hyun

    2016-12-01

    To investigate online search activity of suicide-related words in South Korean adolescents through data mining of social media Web sites as the suicide rate in South Korea is one of the highest in the world. Out of more than 2.35 billion posts for 2 years from January 1, 2011 to December 31, 2012 on 163 social media Web sites in South Korea, 99,693 suicide-related documents were retrieved by Crawler and analyzed using text mining and opinion mining. These data were further combined with monthly employment rate, monthly rental prices index, monthly youth suicide rate, and monthly number of reported bully victims to fit multilevel models as well as structural equation models. The link from grade pressure to suicide risk showed the largest standardized path coefficient (beta = .357, p < .001) in structural models and a significant random effect (p < .01) in multilevel models. Depression was a partial mediator between suicide risk and grade pressure, low body image, victims of bullying, and concerns about disease. The largest total effect was observed in the grade pressure to depression to suicide risk. The multilevel models indicate about 27% of the variance in the daily suicide-related word search activity is explained by month-to-month variations. A lower employment rate, a higher rental prices index, and more bullying were associated with an increased suicide-related word search activity. Academic pressure appears to be the biggest contributor to Korean adolescents' suicide risk. Real-time suicide-related word search activity monitoring and response system needs to be developed. Copyright © 2016 Society for Adolescent Health and Medicine. Published by Elsevier Inc. All rights reserved.

  16. LimTox: a web tool for applied text mining of adverse event and toxicity associations of compounds, drugs and genes

    PubMed Central

    Cañada, Andres; Rabal, Obdulia; Oyarzabal, Julen; Valencia, Alfonso

    2017-01-01

    Abstract A considerable effort has been devoted to retrieve systematically information for genes and proteins as well as relationships between them. Despite the importance of chemical compounds and drugs as a central bio-entity in pharmacological and biological research, only a limited number of freely available chemical text-mining/search engine technologies are currently accessible. Here we present LimTox (Literature Mining for Toxicology), a web-based online biomedical search tool with special focus on adverse hepatobiliary reactions. It integrates a range of text mining, named entity recognition and information extraction components. LimTox relies on machine-learning, rule-based, pattern-based and term lookup strategies. This system processes scientific abstracts, a set of full text articles and medical agency assessment reports. Although the main focus of LimTox is on adverse liver events, it enables also basic searches for other organ level toxicity associations (nephrotoxicity, cardiotoxicity, thyrotoxicity and phospholipidosis). This tool supports specialized search queries for: chemical compounds/drugs, genes (with additional emphasis on key enzymes in drug metabolism, namely P450 cytochromes—CYPs) and biochemical liver markers. The LimTox website is free and open to all users and there is no login requirement. LimTox can be accessed at: http://limtox.bioinfo.cnio.es PMID:28531339

  17. Abandoned mines and their impact on the environment: Case studies from Franklin and Sterling Mines, NJ and Rondout Quarry, NY

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Kolkas, M.M.; Nehru, C.E.

    1995-09-01

    Water logged abandoned mines have an impact on the environment. In this project we selected abandoned mines from two sets of different ore bodies to learn about their environmental impact. Franklin and Sterling Pb-Zn mines, NJ and the limestone quarry in Rondout formation, NY were selected as case study examples. In the Pb-Zn mines metalimestone is the country rock and in the Rondout quarry limestone is the country rock. Soil water samples from selected strategic locations were analyzed for toxic and related heavy metal elements such as Pb, Zn, Cd, Cr and U. The levels of concentrations of these elementsmore » varied from one location to another according to the chemistry of the ore body and the ground movement throughout the area. In particular Cd, Cr and U concentration were variable from Franklin to Sterling mine. However, in the Rondout limestone (cement) quarry, higher concentrations of Cr and lower concentrations of Pb and Zn were noted. We conclude that ore body chemistry, mine dumps and tailing contaminated ponds along with the ground water movement throughout the area have an impact on the ground water and nearby river/stream contaminant chemistry in the areas.« less

  18. The utility of web mining for epidemiological research: studying the association between parity and cancer risk [Web Mining for Epidemiological Research. Assessing its Utility in Exploring the Association Between Parity and Cancer Risk

    DOE PAGES

    Tourassi, Georgia; Yoon, Hong-Jun; Xu, Songhua; ...

    2015-11-27

    Background: The World Wide Web has emerged as a powerful data source for epidemiological studies related to infectious disease surveillance. However, its potential for cancer-related epidemiological discoveries is largely unexplored. Methods: Using advanced web crawling and tailored information extraction procedures we automatically collected and analyzed the text content of 79,394 online obituary articles published between 1998 and 2014. The collected data included 51,911 cancer (27,330 breast; 9,470 lung; 6,496 pancreatic; 6,342 ovarian; 2,273 colon) and 27,483 non-cancer cases. With the derived information, we replicated a case-control study design to investigate the association between parity and cancer risk. Age-adjusted odds ratiosmore » (ORs) with 95% confidence intervals (CIs) were calculated for each cancer type and compared to those reported in large-scale epidemiological studies. Results: Parity was found to be associated with a significantly reduced risk of breast cancer (OR=0.78, 95% CI = 0.75 to 0.82), pancreatic cancer (OR=0.78, 95% CI = 0.72 to 0.83), colon cancer (OR=0.67, 95% CI = 0.60 to 0.74), and ovarian cancer (OR=0.58, 95% CI = 0.54 to 0.62). Marginal association was found for lung cancer prevalence (OR=0.87, 95% CI = 0.81 to 0.92). The linear trend between multi-parity and reduced cancer risk was dramatically more pronounced for breast and ovarian cancer than the other cancers included in the analysis. Conclusion: This large web-mining study on parity and cancer risk produced findings very similar to those reported with traditional observational studies. It may be used as a promising strategy to generate study hypotheses for guiding and prioritizing future epidemiological studies.« less

  19. The utility of web mining for epidemiological research: studying the association between parity and cancer risk [Web Mining for Epidemiological Research. Assessing its Utility in Exploring the Association Between Parity and Cancer Risk

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Tourassi, Georgia; Yoon, Hong-Jun; Xu, Songhua

    Background: The World Wide Web has emerged as a powerful data source for epidemiological studies related to infectious disease surveillance. However, its potential for cancer-related epidemiological discoveries is largely unexplored. Methods: Using advanced web crawling and tailored information extraction procedures we automatically collected and analyzed the text content of 79,394 online obituary articles published between 1998 and 2014. The collected data included 51,911 cancer (27,330 breast; 9,470 lung; 6,496 pancreatic; 6,342 ovarian; 2,273 colon) and 27,483 non-cancer cases. With the derived information, we replicated a case-control study design to investigate the association between parity and cancer risk. Age-adjusted odds ratiosmore » (ORs) with 95% confidence intervals (CIs) were calculated for each cancer type and compared to those reported in large-scale epidemiological studies. Results: Parity was found to be associated with a significantly reduced risk of breast cancer (OR=0.78, 95% CI = 0.75 to 0.82), pancreatic cancer (OR=0.78, 95% CI = 0.72 to 0.83), colon cancer (OR=0.67, 95% CI = 0.60 to 0.74), and ovarian cancer (OR=0.58, 95% CI = 0.54 to 0.62). Marginal association was found for lung cancer prevalence (OR=0.87, 95% CI = 0.81 to 0.92). The linear trend between multi-parity and reduced cancer risk was dramatically more pronounced for breast and ovarian cancer than the other cancers included in the analysis. Conclusion: This large web-mining study on parity and cancer risk produced findings very similar to those reported with traditional observational studies. It may be used as a promising strategy to generate study hypotheses for guiding and prioritizing future epidemiological studies.« less

  20. 29 CFR 570.54 - Logging occupations and occupations in the operation of any sawmill, lath mill, shingle mill, or...

    Code of Federal Regulations, 2010 CFR

    2010-07-01

    ... convenience of the user, the revised text is set forth as follows: § 570.54 Forest fire fighting and forest... performed, or mining operations. Portable sawmill shall mean a sawmilling operation where no office or...

  1. An open data mining framework for the analysis of medical images: application on obstructive nephropathy microscopy images.

    PubMed

    Doukas, Charalampos; Goudas, Theodosis; Fischer, Simon; Mierswa, Ingo; Chatziioannou, Aristotle; Maglogiannis, Ilias

    2010-01-01

    This paper presents an open image-mining framework that provides access to tools and methods for the characterization of medical images. Several image processing and feature extraction operators have been implemented and exposed through Web Services. Rapid-Miner, an open source data mining system has been utilized for applying classification operators and creating the essential processing workflows. The proposed framework has been applied for the detection of salient objects in Obstructive Nephropathy microscopy images. Initial classification results are quite promising demonstrating the feasibility of automated characterization of kidney biopsy images.

  2. A novel web informatics approach for automated surveillance of cancer mortality trends✩

    PubMed Central

    Tourassi, Georgia; Yoon, Hong-Jun; Xu, Songhua

    2016-01-01

    Cancer surveillance data are collected every year in the United States via the National Program of Cancer Registries (NPCR) and the Surveillance, Epidemiology and End Results (SEER) Program of the National Cancer Institute (NCI). General trends are closely monitored to measure the nation's progress against cancer. The objective of this study was to apply a novel web informatics approach for enabling fully automated monitoring of cancer mortality trends. The approach involves automated collection and text mining of online obituaries to derive the age distribution, geospatial, and temporal trends of cancer deaths in the US. Using breast and lung cancer as examples, we mined 23,850 cancer-related and 413,024 general online obituaries spanning the timeframe 2008–2012. There was high correlation between the web-derived mortality trends and the official surveillance statistics reported by NCI with respect to the age distribution (ρ = 0.981 for breast; ρ = 0.994 for lung), the geospatial distribution (ρ = 0.939 for breast; ρ = 0.881 for lung), and the annual rates of cancer deaths (ρ = 0.661 for breast; ρ = 0.839 for lung). Additional experiments investigated the effect of sample size on the consistency of the web-based findings. Overall, our study findings support web informatics as a promising, cost-effective way to dynamically monitor spatiotemporal cancer mortality trends. PMID:27044930

  3. A novel web informatics approach for automated surveillance of cancer mortality trends

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Tourassi, Georgia; Yoon, Hong -Jun; Xu, Songhua

    Cancer surveillance data are collected every year in the United States via the National Program of Cancer Registries (NPCR) and the Surveillance, Epidemiology and End Results (SEER) Program of the National Cancer Institute (NCI). General trends are closely monitored to measure the nation’s progress against cancer. The objective of this study was to apply a novel web informatics approach for enabling fully automated monitoring of cancer mortality trends. The approach involves automated collection and text mining of online obituaries to derive the age distribution, geospatial, and temporal trends of cancer deaths in the US. Using breast and lung cancer asmore » examples, we mined 23,850 cancer-related and 413,024 general online obituaries spanning the timeframe 2008–2012. There was high correlation between the web-derived mortality trends and the official surveillance statistics reported by NCI with respect to the age distribution (ρ = 0.981 for breast; ρ = 0.994 for lung), the geospatial distribution (ρ = 0.939 for breast; ρ = 0.881 for lung), and the annual rates of cancer deaths (ρ = 0.661 for breast; ρ = 0.839 for lung). Additional experiments investigated the effect of sample size on the consistency of the web-based findings. Altogether, our study findings support web informatics as a promising, cost-effective way to dynamically monitor spatiotemporal cancer mortality trends.« less

  4. A novel web informatics approach for automated surveillance of cancer mortality trends

    DOE PAGES

    Tourassi, Georgia; Yoon, Hong -Jun; Xu, Songhua

    2016-04-01

    Cancer surveillance data are collected every year in the United States via the National Program of Cancer Registries (NPCR) and the Surveillance, Epidemiology and End Results (SEER) Program of the National Cancer Institute (NCI). General trends are closely monitored to measure the nation’s progress against cancer. The objective of this study was to apply a novel web informatics approach for enabling fully automated monitoring of cancer mortality trends. The approach involves automated collection and text mining of online obituaries to derive the age distribution, geospatial, and temporal trends of cancer deaths in the US. Using breast and lung cancer asmore » examples, we mined 23,850 cancer-related and 413,024 general online obituaries spanning the timeframe 2008–2012. There was high correlation between the web-derived mortality trends and the official surveillance statistics reported by NCI with respect to the age distribution (ρ = 0.981 for breast; ρ = 0.994 for lung), the geospatial distribution (ρ = 0.939 for breast; ρ = 0.881 for lung), and the annual rates of cancer deaths (ρ = 0.661 for breast; ρ = 0.839 for lung). Additional experiments investigated the effect of sample size on the consistency of the web-based findings. Altogether, our study findings support web informatics as a promising, cost-effective way to dynamically monitor spatiotemporal cancer mortality trends.« less

  5. TOPSAN: a dynamic web database for structural genomics.

    PubMed

    Ellrott, Kyle; Zmasek, Christian M; Weekes, Dana; Sri Krishna, S; Bakolitsa, Constantina; Godzik, Adam; Wooley, John

    2011-01-01

    The Open Protein Structure Annotation Network (TOPSAN) is a web-based collaboration platform for exploring and annotating structures determined by structural genomics efforts. Characterization of those structures presents a challenge since the majority of the proteins themselves have not yet been characterized. Responding to this challenge, the TOPSAN platform facilitates collaborative annotation and investigation via a user-friendly web-based interface pre-populated with automatically generated information. Semantic web technologies expand and enrich TOPSAN's content through links to larger sets of related databases, and thus, enable data integration from disparate sources and data mining via conventional query languages. TOPSAN can be found at http://www.topsan.org.

  6. SA-Search: a web tool for protein structure mining based on a Structural Alphabet

    PubMed Central

    Guyon, Frédéric; Camproux, Anne-Claude; Hochez, Joëlle; Tufféry, Pierre

    2004-01-01

    SA-Search is a web tool that can be used to mine for protein structures and extract structural similarities. It is based on a hidden Markov model derived Structural Alphabet (SA) that allows the compression of three-dimensional (3D) protein conformations into a one-dimensional (1D) representation using a limited number of prototype conformations. Using such a representation, classical methods developed for amino acid sequences can be employed. Currently, SA-Search permits the performance of fast 3D similarity searches such as the extraction of exact words using a suffix tree approach, and the search for fuzzy words viewed as a simple 1D sequence alignment problem. SA-Search is available at http://bioserv.rpbs.jussieu.fr/cgi-bin/SA-Search. PMID:15215446

  7. SA-Search: a web tool for protein structure mining based on a Structural Alphabet.

    PubMed

    Guyon, Frédéric; Camproux, Anne-Claude; Hochez, Joëlle; Tufféry, Pierre

    2004-07-01

    SA-Search is a web tool that can be used to mine for protein structures and extract structural similarities. It is based on a hidden Markov model derived Structural Alphabet (SA) that allows the compression of three-dimensional (3D) protein conformations into a one-dimensional (1D) representation using a limited number of prototype conformations. Using such a representation, classical methods developed for amino acid sequences can be employed. Currently, SA-Search permits the performance of fast 3D similarity searches such as the extraction of exact words using a suffix tree approach, and the search for fuzzy words viewed as a simple 1D sequence alignment problem. SA-Search is available at http://bioserv.rpbs.jussieu.fr/cgi-bin/SA-Search.

  8. Interactive text mining with Pipeline Pilot: a bibliographic web-based tool for PubMed.

    PubMed

    Vellay, S G P; Latimer, N E Miller; Paillard, G

    2009-06-01

    Text mining has become an integral part of all research in the medical field. Many text analysis software platforms support particular use cases and only those. We show an example of a bibliographic tool that can be used to support virtually any use case in an agile manner. Here we focus on a Pipeline Pilot web-based application that interactively analyzes and reports on PubMed search results. This will be of interest to any scientist to help identify the most relevant papers in a topical area more quickly and to evaluate the results of query refinement. Links with Entrez databases help both the biologist and the chemist alike. We illustrate this application with Leishmaniasis, a neglected tropical disease, as a case study.

  9. Evaluation of longitudinal tracking and data mining for an imaging informatics-based multiple sclerosis e-folder (Conference Presentation)

    NASA Astrophysics Data System (ADS)

    Ma, Kevin C.; Forsyth, Sydney; Amezcua, Lilyana; Liu, Brent J.

    2017-03-01

    We have designed and developed a multiple sclerosis eFolder system for patient data storage, image viewing, and automatic lesion quantification results to allow patient tracking. The web-based system aims to be integrated in DICOM-compliant clinical and research environments to aid clinicians in patient treatments and data analysis. The system quantifies lesion volumes, identify and register lesion locations to track shifts in volume and quantity of lesions in a longitudinal study. We aim to evaluate the two most important features of the system, data mining and longitudinal lesion tracking, to demonstrate the MS eFolder's capability in improving clinical workflow efficiency and outcome analysis for research. In order to evaluate data mining capabilities, we have collected radiological and neurological data from 72 patients, 36 Caucasian and 36 Hispanic matched by gender, disease duration, and age. Data analysis on those patients based on ethnicity is performed, and analysis results are displayed by the system's web-based user interface. The data mining module is able to successfully separate Hispanic and Caucasian patients and compare their disease profiles. For longitudinal lesion tracking, we have collected 4 longitudinal cases and simulated different lesion growths over the next year. As a result, the eFolder is able to detect changes in lesion volume and identifying lesions with the most changes. Data mining and lesion tracking evaluation results show high potential of eFolder's usefulness in patientcare and informatics research for multiple sclerosis.

  10. Ecogeochemistry of the subsurface food web at pH 0-2.5 in Iron Mountain, California, U.S.A.

    USGS Publications Warehouse

    Robbins, E.I.; Rodgers, T.M.; Alpers, Charles N.; Nordstrom, D. Kirk

    2000-01-01

    Pyrite oxidation in the underground mining environment of Iron Mountain, California, has created the most acidic pH values ever reported in aquatic systems. Sulfate values as high as 120 000 mg l-1 and iron as high as 27 600 mg l-1 have been measured in the mine water, which also carries abundant other dissolved metals including Al, Zn, Cu, Cd, Mn, Sb and Pb. Extreme acidity and high metal concentrations apparently do not preclude the presence of an underground acidophilic food web, which has developed with bacterial biomass at the base and heliozoans as top predators. Slimes, oil-like films, flexible and inflexible stalactites, sediments, water and precipitates were found to have distinctive communities. A variety of filamentous and non-filamentous bacteria grew in slimes in water having pH values < 1.0. Fungal hyphae colonize stalactites dripping pH 1.0 water; they may help to form these drip structures. Motile hypotrichous ciliates and bdelloid rotifers are particularly abundant in slimes having a pH of 1.5. Holdfasts of the iron bacterium Leptothrix discophora attach to biofilms covering pools of standing water having a pH of 2.5 in the mine. The mine is not a closed environment - people, forced air flow and massive flushing during high intensity rainfall provide intermittent contact between the surface and underground habitats, so the mine ecosystem probably is not a restricted one.

  11. Ecogeochemistry of the subsurface food web at pH 0–2.5 in Iron Mountain, California, U.S.A.

    USGS Publications Warehouse

    Robbins, Eleanora I.; Rodgers , Teresa M.; Alpers, Charles N.; Nordstrom, D. Kirk

    2000-01-01

    Pyrite oxidation in the underground mining environment of Iron Mountain, California, has created the most acidic pH values ever reported in aquatic systems. Sulfate values as high as 120 000 mg l−1 and iron as high as 27 600 mg l−1 have been measured in the mine water, which also carries abundant other dissolved metals including Al, Zn, Cu, Cd, Mn, Sb and Pb. Extreme acidity and high metal concentrations apparently do not preclude the presence of an underground acidophilic food web, which has developed with bacterial biomass at the base and heliozoans as top predators. Slimes, oil-like films, flexible and inflexible stalactites, sediments, water and precipitates were found to have distinctive communities. A variety of filamentous and non-filamentous bacteria grew in slimes in water having pH values <1.0. Fungal hyphae colonize stalactites dripping pH 1.0 water; they may help to form these drip structures. Motile hypotrichous ciliates and bdelloid rotifers are particularly abundant in slimes having a pH of 1.5. Holdfasts of the iron bacterium Leptothrix discophora attach to biofilms covering pools of standing water having a pH of 2.5 in the mine. The mine is not a closed environment – people, forced air flow and massive flushing during high intensity rainfall provide intermittent contact between the surface and underground habitats, so the mine ecosystem probably is not a restricted one.

  12. Web-based pathology practice examination usage.

    PubMed

    Klatt, Edward C

    2014-01-01

    General and subject specific practice examinations for students in health sciences studying pathology were placed onto a free public internet web site entitled web path and were accessed four clicks from the home web site menu. Multiple choice questions were coded into. html files with JavaScript functions for web browser viewing in a timed format. A Perl programming language script with common gateway interface for web page forms scored examinations and placed results into a log file on an internet computer server. The four general review examinations of 30 questions each could be completed in up to 30 min. The 17 subject specific examinations of 10 questions each with accompanying images could be completed in up to 15 min each. The results of scores and user educational field of study from log files were compiled from June 2006 to January 2014. The four general review examinations had 31,639 accesses with completion of all questions, for a completion rate of 54% and average score of 75%. A score of 100% was achieved by 7% of users, ≥90% by 21%, and ≥50% score by 95% of users. In top to bottom web page menu order, review examination usage was 44%, 24%, 17%, and 15% of all accessions. The 17 subject specific examinations had 103,028 completions, with completion rate 73% and average score 74%. Scoring at 100% was 20% overall, ≥90% by 37%, and ≥50% score by 90% of users. The first three menu items on the web page accounted for 12.6%, 10.0%, and 8.2% of all completions, and the bottom three accounted for no more than 2.2% each. Completion rates were higher for shorter 10 questions subject examinations. Users identifying themselves as MD/DO scored higher than other users, averaging 75%. Usage was higher for examinations at the top of the web page menu. Scores achieved suggest that a cohort of serious users fully completing the examinations had sufficient preparation to use them to support their pathology education.

  13. Mining large heterogeneous data sets in drug discovery.

    PubMed

    Wild, David J

    2009-10-01

    Increasingly, effective drug discovery involves the searching and data mining of large volumes of information from many sources covering the domains of chemistry, biology and pharmacology amongst others. This has led to a proliferation of databases and data sources relevant to drug discovery. This paper provides a review of the publicly-available large-scale databases relevant to drug discovery, describes the kinds of data mining approaches that can be applied to them and discusses recent work in integrative data mining that looks for associations that pan multiple sources, including the use of Semantic Web techniques. The future of mining large data sets for drug discovery requires intelligent, semantic aggregation of information from all of the data sources described in this review, along with the application of advanced methods such as intelligent agents and inference engines in client applications.

  14. Online Health Promotion, Early Identification of Difficulties, and Help Seeking in Young People

    ERIC Educational Resources Information Center

    Santor, Darcy A.; Poulin, Christiane; LeBlanc, John C.; Kusumakar, Vivek

    2007-01-01

    Objective: To examine the use and impact of a dedicated health information Web site for adolescents. Method: Five hundred fifty-eight (27.2%) of all students in grades 7 through 12 from 4 schools logged onto the Web site; 1775 (86.4%) of all students in these grades completed a year-end health survey, with 455 (81.5%) of the students who used the…

  15. Multilingual Speech and Language Processing

    DTIC Science & Technology

    2003-04-01

    client software handles the user end of the transaction. Historically, four clients were provided: e-mail, web, FrameMaker , and command line. By...command-line client and an API. The API allows integration of CyberTrans into a number of processes including word processing packages ( FrameMaker ...preservation and logging, and others. The available clients remain e-mail, Web and FrameMaker . Platforms include both Unix and PC for clients, with

  16. Development of a Recommender System based on Personal History

    NASA Astrophysics Data System (ADS)

    Tanaka, Katsuaki; Hori, Koichi; Yamamoto, Masato

    The flood of information on the Internet makes a person who surf it without some strong intention strayed into it. One of the ways to control the balance between a person and the flood is a recommender system by computer, and many web sites use it. These systems work on a web site for the same kind of items. However the field of personal activity is not limited to handle the same kind of thing and a web site, but also offline area in the real world. To handle personal offline activities, LifeLog is proposed as method to record it, but the main purpose of LifeLog is recording a personal history. How to use a history has still been studied stage. The authors have developed a recommender system that captures personal context from history of personal online and offline activities, treats information on web sites as a large set of context, and finds out and extends overlap of them, then recommends information located there. The aim of the system is that a person can enjoy waves of information again. The system worked as a part of My-life Assist Service. It was a service for mobile phones provided by NTT DoCoMo, Inc. as a field experiment from Dec. 2007 to Feb. 2008.

  17. RPM-WEBBSYS: A web-based computer system to apply the rational polynomial method for estimating static formation temperatures of petroleum and geothermal wells

    NASA Astrophysics Data System (ADS)

    Wong-Loya, J. A.; Santoyo, E.; Andaverde, J. A.; Quiroz-Ruiz, A.

    2015-12-01

    A Web-Based Computer System (RPM-WEBBSYS) has been developed for the application of the Rational Polynomial Method (RPM) to estimate static formation temperatures (SFT) of geothermal and petroleum wells. The system is also capable to reproduce the full thermal recovery processes occurred during the well completion. RPM-WEBBSYS has been programmed using advances of the information technology to perform more efficiently computations of SFT. RPM-WEBBSYS may be friendly and rapidly executed by using any computing device (e.g., personal computers and portable computing devices such as tablets or smartphones) with Internet access and a web browser. The computer system was validated using bottomhole temperature (BHT) measurements logged in a synthetic heat transfer experiment, where a good matching between predicted and true SFT was achieved. RPM-WEBBSYS was finally applied to BHT logs collected from well drilling and shut-in operations, where the typical problems of the under- and over-estimation of the SFT (exhibited by most of the existing analytical methods) were effectively corrected.

  18. Data-driven decision support for radiologists: re-using the National Lung Screening Trial dataset for pulmonary nodule management.

    PubMed

    Morrison, James J; Hostetter, Jason; Wang, Kenneth; Siegel, Eliot L

    2015-02-01

    Real-time mining of large research trial datasets enables development of case-based clinical decision support tools. Several applicable research datasets exist including the National Lung Screening Trial (NLST), a dataset unparalleled in size and scope for studying population-based lung cancer screening. Using these data, a clinical decision support tool was developed which matches patient demographics and lung nodule characteristics to a cohort of similar patients. The NLST dataset was converted into Structured Query Language (SQL) tables hosted on a web server, and a web-based JavaScript application was developed which performs real-time queries. JavaScript is used for both the server-side and client-side language, allowing for rapid development of a robust client interface and server-side data layer. Real-time data mining of user-specified patient cohorts achieved a rapid return of cohort cancer statistics and lung nodule distribution information. This system demonstrates the potential of individualized real-time data mining using large high-quality clinical trial datasets to drive evidence-based clinical decision-making.

  19. Biomedical data mining in clinical routine: expanding the impact of hospital information systems.

    PubMed

    Müller, Marcel; Markó, Kornel; Daumke, Philipp; Paetzold, Jan; Roesner, Arnold; Klar, Rüdiger

    2007-01-01

    In this paper we want to describe how the promising technology of biomedical data mining can improve the use of hospital information systems: a large set of unstructured, narrative clinical data from a dermatological university hospital like discharge letters or other dermatological reports were processed through a morpho-semantic text retrieval engine ("MorphoSaurus") and integrated with other clinical data using a web-based interface and brought into daily clinical routine. The user evaluation showed a very high user acceptance - this system seems to meet the clinicians' requirements for a vertical data mining in the electronic patient records. What emerges is the need for integration of biomedical data mining into hospital information systems for clinical, scientific, educational and economic reasons.

  20. pubmed.mineR: an R package with text-mining algorithms to analyse PubMed abstracts.

    PubMed

    Rani, Jyoti; Shah, A B Rauf; Ramachandran, Srinivasan

    2015-10-01

    The PubMed literature database is a valuable source of information for scientific research. It is rich in biomedical literature with more than 24 million citations. Data-mining of voluminous literature is a challenging task. Although several text-mining algorithms have been developed in recent years with focus on data visualization, they have limitations such as speed, are rigid and are not available in the open source. We have developed an R package, pubmed.mineR, wherein we have combined the advantages of existing algorithms, overcome their limitations, and offer user flexibility and link with other packages in Bioconductor and the Comprehensive R Network (CRAN) in order to expand the user capabilities for executing multifaceted approaches. Three case studies are presented, namely, 'Evolving role of diabetes educators', 'Cancer risk assessment' and 'Dynamic concepts on disease and comorbidity' to illustrate the use of pubmed.mineR. The package generally runs fast with small elapsed times in regular workstations even on large corpus sizes and with compute intensive functions. The pubmed.mineR is available at http://cran.rproject. org/web/packages/pubmed.mineR.

  1. Exploration of spatio-temporal patterns of students' movement in field trip by visualizing the log data

    NASA Astrophysics Data System (ADS)

    Cho, Nahye; Kang, Youngok

    2018-05-01

    A numerous log data in addition to user input data are being generated as mobile and web users continue to increase recently, and the studies in order to explore the patterns and meanings of various movement activities by making use of these log data are also rising rapidly. On the other hand, in the field of education, people have recognized the importance of field trip as the creative education is highlighted. Also, the examples which utilize the mobile devices in the field trip in accordance to the development of information technology are growing. In this study, we try to explore the patterns of student's activity by visualizing the log data generated from high school students' field trip with mobile device.

  2. Data Mining of Network Logs

    NASA Technical Reports Server (NTRS)

    Collazo, Carlimar

    2011-01-01

    The statement of purpose is to analyze network monitoring logs to support the computer incident response team. Specifically, gain a clear understanding of the Uniform Resource Locator (URL) and its structure, and provide a way to breakdown a URL based on protocol, host name domain name, path, and other attributes. Finally, provide a method to perform data reduction by identifying the different types of advertisements shown on a webpage for incident data analysis. The procedures used for analysis and data reduction will be a computer program which would analyze the URL and identify and advertisement links from the actual content links.

  3. Optimizing Earth Data Search Ranking using Deep Learning and Real-time User Behaviour

    NASA Astrophysics Data System (ADS)

    Jiang, Y.; Yang, C. P.; Armstrong, E. M.; Huang, T.; Moroni, D. F.; McGibbney, L. J.; Greguska, F. R., III

    2017-12-01

    Finding Earth science data has been a challenging problem given both the quantity of data available and the heterogeneity of the data across a wide variety of domains. Current search engines in most geospatial data portals tend to induce end users to focus on one single data characteristic dimension (e.g., term frequency-inverse document frequency (TF-IDF) score, popularity, release date, etc.). This approach largely fails to take account of users' multidimensional preferences for geospatial data, and hence may likely result in a less than optimal user experience in discovering the most applicable dataset out of a vast range of available datasets. With users interacting with search engines, sufficient information is already hidden in the log files. Compared with explicit feedback data, information that can be derived/extracted from log files is virtually free and substantially more timely. In this dissertation, I propose an online deep learning framework that can quickly update the learning function based on real-time user clickstream data. The contributions of this framework include 1) a log processor that can ingest, process and create training data from web logs in a real-time manner; 2) a query understanding module to better interpret users' search intent using web log processing results and metadata; 3) a feature extractor that identifies ranking features representing users' multidimensional interests of geospatial data; and 4) a deep learning based ranking algorithm that can be trained incrementally using user behavior data. The search ranking results will be evaluated using precision at K and normalized discounted cumulative gain (NDCG).

  4. Tracking reflective practice-based learning by medical students during an ambulatory clerkship.

    PubMed

    Thomas, Patricia A; Goldberg, Harry

    2007-11-01

    To explore the use of web and palm digital assistant (PDA)-based patient logs to facilitate reflective learning in an ambulatory medicine clerkship. Thematic analysis of convenience sample of three successive rotations of medical students' patient log entries. Johns Hopkins University School of Medicine. MS3 and MS4 students rotating through a required block ambulatory medicine clerkship. Students are required to enter patient encounters into a web-based log system during the clerkship. Patient-linked entries included an open text field entitled, "Learning Need." Students were encouraged to use this field to enter goals for future study or teaching points related to the encounter. The logs of 59 students were examined. These students entered 3,051 patient encounters, and 51 students entered 1,347 learning need entries (44.1% of encounters). The use of the "Learning Need" field was not correlated with MS year, gender or end-of-clerkship knowledge test performance. There were strong correlations between the use of diagnostic thinking comments and observations of therapeutic relationships (Pearson's r=.42, p<0.001), and between diagnostic thinking and primary interpretation skills (Pearson's r=.60, p<0.001), but not between diagnostic thinking and factual knowledge (Pearson's r =.10, p=.46). We found that when clerkship students were cued to reflect on each patient encounter with the electronic log system, student entries grouped into categories that suggested different levels of reflective thinking. Future efforts should explore the use of such entries to encourage and track habits of reflective practice in the clinical curriculum.

  5. Microbial and geochemical assessment of bauxitic un-mined and post-mined chronosequence soils from Mocho Mountains, Jamaica.

    PubMed

    Lewis, Dawn E; Chauhan, Ashvini; White, John R; Overholt, Will; Green, Stefan J; Jasrotia, Puja; Wafula, Denis; Jagoe, Charles

    2012-10-01

    Microorganisms are very sensitive to environmental change and can be used to gauge anthropogenic impacts and even predict restoration success of degraded environments. Here, we report assessment of bauxite mining activities on soil biogeochemistry and microbial community structure using un-mined and three post-mined sites in Jamaica. The post-mined soils represent a chronosequence, undergoing restoration since 1987, 1997, and 2007. Soils were collected during dry and wet seasons and analyzed for pH, organic matter (OM), total carbon (TC), nitrogen (TN), and phosphorus. The microbial community structure was assessed through quantitative PCR and massively parallel bacterial ribosomal RNA (rRNA) gene sequencing. Edaphic factors and microbial community composition were analyzed using multivariate statistical approaches and revealed a significant, negative impact of mining on soil that persisted even after greater than 20 years of restoration. Seasonal fluctuations contributed to variation in measured soil properties and community composition, but they were minor in comparison to long-term effects of mining. In both seasons, post-mined soils were higher in pH but OM, TC, and TN decreased. Bacterial rRNA gene analyses demonstrated a general decrease in diversity in post-mined soils and up to a 3-log decrease in rRNA gene abundance. Community composition analyses demonstrated that bacteria from the Proteobacteria (α, β, γ, δ), Acidobacteria, and Firmicutes were abundant in all soils. The abundance of Firmicutes was elevated in newer post-mined soils relative to the un-mined soil, and this contrasted a decrease, relative to un-mined soils, in proteobacterial and acidobacterial rRNA gene abundances. Our study indicates long-lasting impacts of mining activities to soil biogeochemical and microbial properties with impending loss in soil productivity.

  6. Multiple-Feature Extracting Modules Based Leak Mining System Design

    PubMed Central

    Cho, Ying-Chiang; Pan, Jen-Yi

    2013-01-01

    Over the years, human dependence on the Internet has increased dramatically. A large amount of information is placed on the Internet and retrieved from it daily, which makes web security in terms of online information a major concern. In recent years, the most problematic issues in web security have been e-mail address leakage and SQL injection attacks. There are many possible causes of information leakage, such as inadequate precautions during the programming process, which lead to the leakage of e-mail addresses entered online or insufficient protection of database information, a loophole that enables malicious users to steal online content. In this paper, we implement a crawler mining system that is equipped with SQL injection vulnerability detection, by means of an algorithm developed for the web crawler. In addition, we analyze portal sites of the governments of various countries or regions in order to investigate the information leaking status of each site. Subsequently, we analyze the database structure and content of each site, using the data collected. Thus, we make use of practical verification in order to focus on information security and privacy through black-box testing. PMID:24453892

  7. Multiple-feature extracting modules based leak mining system design.

    PubMed

    Cho, Ying-Chiang; Pan, Jen-Yi

    2013-01-01

    Over the years, human dependence on the Internet has increased dramatically. A large amount of information is placed on the Internet and retrieved from it daily, which makes web security in terms of online information a major concern. In recent years, the most problematic issues in web security have been e-mail address leakage and SQL injection attacks. There are many possible causes of information leakage, such as inadequate precautions during the programming process, which lead to the leakage of e-mail addresses entered online or insufficient protection of database information, a loophole that enables malicious users to steal online content. In this paper, we implement a crawler mining system that is equipped with SQL injection vulnerability detection, by means of an algorithm developed for the web crawler. In addition, we analyze portal sites of the governments of various countries or regions in order to investigate the information leaking status of each site. Subsequently, we analyze the database structure and content of each site, using the data collected. Thus, we make use of practical verification in order to focus on information security and privacy through black-box testing.

  8. SeMPI: a genome-based secondary metabolite prediction and identification web server.

    PubMed

    Zierep, Paul F; Padilla, Natàlia; Yonchev, Dimitar G; Telukunta, Kiran K; Klementz, Dennis; Günther, Stefan

    2017-07-03

    The secondary metabolism of bacteria, fungi and plants yields a vast number of bioactive substances. The constantly increasing amount of published genomic data provides the opportunity for an efficient identification of gene clusters by genome mining. Conversely, for many natural products with resolved structures, the encoding gene clusters have not been identified yet. Even though genome mining tools have become significantly more efficient in the identification of biosynthetic gene clusters, structural elucidation of the actual secondary metabolite is still challenging, especially due to as yet unpredictable post-modifications. Here, we introduce SeMPI, a web server providing a prediction and identification pipeline for natural products synthesized by polyketide synthases of type I modular. In order to limit the possible structures of PKS products and to include putative tailoring reactions, a structural comparison with annotated natural products was introduced. Furthermore, a benchmark was designed based on 40 gene clusters with annotated PKS products. The web server of the pipeline (SeMPI) is freely available at: http://www.pharmaceutical-bioinformatics.de/sempi. © The Author(s) 2017. Published by Oxford University Press on behalf of Nucleic Acids Research.

  9. Talking with the American Public: Blogs, Facebook, and YouTube, and Public Affairs

    DTIC Science & Technology

    2009-02-20

    with news concerning Marines. News sources are not limited to news websites though. Personal websites, web logs (blogs), and social networking sites are...2.0 tools such as blogs, social networking sites , and web-article comments are conversations; they are an effective way for PA to ensure the...internet. More importantly, they get it not only from traditional news media sites, but also from social networking sites where ideas, perspectives

  10. Characterization of Geologic Structures and Host Rock Properties Relevant to the Hydrogeology of the Standard Mine in Elk Basin, Gunnison County, Colorado

    USGS Publications Warehouse

    Caine, Jonathan S.; Manning, Andrew H.; Berger, Byron R.; Kremer, Yannick; Guzman, Mario A.; Eberl, Dennis D.; Schuller, Kathryn

    2010-01-01

    The Standard Mine Superfund Site is a source of mine drainage and associated heavy metal contamination of surface and groundwaters. The site contains Tertiary polymetallic quartz veins and fault zones that host precious and base metal sulfide mineralization common in Colorado. To assist the U.S. Environmental Protection Agency in its effort to remediate mine-related contamination, we characterized geologic structures, host rocks, and their potential hydraulic properties to better understand the sources of contaminants and the local hydrogeology. Real time kinematic and handheld global positioning systems were used to locate and map precisely the geometry of the surface traces of structures and mine-related features, such as portals. New reconnaissance geologic mapping, field and x-ray diffraction mineralogy, rock sample collection, thin-section analysis, and elemental geochemical analysis were completed to characterize hydrothermal alteration, mineralization, and subsequent leaching of metallic phases. Surface and subsurface observations, fault vein and fracture network characterization, borehole geophysical logging, and mercury injection capillary entry pressure data were used to document potential controls on the hydrologic system.

  11. Diamond Eye: a distributed architecture for image data mining

    NASA Astrophysics Data System (ADS)

    Burl, Michael C.; Fowlkes, Charless; Roden, Joe; Stechert, Andre; Mukhtar, Saleem

    1999-02-01

    Diamond Eye is a distributed software architecture, which enables users (scientists) to analyze large image collections by interacting with one or more custom data mining servers via a Java applet interface. Each server is coupled with an object-oriented database and a computational engine, such as a network of high-performance workstations. The database provides persistent storage and supports querying of the 'mined' information. The computational engine provides parallel execution of expensive image processing, object recognition, and query-by-content operations. Key benefits of the Diamond Eye architecture are: (1) the design promotes trial evaluation of advanced data mining and machine learning techniques by potential new users (all that is required is to point a web browser to the appropriate URL), (2) software infrastructure that is common across a range of science mining applications is factored out and reused, and (3) the system facilitates closer collaborations between algorithm developers and domain experts.

  12. Mercury and other trace elements in a pelagic Arctic marine food web (Northwater Polynya, Baffin Bay).

    PubMed

    Campbell, Linda M; Norstrom, Ross J; Hobson, Keith A; Muir, Derek C G; Backus, Sean; Fisk, Aaron T

    2005-12-01

    Total mercury (THg), methylmercury (MeHg) and 22 other trace elements were measured in ice algae, three species of zooplankton, mixed zooplankton samples, Arctic cod (Boreogadus saida), ringed seals (Phoca hispida) and eight species of seabirds to examine the trophodynamics of these metals in an Arctic marine food web. All samples were collected in 1998 in the Northwater Polynya (NOW) located between Ellesmere Island and Greenland in Baffin Bay. THg and MeHg were found to biomagnify through the NOW food web, based on significant positive relationships between log THg and log MeHg concentrations vs. delta15N muscle and liver . The slope of these relationships for muscle THg and MeHg concentrations (slope=0.197 and 0.223, respectively) were similar to those reported for other aquatic food webs. The food web behavior of THg and delta15N appears constant, regardless of trophic state (eutrophic vs. oligotrophic), latitude (Arctic vs. tropical) or salinity (marine vs. freshwater) of the ecosystem. Rb in both liver and muscle tissue and Zn in muscle tissue were also found to biomagnify through this food web, although at a rate that is approximately 25% of that of THg. A number of elements (Cd, Pb and Ni in muscle tissue and Cd and Li in seabird liver tissue) were found to decrease trophically through the food web, as indicated by significantly negative relationships with tissue-specific delta15N. A diverse group of metals (Ag, Ba, La, Li, Sb, Sr, U and V) were found to have higher concentrations in zooplankton than seabirds or marine mammals due to bioconcentration from seawater. The remaining metals (As, Co, Cu, Ga, Mn, Mo and Se in muscle tissue) showed no relationship with trophic position, as indicated by delta15N values, although As in liver tissue showed significant biomagnification in the seabird portion of the food web.

  13. Calypso: a user-friendly web-server for mining and visualizing microbiome-environment interactions.

    PubMed

    Zakrzewski, Martha; Proietti, Carla; Ellis, Jonathan J; Hasan, Shihab; Brion, Marie-Jo; Berger, Bernard; Krause, Lutz

    2017-03-01

    Calypso is an easy-to-use online software suite that allows non-expert users to mine, interpret and compare taxonomic information from metagenomic or 16S rDNA datasets. Calypso has a focus on multivariate statistical approaches that can identify complex environment-microbiome associations. The software enables quantitative visualizations, statistical testing, multivariate analysis, supervised learning, factor analysis, multivariable regression, network analysis and diversity estimates. Comprehensive help pages, tutorials and videos are provided via a wiki page. The web-interface is accessible via http://cgenome.net/calypso/ . The software is programmed in Java, PERL and R and the source code is available from Zenodo ( https://zenodo.org/record/50931 ). The software is freely available for non-commercial users. l.krause@uq.edu.au. Supplementary data are available at Bioinformatics online. © The Author 2016. Published by Oxford University Press.

  14. Impacts of gold mine waste disposal on a tropical pelagic ecosystem.

    PubMed

    Brewer, D T; Morello, E B; Griffiths, S; Fry, G; Heales, D; Apte, S C; Venables, W N; Rothlisberg, P C; Moeseneder, C; Lansdell, M; Pendrey, R; Coman, F; Strzelecki, J; Jarolimek, C V; Jung, R F; Richardson, A J

    2012-12-01

    We used a comparative approach to investigate the impact of the disposal of gold mine tailings into the ocean near the Lihir mine (Niolam Island, Papua New Guinea). We found abundance and diversity of zooplankton, micronekton and pelagic fish to be similar or higher in the mine region compared to the reference site. We also found relatively high trace metal concentrations in lower trophic level groups, especially zooplankton, near the mine discharge, but few differences in tissue concentrations of micronekton, baitfish and pelagic fish between the two regions. Biomagnification of some trace metals by micronekton, and of mercury by fish was evident in both regions. We conclude that ocean mine waste disposal at Niolam Island has a local impact on the smaller and less mobile pelagic communities in terms of trace metal concentrations, but has little effect on the abundance and biodiversity of the local food web. Crown Copyright © 2012. Published by Elsevier Ltd. All rights reserved.

  15. Using native epiphytic ferns to estimate the atmospheric mercury levels in a small-scale gold mining area of West Java, Indonesia.

    PubMed

    Kono, Yuriko; Rahajoe, Joeni S; Hidayati, Nuril; Kodamatani, Hitoshi; Tomiyasu, Takashi

    2012-09-01

    Mercury pollution is caused by artisanal and small-scale gold mining (ASGM) operations along the Cikaniki River (West Java, Indonesia). The atmosphere is one of the primary media through which mercury can disperse. In this study, atmospheric mercury levels are estimated using the native epiphytic fern Asplenium nidus complex (A. nidus) as a biomonitor; these estimates shed light on the atmospheric dispersion of mercury released during mining. Samples were collected from 8 sites along the Cikaniki Basin during September-November, 2008 and September-November, 2009. The A. nidus fronds that were attached to tree trunks 1-3m above the ground were collected and measured for total mercury concentration using cold vapor atomic absorption spectrometry (CVAAS) after acid-digestion. The atmospheric mercury was collected using porous gold collectors, and the concentrations were determined using double-amalgam CVAAS. The highest atmospheric mercury concentration, 1.8 × 10(3) ± 1.6 × 10(3) ngm(-3), was observed at the mining hot spot, and the lowest concentration of mercury, 5.6 ± 2.0 ngm(-3), was observed at the remote site from the Cikaniki River in 2009. The mercury concentrations in A. nidus were higher at the mining village (5.4 × 10(3) ± 1.6 × 10(3) ngg(-1)) than at the remote site (70 ± 30 ngg(-1)). The distribution of mercury in A. nidus was similar to that in the atmosphere; a significant correlation was observed between the mercury concentrations in the air and in A. nidus (r=0.895, P<0.001, n=14). The mercury levels in the atmosphere can be estimated from the mercury concentration in A. nidus using a regression equation: log (Hg(A.nidu)/ngg(-1))=0.740 log (Hg(Air)/ngm (-3)) - 1.324. Copyright © 2012 Elsevier Ltd. All rights reserved.

  16. Aligning observed and modelled behaviour based on workflow decomposition

    NASA Astrophysics Data System (ADS)

    Wang, Lu; Du, YuYue; Liu, Wei

    2017-09-01

    When business processes are mostly supported by information systems, the availability of event logs generated from these systems, as well as the requirement of appropriate process models are increasing. Business processes can be discovered, monitored and enhanced by extracting process-related information. However, some events cannot be correctly identified because of the explosion of the amount of event logs. Therefore, a new process mining technique is proposed based on a workflow decomposition method in this paper. Petri nets (PNs) are used to describe business processes, and then conformance checking of event logs and process models is investigated. A decomposition approach is proposed to divide large process models and event logs into several separate parts that can be analysed independently; while an alignment approach based on a state equation method in PN theory enhances the performance of conformance checking. Both approaches are implemented in programmable read-only memory (ProM). The correctness and effectiveness of the proposed methods are illustrated through experiments.

  17. Understanding the usage of content in a mental health intervention for depression: an analysis of log data.

    PubMed

    Van Gemert-Pijnen, Julia Ewc; Kelders, Saskia M; Bohlmeijer, Ernst T

    2014-01-31

    Web-based interventions for the early treatment of depressive symptoms can be considered effective in reducing mental complaints. However, there is a limited understanding of which elements in an intervention contribute to effectiveness. For efficiency and effectiveness of interventions, insight is needed into the use of content and persuasive features. The aims of this study were (1) to illustrate how log data can be used to understand the uptake of the content of a Web-based intervention that is based on the acceptance and commitment therapy (ACT) and (2) to discover how log data can be of value for improving the incorporation of content in Web-based interventions. Data from 206 participants (out of the 239) who started the first nine lessons of the Web-based intervention, Living to the Full, were used for a secondary analysis of a subset of the log data of the parent study about adherence to the intervention. The log files used in this study were per lesson: login, start mindfulness, download mindfulness, view success story, view feedback message, start multimedia, turn on text-message coach, turn off text-message coach, and view text message. Differences in usage between lessons were explored with repeated measures ANOVAs (analysis of variance). Differences between groups were explored with one-way ANOVAs. To explore the possible predictive value of the login per lesson quartiles on the outcome measures, four linear regressions were used with login quartiles as predictor and with the outcome measures (Center for Epidemiologic Studies-Depression [CES-D] and the Hospital Anxiety and Depression Scale-Anxiety [HADS-A] on post-intervention and follow-up) as dependent variables. A significant decrease in logins and in the use of content and persuasive features over time was observed. The usage of features varied significantly during the treatment process. The usage of persuasive features increased during the third part of the ACT (commitment to value-based living), which might indicate that at that stage motivational support was relevant. Higher logins over time (9 weeks) corresponded with a higher usage of features (in most cases significant); when predicting depressive symptoms at post-intervention, the linear regression yielded a significant model with login quartile as a significant predictor (explained variance is 2.7%). A better integration of content and persuasive features in the design of the intervention and a better intra-usability of features within the system are needed to identify which combination of features works best for whom. Pattern recognition can be used to tailor the intervention based on usage patterns from the earlier lessons and to support the uptake of content essential for therapy. An adaptable interface for a modular composition of therapy features supposes a dynamic approach for Web-based treatment; not a predefined path for all, but a flexible way to go through all features that have to be used.

  18. Understanding the Usage of Content in a Mental Health Intervention for Depression: An Analysis of Log Data

    PubMed Central

    2014-01-01

    Background Web-based interventions for the early treatment of depressive symptoms can be considered effective in reducing mental complaints. However, there is a limited understanding of which elements in an intervention contribute to effectiveness. For efficiency and effectiveness of interventions, insight is needed into the use of content and persuasive features. Objective The aims of this study were (1) to illustrate how log data can be used to understand the uptake of the content of a Web-based intervention that is based on the acceptance and commitment therapy (ACT) and (2) to discover how log data can be of value for improving the incorporation of content in Web-based interventions. Methods Data from 206 participants (out of the 239) who started the first nine lessons of the Web-based intervention, Living to the Full, were used for a secondary analysis of a subset of the log data of the parent study about adherence to the intervention. The log files used in this study were per lesson: login, start mindfulness, download mindfulness, view success story, view feedback message, start multimedia, turn on text-message coach, turn off text-message coach, and view text message. Differences in usage between lessons were explored with repeated measures ANOVAs (analysis of variance). Differences between groups were explored with one-way ANOVAs. To explore the possible predictive value of the login per lesson quartiles on the outcome measures, four linear regressions were used with login quartiles as predictor and with the outcome measures (Center for Epidemiologic Studies—Depression [CES-D] and the Hospital Anxiety and Depression Scale—Anxiety [HADS-A] on post-intervention and follow-up) as dependent variables. Results A significant decrease in logins and in the use of content and persuasive features over time was observed. The usage of features varied significantly during the treatment process. The usage of persuasive features increased during the third part of the ACT (commitment to value-based living), which might indicate that at that stage motivational support was relevant. Higher logins over time (9 weeks) corresponded with a higher usage of features (in most cases significant); when predicting depressive symptoms at post-intervention, the linear regression yielded a significant model with login quartile as a significant predictor (explained variance is 2.7%). Conclusions A better integration of content and persuasive features in the design of the intervention and a better intra-usability of features within the system are needed to identify which combination of features works best for whom. Pattern recognition can be used to tailor the intervention based on usage patterns from the earlier lessons and to support the uptake of content essential for therapy. An adaptable interface for a modular composition of therapy features supposes a dynamic approach for Web-based treatment; not a predefined path for all, but a flexible way to go through all features that have to be used. PMID:24486914

  19. Facilitating Decision Making, Re-Use and Collaboration: A Knowledge Management Approach to Acquisition Program Self-Awareness

    DTIC Science & Technology

    2009-06-01

    capabilities: web-based, relational/multi-dimensional, client/server, and metadata (data about data) inclusion (pp. 39-40). Text mining, on the other...and Organizational Systems ( CASOS ) (Carley, 2005). Although AutoMap can be used to conduct text-mining, it was utilized only for its visualization...provides insight into how the GMCOI is using the terms, and where there might be redundant terms and need for de -confliction and standardization

  20. FlyMine: an integrated database for Drosophila and Anopheles genomics

    PubMed Central

    Lyne, Rachel; Smith, Richard; Rutherford, Kim; Wakeling, Matthew; Varley, Andrew; Guillier, Francois; Janssens, Hilde; Ji, Wenyan; Mclaren, Peter; North, Philip; Rana, Debashis; Riley, Tom; Sullivan, Julie; Watkins, Xavier; Woodbridge, Mark; Lilley, Kathryn; Russell, Steve; Ashburner, Michael; Mizuguchi, Kenji; Micklem, Gos

    2007-01-01

    FlyMine is a data warehouse that addresses one of the important challenges of modern biology: how to integrate and make use of the diversity and volume of current biological data. Its main focus is genomic and proteomics data for Drosophila and other insects. It provides web access to integrated data at a number of different levels, from simple browsing to construction of complex queries, which can be executed on either single items or lists. PMID:17615057

  1. LimTox: a web tool for applied text mining of adverse event and toxicity associations of compounds, drugs and genes.

    PubMed

    Cañada, Andres; Capella-Gutierrez, Salvador; Rabal, Obdulia; Oyarzabal, Julen; Valencia, Alfonso; Krallinger, Martin

    2017-07-03

    A considerable effort has been devoted to retrieve systematically information for genes and proteins as well as relationships between them. Despite the importance of chemical compounds and drugs as a central bio-entity in pharmacological and biological research, only a limited number of freely available chemical text-mining/search engine technologies are currently accessible. Here we present LimTox (Literature Mining for Toxicology), a web-based online biomedical search tool with special focus on adverse hepatobiliary reactions. It integrates a range of text mining, named entity recognition and information extraction components. LimTox relies on machine-learning, rule-based, pattern-based and term lookup strategies. This system processes scientific abstracts, a set of full text articles and medical agency assessment reports. Although the main focus of LimTox is on adverse liver events, it enables also basic searches for other organ level toxicity associations (nephrotoxicity, cardiotoxicity, thyrotoxicity and phospholipidosis). This tool supports specialized search queries for: chemical compounds/drugs, genes (with additional emphasis on key enzymes in drug metabolism, namely P450 cytochromes-CYPs) and biochemical liver markers. The LimTox website is free and open to all users and there is no login requirement. LimTox can be accessed at: http://limtox.bioinfo.cnio.es. © The Author(s) 2017. Published by Oxford University Press on behalf of Nucleic Acids Research.

  2. Physical exploration for uranium during 1951 in the Silver Reef district, Washington County, Utah

    USGS Publications Warehouse

    Stugard, Frederick

    1954-01-01

    During 1951 a joint exploration program of the most promising uraniferous areas in the Silver Reef district was made by the U.S. Geological Survey and the U.S. atomic Energy Commission. A U.S. Bureau of Mines drill crew, on contract to the Atomic Energy Commission, did 2,450 feet of diamond drilling under the geological supervision of the U.S. Geological Survey. The purpose of the drilling was to delineate broadly the favorable ground for commercial development of the uranium deposits. Ten drill holes were located around Pumpkin Point, which is the northeastern end of Buckeye Reef, to probe for extensions of small ore sheets mined on the Point in fine-grained sandstones of the Chinle formation. Three additional holes were located around Tecumseh Hill to probe for extensions of the small showings of uranium-bearing rocks of Buckeye Reef. Only one trace of uranium mineral was detected in the 13 drill holes by logging of drill cores, gamma-ray logging of the holes, and analysis of many core splits from favorable lithology. Extensive traversing with Geiger counters throughout the district and detailed geologic mapping of areas on Buckeye Reef and on East Reef indicate that the chances of discovering significant uranium deposits in the Silver Reef district are very poor, because of: highly variable lithology, closely faulted structure, and obliteration of the shallow uranium-bearing lenses by silver mining. Most of the available ore in the district was in the Pumpkin Point area and has been mined during 1950 to 1953. No ore reserves can be computed for the district before further development work. The most favorable remaining area in the district is now being explored by the operators with Atomic Energy Commission supervision.

  3. The key image and case log application: new radiology software for teaching file creation and case logging that incorporates elements of a social network.

    PubMed

    Rowe, Steven P; Siddiqui, Adeel; Bonekamp, David

    2014-07-01

    To create novel radiology key image software that is easy to use for novice users, incorporates elements adapted from social networking Web sites, facilitates resident and fellow education, and can serve as the engine for departmental sharing of interesting cases and follow-up studies. Using open-source programming languages and software, radiology key image software (the key image and case log application, KICLA) was developed. This system uses a lightweight interface with the institutional picture archiving and communications systems and enables the storage of key images, image series, and cine clips. It was designed to operate with minimal disruption to the radiologists' daily workflow. Many features of the user interface have been inspired by social networking Web sites, including image organization into private or public folders, flexible sharing with other users, and integration of departmental teaching files into the system. We also review the performance, usage, and acceptance of this novel system. KICLA was implemented at our institution and achieved widespread popularity among radiologists. A large number of key images have been transmitted to the system since it became available. After this early experience period, the most commonly encountered radiologic modalities are represented. A survey distributed to users revealed that most of the respondents found the system easy to use (89%) and fast at allowing them to record interesting cases (100%). Hundred percent of respondents also stated that they would recommend a system such as KICLA to their colleagues. The system described herein represents a significant upgrade to the Digital Imaging and Communications in Medicine teaching file paradigm with efforts made to maximize its ease of use and inclusion of characteristics inspired by social networking Web sites that allow the system additional functionality such as individual case logging. Copyright © 2014 AUR. Published by Elsevier Inc. All rights reserved.

  4. Study on Personalized Recommendation Model of Internet Advertisement

    NASA Astrophysics Data System (ADS)

    Zhou, Ning; Chen, Yongyue; Zhang, Huiping

    With the rapid development of E-Commerce, the audiences put forward higher requirements on personalized Internet advertisement than before. The main function of Personalized Advertising System is to provide the most suitable advertisements for anonymous users on Web sites. The paper offers a personalized Internet advertisement recommendation model. By mining the audiences' historical and current behavior, and the advertisers' and publisher's web site content, etc, the system can recommend appropriate advertisements to corresponding audiences.

  5. Data Mining Meets HCI: Making Sense of Large Graphs

    DTIC Science & Technology

    2012-07-01

    graph algo- rithms, won the Open Source Software World Challenge, Silver Award. We have released Pegasus as free , open-source software, downloaded by...METIS [77], spectral clustering [108], and the parameter- free “Cross-associations” (CA) [26]. Belief Propagation can also be used for clus- tering, as...number of tools have been developed to support “ landscape ” views of information. These include WebBook and Web- Forager [23], which use a book metaphor

  6. Phone and Web-Based Tobacco Cessation Treatment: Real-World Utilization Patterns and Outcomes for 11,000 Tobacco Users

    PubMed Central

    Hapgood, Jenny; Smucker Barnwell, Sara; McAfee, Tim

    2008-01-01

    Background Phone-based tobacco cessation programs have been proven effective and widely adopted. Web-based solutions exist; however, the evidence base is not yet well established. Many cessation treatments are commercially available, but few integrate the phone and Web for delivery and no published studies exist for integrated programs. Objective This paper describes a comprehensive integrated phone/Web tobacco cessation program and the characteristics, experience, and outcomes of smokers enrolled in this program from a real-world evaluation. Methods We tracked program utilization (calls completed, Web log-ins), quit status, satisfaction, and demographics of 11,143 participants who enrolled in the Free & Clear Quit For Life Program between May 2006 and October 2007. All participants received up to five proactive phone counseling sessions with Quit Coaches, unlimited access to an interactive website, up to 20 tailored emails, printed Quit Guides, and cessation medication information. The program was designed to encourage use of all program components rather than asking participants to choose which components they wanted to use while quitting. Results We found that participants tended to use phone services more than Web services. On average, participants completed 2-2.5 counseling calls and logged in to the online program 1-2 times. Women were more adherent to the overall program; women utilized Web and phone services significantly (P = .003) more than men. Older smokers (> 26 years) and moderate smokers (15-20 cigarettes/day) utilized services more (P < .001) than younger (< 26 years) and light or heavy smokers. Satisfaction with services was high (92% to 95%) and varied somewhat with Web utilization. Thirty-day quit rates at the 6-month follow-up were 41% using responder analysis and 21% using intent-to-treat analysis. Web utilization was significantly associated with increased call completion and tobacco abstinence rates at the 6-month follow-up evaluation. Conclusions This paper expands our understanding of a real-world treatment program combining two mediums, phone and Web. Greater adherence to the program, as defined by using both the phone and Web components, is associated with higher quit rates. This study has implications for reaching and treating tobacco users with an integrated phone/Web program and offers evidence regarding the effectiveness of integrated cessation programs. PMID:19017583

  7. Ultrabroadband photonic internet: safety aspects

    NASA Astrophysics Data System (ADS)

    Kalicki, Arkadiusz; Romaniuk, Ryszard

    2008-11-01

    Web applications became most popular medium in the Internet. Popularity, easiness of web application frameworks together with careless development results in high number of vulnerabilities and attacks. There are several types of attacks possible because of improper input validation. SQL injection is ability to execute arbitrary SQL queries in a database through an existing application. Cross-site scripting is the vulnerability which allows malicious web users to inject code into the web pages viewed by other users. Cross-Site Request Forgery (CSRF) is an attack that tricks the victim into loading a page that contains malicious request. Web spam in blogs. There are several techniques to mitigate attacks. Most important are web application strong design, correct input validation, defined data types for each field and parameterized statements in SQL queries. Server hardening with firewall, modern security policies systems and safe web framework interpreter configuration are essential. It is advised to keep proper security level on client side, keep updated software and install personal web firewalls or IDS/IPS systems. Good habits are logging out from services just after finishing work and using even separate web browser for most important sites, like e-banking.

  8. Prediction of the thermodynamic properties of metal-arsenate and metal-arsenite aqueous complexes to high temperatures and pressures and some geological consequences

    NASA Astrophysics Data System (ADS)

    Marini, Luigi; Accornero, Marina

    2007-07-01

    The standard thermodynamic properties at 25°C, 1 bar (Δ G {f/o}, Δ H {f/o}, S o, C {P/o}, V o, ω) and the coefficients of the revised Helgeson-Kirkham-Flowers equations of state were evaluated for several aqueous complexes formed by dissolved metals and either arsenate or arsenite ions. The guidelines of Shock and Helgeson (Geochim Cosmochim Acta 52:2009-2036, 1988) and Sverjensky et al. (Geochim Cosmochim Acta 61:1359-1412, 1997) were followed and corroborated with alternative approaches, whenever possible. The SUPCRT92 computer code was used to generate the log K of the destruction reactions of these metal-arsenate and metal-arsenite aqueous complexes at pressures and temperatures required by the EQ3/6 software package, version 7.2b. Apart from the AlAsO{4/o} and FeAsO{4/o} complexes, our log K at 25°C, 1 bar are in fair agreement with those of Whiting (MS Thesis, Colorado School of Mines, Golden, CO, 1992). Moreover, the equilibrium constants evaluated in this study are in good to fair agreement with those determined experimentally for the Ca-dihydroarsenate and Ca-hydroarsenate complexes at 40°C (Mironov et al., Russ J Inorg Chem 40:1690, 1995) and for Fe(III)-hydroarsenate complex at 25°C (Raposo et al., J Sol Chem 35:79-94, 2006), whereas the disagreement with the log K measured for the Ca-arsenate complex at 40°C (Mironov et al., Russ J Inorg Chem 40:1690, 1995) might be due to uncertainties in this measured value. The implications of aqueous complexing between dissolved metals and arsenate/arsenite ions were investigated for seawater, high-temperature geothermal liquids and acid mine drainage and aqueous solutions deriving from mixing of acid mine waters and surface waters.

  9. Protective immune response of chickens to oral vaccination with thermostable live Fowlpox virus vaccine (strain TPV-1) coated on oiled rice.

    PubMed

    Wambura, Philemon N; Godfrey, S K

    2010-03-01

    The objective of the present study was to develop and evaluate a local vaccine (strain TPV-1) against Fowl pox (FP) in chickens. Two separate groups of chickens were vaccinated with FP vaccine through oral (coated on oiled rice) and wing web stab routes, respectively. The results showed that the haemagglutination-inhibition (HI) antibody titres in both vaccinated groups were comparable and significantly higher (P < 0.05) than the control chickens. It was further revealed that 14 days after vaccination HI GMT of > or =2 log(2) was recorded in chickens vaccinated by oral and wing web stab routes whereas 35 days after vaccination the HI antibody titres reached 5.6 log(2) and 6.3 log(2), respectively. Moreover, in both groups the birds showed 100% protection against challenge virus at 35 days after vaccination. The findings from the present study have shown that oral route is equally effective as wing web stab route for vaccination of chickens against FP. However, the oral route can be used in mass vaccination of birds thus avoid catching individual birds for vaccination. It was noteworthy that strain TPV-1 virus could be propagated by a simple allantoic cavity inoculation and harvesting of allantoic fluid where it survived exposure at 57 degrees C for 2 hours. If the oral vaccination technique is optimized it may be used in controlling FP in scavenging and feral chickens. In conclusion, the present study has shown that FP vaccine (strain TPV-1) was safe, thermostable, immunogenic and efficacious in vaccinated chickens.

  10. Worldwide Research, Worldwide Participation: Web-Based Test Logger

    NASA Technical Reports Server (NTRS)

    Clark, David A.

    1998-01-01

    Thanks to the World Wide Web, a new paradigm has been born. ESCORT (steady state data system) facilities can now be configured to use a Web-based test logger, enabling worldwide participation in tests. NASA Lewis Research Center's new Web-based test logger for ESCORT automatically writes selected test and facility parameters to a browser and allows researchers to insert comments. All data can be viewed in real time via Internet connections, so anyone with a Web browser and the correct URL (universal resource locator, or Web address) can interactively participate. As the test proceeds and ESCORT data are taken, Web browsers connected to the logger are updated automatically. The use of this logger has demonstrated several benefits. First, researchers are free from manual data entry and are able to focus more on the tests. Second, research logs can be printed in report format immediately after (or during) a test. And finally, all test information is readily available to an international public.

  11. Proactive Supply Chain Performance Management with Predictive Analytics

    PubMed Central

    Stefanovic, Nenad

    2014-01-01

    Today's business climate requires supply chains to be proactive rather than reactive, which demands a new approach that incorporates data mining predictive analytics. This paper introduces a predictive supply chain performance management model which combines process modelling, performance measurement, data mining models, and web portal technologies into a unique model. It presents the supply chain modelling approach based on the specialized metamodel which allows modelling of any supply chain configuration and at different level of details. The paper also presents the supply chain semantic business intelligence (BI) model which encapsulates data sources and business rules and includes the data warehouse model with specific supply chain dimensions, measures, and KPIs (key performance indicators). Next, the paper describes two generic approaches for designing the KPI predictive data mining models based on the BI semantic model. KPI predictive models were trained and tested with a real-world data set. Finally, a specialized analytical web portal which offers collaborative performance monitoring and decision making is presented. The results show that these models give very accurate KPI projections and provide valuable insights into newly emerging trends, opportunities, and problems. This should lead to more intelligent, predictive, and responsive supply chains capable of adapting to future business environment. PMID:25386605

  12. Proactive supply chain performance management with predictive analytics.

    PubMed

    Stefanovic, Nenad

    2014-01-01

    Today's business climate requires supply chains to be proactive rather than reactive, which demands a new approach that incorporates data mining predictive analytics. This paper introduces a predictive supply chain performance management model which combines process modelling, performance measurement, data mining models, and web portal technologies into a unique model. It presents the supply chain modelling approach based on the specialized metamodel which allows modelling of any supply chain configuration and at different level of details. The paper also presents the supply chain semantic business intelligence (BI) model which encapsulates data sources and business rules and includes the data warehouse model with specific supply chain dimensions, measures, and KPIs (key performance indicators). Next, the paper describes two generic approaches for designing the KPI predictive data mining models based on the BI semantic model. KPI predictive models were trained and tested with a real-world data set. Finally, a specialized analytical web portal which offers collaborative performance monitoring and decision making is presented. The results show that these models give very accurate KPI projections and provide valuable insights into newly emerging trends, opportunities, and problems. This should lead to more intelligent, predictive, and responsive supply chains capable of adapting to future business environment.

  13. A study of medical and health queries to web search engines.

    PubMed

    Spink, Amanda; Yang, Yin; Jansen, Jim; Nykanen, Pirrko; Lorence, Daniel P; Ozmutlu, Seda; Ozmutlu, H Cenk

    2004-03-01

    This paper reports findings from an analysis of medical or health queries to different web search engines. We report results: (i). comparing samples of 10000 web queries taken randomly from 1.2 million query logs from the AlltheWeb.com and Excite.com commercial web search engines in 2001 for medical or health queries, (ii). comparing the 2001 findings from Excite and AlltheWeb.com users with results from a previous analysis of medical and health related queries from the Excite Web search engine for 1997 and 1999, and (iii). medical or health advice-seeking queries beginning with the word 'should'. Findings suggest: (i). a small percentage of web queries are medical or health related, (ii). the top five categories of medical or health queries were: general health, weight issues, reproductive health and puberty, pregnancy/obstetrics, and human relationships, and (iii). over time, the medical and health queries may have declined as a proportion of all web queries, as the use of specialized medical/health websites and e-commerce-related queries has increased. Findings provide insights into medical and health-related web querying and suggests some implications for the use of the general web search engines when seeking medical/health information.

  14. Advanced Query and Data Mining Capabilities for MaROS

    NASA Technical Reports Server (NTRS)

    Wang, Paul; Wallick, Michael N.; Allard, Daniel A.; Gladden, Roy E.; Hy, Franklin H.

    2013-01-01

    The Mars Relay Operational Service (MaROS) comprises a number of tools to coordinate, plan, and visualize various aspects of the Mars Relay network. These levels include a Web-based user interface, a back-end "ReSTlet" built in Java, and databases that store the data as it is received from the network. As part of MaROS, the innovators have developed and implemented a feature set that operates on several levels of the software architecture. This new feature is an advanced querying capability through either the Web-based user interface, or through a back-end REST interface to access all of the data gathered from the network. This software is not meant to replace the REST interface, but to augment and expand the range of available data. The current REST interface provides specific data that is used by the MaROS Web application to display and visualize the information; however, the returned information from the REST interface has typically been pre-processed to return only a subset of the entire information within the repository, particularly only the information that is of interest to the GUI (graphical user interface). The new, advanced query and data mining capabilities allow users to retrieve the raw data and/or to perform their own data processing. The query language used to access the repository is a restricted subset of the structured query language (SQL) that can be built safely from the Web user interface, or entered as freeform SQL by a user. The results are returned in a CSV (Comma Separated Values) format for easy exporting to third party tools and applications that can be used for data mining or user-defined visualization and interpretation. This is the first time that a service is capable of providing access to all cross-project relay data from a single Web resource. Because MaROS contains the data for a variety of missions from the Mars network, which span both NASA and ESA, the software also establishes an access control list (ACL) on each data record in the database repository to enforce user access permissions through a multilayered approach.

  15. MINING ENVIRONMENTAL TOXICOLOGY INFORMATION WEB RESOURCES

    EPA Science Inventory

    Environmental toxicology is the study of the ecological effects of anthropogenic substances released into the environment. It is a relatively diverse field addressing impacts to aquatic and terrestrial organisms and communities. The determination of potential risk associated with...

  16. AMP: A platform for managing and mining data in the treatment of Autism Spectrum Disorder.

    PubMed

    Linstead, Erik; Burns, Ryan; Duy Nguyen; Tyler, David

    2016-08-01

    We introduce AMP (Autism Management Platform), an integrated health care information system for capturing, analyzing, and managing data associated with the diagnosis and treatment of Autism Spectrum Disorder in children. AMP's mobile application simplifies the means by which parents, guardians, and clinicians can collect and share multimedia data with one another, facilitating communication and reducing data redundancy, while simplifying retrieval. Additionally, AMP provides an intelligent web interface and analytics platform which allow physicians and specialists to aggregate and mine patient data in real-time, as well as give relevant feedback to automatically learn data filtering preferences over time. Together AMP's mobile app, web client, and analytics engine implement a rich set of features that streamline the data collection and analysis process in the context of a secure and easy-to-use system so that data may be more effectively leveraged to guide treatment.

  17. Emergency medicine clerkship encounter and procedure logging using handheld computers.

    PubMed

    Penciner, Rick; Siddiqui, Sanam; Lee, Shirley

    2007-08-01

    Tracking medical student clinical encounters is now an accreditation requirement of medical schools. The use of handheld computers for electronic logging is emerging as a strategy to achieve this. To evaluate the technical feasibility and student satisfaction of a novel electronic logging and feedback program using handheld computers in the emergency department. This was a survey study of fourth-year medical student satisfaction with the use of their handheld computers for electronic logging of patient encounters and procedures. The authors also included an analysis of this technology. Forty-six students participated in this pilot project, logging a total of 2,930 encounters. Students used the logs an average of 7.6 shifts per rotation, logging an average of 8.3 patients per shift. Twenty-nine students (63%) responded to the survey. Students generally found it easy to complete each encounter (69%) and easy to synchronize their handheld computer with the central server (83%). However, half the students (49%) never viewed the feedback Web site and most (79%) never reviewed their logs with their preceptors. Overall, only 17% found the logging program beneficial as a learning tool. Electronic logging by medical students during their emergency medicine clerkship has many potential benefits as a method to document clinical encounters and procedures performed. However, this study demonstrated poor compliance and dissatisfaction with the process. In order for electronic logging using handheld computers to be a beneficial educational tool for both learners and educators, obstacles to effective implementation need to be addressed.

  18. From gold leaf to thermal neutrons: One hundred years of radioactivity and geological exploration (Invited)

    NASA Astrophysics Data System (ADS)

    Howarth, R. J.

    2010-12-01

    In 1789 Klaproth extracted ‘Uranit,’ from shiny black ‘Bechblende’ ore obtained from the George Wagsfort silver mine at Johanngeorgenstadt in the Erzegebirge (Ore Mountains), Saxony, Germany. He believed it to be a new chemical element (but what he had obtained was actually an oxide; uranium was first isolated in its pure metallic form by Peligot in 1856, following an earlier attempt in 1841). By 1816 pitchblende had also been found in Hungary & Cornwall, England. In1871, an English metallurgist, Richard Pearce, visiting the USA, discovered two cwt. of pitchblende ‘thrown away on a refuse-heap’ at the Wood Mine, Gilpin Co., Colorado. He returned the following year and leased the mine, which subsequently supplied the ore to MMe. Curie. In 1867, Saint-Victor had noticed the blackening of silver halide emulsion by uranium compounds, but it was Becquerel who, in a series of experiments in 1896, first showed that they emitted a radiation which was not the same as that from a Crookes tube (X-rays). In 1898 both Mme Curie and Schmidt independently described the radioactivity of thorium and its compounds but she realised, from her electrometer experiments, that uranium minerals might contain a yet more active element than either U or Th. By the end of 1900, the U-decay series had been elucidated, and during 1903-4, a radioactive gas found to exist in soil, water, air and crude petroleum, was shown to be identical with radium ‘emanation.’ Wolcott (1904) suggested that it might be used to prospect for U and Th ores. Nevertheless, for the next 25 years, investigations were largely confined to the laboratory with instrumental development and studies of radiochemisty, mineralogy, autoradiography, pleochroic haloes and contemplating radioactive heating of the Earth. However, there were exceptions: In 1905 von dem Borne used electrometer measurements to locate veins of pitchblende in a mine, and Ambronn (1921) measured the activity of successive core samples taken down an oil well to make a down-hole radioactivity profile. Technical advances were rapidly reflected in prospecting on foot, by car, and in the air, with successive adoption of the electrometer (1927); the Geiger-Müller (1945), scintillation (1952) and Hare (1954) counters; and the gamma-spectrometer (1960). The modern era of well-logging began with the patenting by Fearon in 1937 of logs using gamma rays (discovered by Viellard, 1900; named by Rutherford, 1914) and neutrons (discovered by Chadwick, 1932), although the term ‘gamma ray log’ is reported as having first been used on 29 October 1938. A simultaneous gamma and neutron logging device was developed by Sherbatskoy in 1951. Neutron-gamma and gamma-gamma logs followed in the next two years and, by the time it was possible to undertake this with a single instrument (Monaghan 1961), further tools had been developed to attempt detection of both hydrocarbons and salt water in the formations passed through. One-hundred years after Pearce’s discovery, the Thermal Neutron Decay Time Log was introduced; the marriage of radioactivity and geology had truly come of age.

  19. Keynote Talk: Mining the Web 2.0 for Improved Image Search

    NASA Astrophysics Data System (ADS)

    Baeza-Yates, Ricardo

    There are several semantic sources that can be found in the Web that are either explicit, e.g. Wikipedia, or implicit, e.g. derived from Web usage data. Most of them are related to user generated content (UGC) or what is called today the Web 2.0. In this talk we show how to use these sources of evidence in Flickr, such as tags, visual annotations or clicks, which represent the the wisdom of crowds behind UGC, to improve image search. These results are the work of the multimedia retrieval team at Yahoo! Research Barcelona and they are already being used in Yahoo! image search. This work is part of a larger effort to produce a virtuous data feedback circuit based on the right combination many different technologies to leverage the Web itself.

  20. Uncovering text mining: A survey of current work on web-based epidemic intelligence

    PubMed Central

    Collier, Nigel

    2012-01-01

    Real world pandemics such as SARS 2002 as well as popular fiction like the movie Contagion graphically depict the health threat of a global pandemic and the key role of epidemic intelligence (EI). While EI relies heavily on established indicator sources a new class of methods based on event alerting from unstructured digital Internet media is rapidly becoming acknowledged within the public health community. At the heart of automated information gathering systems is a technology called text mining. My contribution here is to provide an overview of the role that text mining technology plays in detecting epidemics and to synthesise my existing research on the BioCaster project. PMID:22783909

  1. SalanderMaps: A rapid overview about felt earthquakes through data mining of web-accesses

    NASA Astrophysics Data System (ADS)

    Kradolfer, Urs

    2013-04-01

    While seismological observatories detect and locate earthquakes based on measurements of the ground motion, they neither know a priori whether an earthquake has been felt by the public nor is it known, where it has been felt. Such information is usually gathered by evaluating feedback reported by the public through on-line forms on the web. However, after a felt earthquake in Switzerland, many people visit the webpages of the Swiss Seismological Service (SED) at the ETH Zurich and each such visit leaves traces in the logfiles on our web-servers. Data mining techniques, applied to these logfiles and mining publicly available data bases on the internet open possibilities to obtain previously unknown information about our virtual visitors. In order to provide precise information to authorities and the media, it would be desirable to rapidly know from which locations these web-accesses origin. The method 'Salander' (Seismic Activitiy Linked to Area codes - Nimble Detection of Earthquake Rumbles) will be introduced and it will be explained, how the IP-addresses (each computer or router directly connected to the internet has a unique IP-address; an example would be 129.132.53.5) of a sufficient amount of our virtual visitors were linked to their geographical area. This allows us to unprecedentedly quickly know whether and where an earthquake was felt in Switzerland. It will also be explained, why the method Salander is superior to commercial so-called geolocation products. The corresponding products of the Salander method, animated SalanderMaps, which are routinely generated after each earthquake with a magnitude of M>2 in Switzerland (http://www.seismo.ethz.ch/prod/salandermaps/, available after March 2013), demonstrate how the wavefield of earthquakes propagates through Switzerland and where it was felt. Often, such information is available within less than 60 seconds after origin time, and we always get a clear picture within already five minutes after origin time. Furthermore, the method allows to detect earthquakes solely on the analysis of accesses to our web-servers. Analyzing more than 170 million web-accesses since 2003, all seismic events within or near Switzerland with magnitudes M>4 and most felt events with magnitudes between 3 and 4 were detected. The current system is very robust, as we only had one false alarm while re-processing the web-access logfiles of the past almost 10 years. We anticipate that this method will produce even faster results in the future as the number of both commercial and private internet users is - according to the statistics of our logfiles - still increasing.

  2. Feasibility of using a web-based nutrition intervention among residents of multiethnic working-class neighborhoods.

    PubMed

    McNeill, Lorna H; Viswanath, K; Bennett, Gary G; Puleo, Elaine; Emmons, Karen M

    2007-07-01

    Using the Internet to promote behavior change is becoming more desirable as Internet use continues to increase among diverse audiences. Yet we know very little about whether this medium is useful or about different strategies to encourage Internet use by various populations. This pilot study tested the usefulness of a Web-based intervention designed to deliver nutrition-related information to and increase fruit and vegetable consumption among adults from working-class neighborhoods. Participants (N = 52) had access to the Web site for 6 weeks and received three e-mail reminders encouraging them to eat fruits and vegetables. The Web site provided information about overcoming barriers to healthy eating, accessing social support for healthy eating, setting goals for healthy eating, and maintaining a healthy diet, including recipes. We collected data on participants' use of the Web site, their Internet access and use, and their fruit and vegetable consumption. The mean age of the participants was 46 years, 73% were white, 46% did not have a college degree, and 12% had household incomes at or below 185% of the federal poverty index. They reported consuming an average of 3.4 servings of fruits and vegetables per day. More than half of the participants owned a computer, 75% logged onto the Web site at least once, and those who visited the site averaged 3.8 visits and viewed an average of 24.5 pages. The number of log-ons per day declined over the study period; however, reminder e-mails appeared to motivate participants to return to the Web site. Roughly 74% of participants viewed information on goal setting, 72% viewed information on dietary tracking, and 56% searched for main course recipes. The results of this pilot study suggest that Internet-based health messages have the potential to reach a large percentage of adults from working-class neighborhoods who have access to the Internet.

  3. Impacts of food web structure and feeding behavior on mercury exposure in Greenland Sharks (Somniosus microcephalus).

    PubMed

    McMeans, Bailey C; Arts, Michael T; Fisk, Aaron T

    2015-03-15

    Benthic and pelagic food web components in Cumberland Sound, Canada were explored as sources of total mercury (THg) to Greenland Sharks (Somniosus microcephalus) via both bottom-up food web transfer and top-down shark feeding behavior. Log10THg increased significantly with δ(15)N and trophic position from invertebrates (0.01 ± 0.01 μg · g(-1) [113 ± 1 ng · g(-1)] dw in copepods) to Greenland Sharks (3.54 ± 1.02 μg · g(-1)). The slope of the log10THg vs. δ(15)N linear regression was higher for pelagic compared to benthic food web components (excluding Greenland Sharks, which could not be assigned to either food web), which resulted from THg concentrations being higher at the base of the benthic food web (i.e., in benthic than pelagic primary consumers). However, feeding habitat is unlikely to consistently influence shark THg exposure in Cumberland Sound because THg concentrations did not consistently differ between benthic and pelagic shark prey. Further, size, gender and feeding behavior (inferred from stable isotopes and fatty acids) were unable to significantly explain THg variability among individual Greenland Sharks. Possible reasons for this result include: 1) individual sharks feeding as generalists, 2) high overlap in THg among shark prey, and 3) differences in turnover time between ecological tracers and THg. This first assessment of Greenland Shark THg within an Arctic food web revealed high concentrations consistent with biomagnification, but low ability to explain intra-specific THg variability. Our findings of high THg levels and consumption of multiple prey types, however, suggest that Greenland Sharks acquire THg through a variety of trophic pathways and are a significant contributor to the total biotic THg pool in northern seas. Copyright © 2014 Elsevier B.V. All rights reserved.

  4. Learning Relational Policies from Electronic Health Record Access Logs

    PubMed Central

    Malin, Bradley; Nyemba, Steve; Paulett, John

    2011-01-01

    Modern healthcare organizations (HCOs) are composed of complex dynamic teams to ensure clinical operations are executed in a quick and competent manner. At the same time, the fluid nature of such environments hinders administrators' efforts to define access control policies that appropriately balance patient privacy and healthcare functions. Manual efforts to define these policies are labor-intensive and error-prone, often resulting in systems that endow certain care providers with overly broad access to patients' medical records while restricting other providers from legitimate and timely use. In this work, we propose an alternative method to generate these policies by automatically mining usage patterns from electronic health record (EHR) systems. EHR systems are increasingly being integrated into clinical environments and our approach is designed to be generalizable across HCOs, thus assisting in the design and evaluation of local access control policies. Our technique, which is grounded in data mining and social network analysis theory, extracts a statistical model of the organization from the access logs of its EHRs. In doing so, our approach enables the review of predefined policies, as well as the discovery of unknown behaviors. We evaluate our approach with five months of access logs from the Vanderbilt University Medical Center and confirm the existence of stable social structures and intuitive business operations. Additionally, we demonstrate that there is significant turnover in the interactions between users in the HCO and that policies learned at the department level afford greater stability over time. PMID:21277996

  5. Integrating Data Mining in Program Evaluation of K-12 Online Education

    ERIC Educational Resources Information Center

    Hung, Jui-Long; Hsu, Yu-Chang; Rice, Kerry

    2012-01-01

    This study investigated an innovative approach of program evaluation through analyses of student learning logs, demographic data, and end-of-course evaluation surveys in an online K-12 supplemental program. The results support the development of a program evaluation model for decision making on teaching and learning at the K-12 level. A case study…

  6. Characterizing meadow vegetation with multitemporal Landsat thematic mapper remote sensing.

    Treesearch

    Alan A. Ager; Karen E. Owens

    2004-01-01

    Wet meadows are important biological components in the Blue Mountains of eastern Oregon. Many meadows in the Blue Mountains and elsewhere in the Western United States are in a state of change owing to grazing, mining, logging, road development, and other factors. This project evaluated the utility of remotely sensed data to characterize and monitor meadow vegetation...

  7. Determination of the Partition Coefficients of Organophosphorus Compounds Using High-Performance Liquid Chromatography.

    DTIC Science & Technology

    1987-12-01

    have claimed an advantage to deter- mining values of k’ in 100% aqueous mobile phases by extrapolation of linear plots of log k’ vs. percent organic...im parti- cle size chemically bonded octadecylsilane (ODS) packing ( Alltech Econo- sphere). As required, this column was saturated with I-octanol by in

  8. Geology and mining history of the Southeast Missouri Barite District and the Valles Mines, Washington, Jefferson, and St. Francois Counties, Missouri

    USGS Publications Warehouse

    Mugel, Douglas N.

    2017-03-09

    The Southeast Missouri Barite District and the Valles Mines are located in Washington, Jefferson, and St. Francois Counties, Missouri, where barite and lead ore are present together in surficial and near-surface deposits. Lead mining in the area began in the early 1700’s and extended into the early 1900’s. Hand mining of lead in the residuum resulted in widespread pits (also called shafts or diggings), and there was some underground mining of lead in bedrock. By the 1860’s barite was recovered from the residuum by hand mining, also resulting in widespread diggings, but generally not underground mines in bedrock. Mechanized open-pit mining of the residuum for barite began in the 1920’s. Barite production slowed by the 1980’s, and there has not been any barite mining since 1998. Mechanized barite mining resulted in large mined areas and tailings ponds containing waste from barite mills.The U.S. Environmental Protection Agency (EPA) has determined that lead is present in surface soils in Washington and Jefferson Counties at concentrations exceeding health-based screening levels. Also, elevated concentrations of barium, arsenic, and cadmium have been identified in surface soils, and lead concentrations exceeding the Federal drinking-water standard of 15 micrograms per liter have been identified in private drinking-water wells. Potential sources of these contaminants are wastes associated with barite mining, wastes associated with lead mining, or unmined natural deposits of barium, lead, and other metals. As a first step in helping EPA determine the source of soil and groundwater contamination, the U.S. Geological Survey (USGS), in cooperation with the EPA, investigated the geology and mining history of the Southeast Missouri Barite District and the Valles Mines.Ore minerals are barite (barium sulfate), galena (lead sulfide), cerussite (lead carbonate), anglesite (lead sulfate), sphalerite (zinc sulfide), smithsonite (zinc carbonate), and chalcopyrite (copper-iron sulfide). The Cambrian Potosi Dolomite is the most important formation for the ore deposits, followed by the Eminence Dolomite. Because galena, sphalerite, and barite are less soluble than dolomite, chemical weathering of the ore-bearing dolomite bedrock resulted in the concentration of ore minerals in the residuum. Most of the barite and lead mining was in the residuum, which averages 10 to 15 feet thick.Lead mining by French explorers may have begun in 1719 along Old Mines Creek at Cabanage de Renaudiere, which was followed shortly by the discovery of lead and the development of lead mines at Mine Renault (also called Forche a Renault Mine), Old Mines, and at other places along the Big River, Mineral Fork, and Forche a Renault Creek. Lead mining began sometime between 1775 and 1780 at Mine a Breton, the name of which was later changed to Potosi. Other mining areas were developed in the early part of the 19th century, including Fourche a Courtois (Palmer Mines), the French Diggings, and the Richwoods Mines. Zinc became a valuable resource after the Civil War, and the Valles Mines was an important supplier of zinc as well as lead, with at least some production up until the 1920’s. Lead mining declined in the early part of the 20th century as mining in the Old Lead Belt, Mine La Motte, and the Tri-State District expanded.The earliest lead mines were diggings in the residuum and were round holes (shafts) about 4 feet in diameter dug with pick and shovel about 15–20 feet deep, with drifts dug a short distance laterally from the bottom of the shafts. This mining process was repeated a short distance away until a large area was covered with pits. Some mining in bedrock began by about 1800, with shafts as deep as 170 feet and as much as several hundred feet of lateral drifts.Smelting of the lead ore to elemental lead was first done using a log furnace, which was inefficient; estimates have been made that only about 50 percent of the lead was recovered, and the remainder was lost to the ashes (slags) and to volatilization. Starting in 1798, ash furnaces were used to smelt the ashes from the log furnaces. These two furnaces were worked in tandem for many years but were gradually replaced by other furnaces, including the Scotch hearth. Estimates of lead recovery as high as 80–90 percent have been made for the Scotch hearth. By the mid-1870’s the air furnace was being used, also with estimated lead recovery as high as 80–90 percent. Zinc furnaces were built when zinc became a valuable commodity, but much of the zinc ore was shipped out of the area, either to a smelter in St. Louis, Missouri, or to other smelters.The total lead and zinc production from the Southeast Missouri Barite District and the Valles Mines is estimated at 180,000 tons of lead and 60,000 tons of zinc. An estimated 97,000 tons of lead and an estimated 120,000 tons of zinc were lost during smelting. The estimated losses do not include losses at the mine site during mining and preparation for smelting, such as the loss of fine-grained galena during hand cleaning or the discarding of zinc ore before its value was known, for which no estimates are available.Hand mining for barite in the residuum was active by at least the 1860’s and peaked from 1905 to the 1930’s when several thousand people were engaged in barite mining. Hand mining (diggings) and cleaning of the ore was done in much the same way as earlier lead mining, with the additional use of a rattle box to further clean the barite. Mechanized open-pit mining of old barite diggings began in 1924 to recover barite left behind by hand mining, and washing plants were used to clean the clay from the barite. Hand mining, however, continued to thrive, and washer plants began to close temporarily in 1931; nearly all of the barite produced before 1937 was by hand mining. By the 1940’s, however, all barite mining was mechanized.Mechanized mining used shovels powered by steam, gasoline, or electricity (and by the 1950’s draglines and front-end loaders) to mine the residuum. The ore was loaded onto rail cars (and by the 1940’s, trucks) for shipment to washer plants. Clay was removed from the barite using a log washer, and a jig was used to concentrate the barite. Overflow from the log washers was waste and went to a mud (tailings) pond. The coarse jig tailings went to tailings piles or were used as railroad ballast and, later, to create roads within the mine pit. Some barite was ground, depending on its final use, and some ground barite was bleached using a hot solution of sulfuric acid to remove impurities such as iron minerals and lead sulfide (galena). An earlier bleaching process used lead-lined tanks.Large quantities of water were required for milling the barite; some was recirculated water and the remainder came from dammed streams or was pumped from wells. Tailings and wastewater were impounded behind dikes that were built across small valleys and were increased in height as necessary using washer waste and any overburden that had been stripped. In some cases, dikes were built across valleys that had already been mined for barite.The total production of barite from the Southeast Missouri Barite District and the Valles Mines is estimated to have been about 13.1 million tons. Most of the barite production was from Washington County. Hand mining and processing of barite was inefficient. Estimates of barite recovery range from less than one-fourth to about one-half because pillars between the shafts in the residuum needed to be left unmined for stability. With mechanized mining, large amounts of barite were lost during the milling process. It has been estimated that about 30 percent of the barite was lost and that about two-thirds of the lost barite was fine-grained and was discharged to the tailings ponds. Some galena was lost to the tailings ponds.A 1972 inventory of tailings ponds by the Missouri Geological Survey identified 67 ponds in the Southeast Missouri Barite District (there are more than this currently documented). Results from samples from four ponds that were drilled were used to estimate that the 67 ponds contained almost 39 million tons (or cubic yards) of tailings averaging about 5 percent barite, for a potential reserve of 1.935 million tons of barite.It is not known how much lead was removed during barite mining, either by hand or mechanized mining and processing, how much lead was recovered, or how much lead went as fines to the tailing ponds or as coarse material to mine roads or was otherwise lost.

  9. Semantic Web Services with Web Ontology Language (OWL-S) - Specification of Agent-Services for DARPA Agent Markup Language (DAML)

    DTIC Science & Technology

    2006-08-01

    effective for describing taxonomic categories and properties of things, the structures found in SWRL and SPARQL are better suited to describing conditions...up the query processing time, which may occur many times and furthermore it is time critical. In order to maintain information about the...that time spent during this phase does not depend linearly on the number of concepts present in the data structure , but in the order of log of concepts

  10. Google Analytics: Single Page Traffic Reports

    EPA Pesticide Factsheets

    These are pages that live outside of Google Analytics (GA) but allow you to view GA data for any individual page on either the public EPA web or EPA intranet. You do need to log in to Google Analytics to view them.

  11. 78 FR 76408 - BMW of North America, LLC, Grant of Petition for Decision of Inconsequential Noncompliance

    Federal Register 2010, 2011, 2012, 2013, 2014

    2013-12-17

    .... company that manufactures and imports motor vehicles. \\2\\ BMW AG is a German company that manufactures... petition, the comment and all supporting documents log onto the Federal Docket Management System (FDMS) Web...

  12. AstrodyToolsWeb an e-Science project in Astrodynamics and Celestial Mechanics fields

    NASA Astrophysics Data System (ADS)

    López, R.; San-Juan, J. F.

    2013-05-01

    Astrodynamics Web Tools, AstrodyToolsWeb (http://tastrody.unirioja.es), is an ongoing collaborative Web Tools computing infrastructure project which has been specially designed to support scientific computation. AstrodyToolsWeb provides project collaborators with all the technical and human facilities in order to wrap, manage, and use specialized noncommercial software tools in Astrodynamics and Celestial Mechanics fields, with the aim of optimizing the use of resources, both human and material. However, this project is open to collaboration from the whole scientific community in order to create a library of useful tools and their corresponding theoretical backgrounds. AstrodyToolsWeb offers a user-friendly web interface in order to choose applications, introduce data, and select appropriate constraints in an intuitive and easy way for the user. After that, the application is executed in real time, whenever possible; then the critical information about program behavior (errors and logs) and output, including the postprocessing and interpretation of its results (graphical representation of data, statistical analysis or whatever manipulation therein), are shown via the same web interface or can be downloaded to the user's computer.

  13. The Comprehensive Microbial Resource.

    PubMed

    Peterson, J D; Umayam, L A; Dickinson, T; Hickey, E K; White, O

    2001-01-01

    One challenge presented by large-scale genome sequencing efforts is effective display of uniform information to the scientific community. The Comprehensive Microbial Resource (CMR) contains robust annotation of all complete microbial genomes and allows for a wide variety of data retrievals. The bacterial information has been placed on the Web at http://www.tigr.org/CMR for retrieval using standard web browsing technology. Retrievals can be based on protein properties such as molecular weight or hydrophobicity, GC-content, functional role assignments and taxonomy. The CMR also has special web-based tools to allow data mining using pre-run homology searches, whole genome dot-plots, batch downloading and traversal across genomes using a variety of datatypes.

  14. Predicting hospital visits from geo-tagged Internet search logs.

    PubMed

    Agarwal, Vibhu; Han, Lichy; Madan, Isaac; Saluja, Shaurya; Shidham, Aaditya; Shah, Nigam H

    2016-01-01

    The steady rise in healthcare costs has deprived over 45 million Americans of healthcare services (1, 2) and has encouraged healthcare providers to look for opportunities to improve their operational efficiency. Prior studies have shown that evidence of healthcare seeking intent in Internet searches correlates well with healthcare resource utilization. Given the ubiquitous nature of mobile Internet search, we hypothesized that analyzing geo-tagged mobile search logs could enable us to machine-learn predictors of future patient visits. Using a de-identified dataset of geo-tagged mobile Internet search logs, we mined text and location patterns that are predictors of healthcare resource utilization and built statistical models that predict the probability of a user's future visit to a medical facility. Our efforts will enable the development of innovative methods for modeling and optimizing the use of healthcare resources-a crucial prerequisite for securing healthcare access for everyone in the days to come.

  15. Human dynamics revealed through Web analytics

    NASA Astrophysics Data System (ADS)

    Gonçalves, Bruno; Ramasco, José J.

    2008-08-01

    The increasing ubiquity of Internet access and the frequency with which people interact with it raise the possibility of using the Web to better observe, understand, and monitor several aspects of human social behavior. Web sites with large numbers of frequently returning users are ideal for this task. If these sites belong to companies or universities, their usage patterns can furnish information about the working habits of entire populations. In this work, we analyze the properly anonymized logs detailing the access history to Emory University’s Web site. Emory is a medium-sized university located in Atlanta, Georgia. We find interesting structure in the activity patterns of the domain and study in a systematic way the main forces behind the dynamics of the traffic. In particular, we find that linear preferential linking, priority-based queuing, and the decay of interest for the contents of the pages are the essential ingredients to understand the way users navigate the Web.

  16. Cadmium Accumulation in Periphyton from an Abandoned Mining District in the Buffalo National River, Arkansas.

    PubMed

    McCauley, Jacob R; Bouldin, Jennifer L

    2016-06-01

    The Rush Mining District along the Buffalo River in Arkansas has a significant history of zinc and lead mining operations. The tails and spoils of these operations deposit heavy amounts of raw ore into streams. One element commonly found in the earth's crust that becomes a minor constituent of the deposition is cadmium. Periphyton samples from Rush Creek and Clabber Creek, two creeks within the Rush Mining District were measured for cadmium as well as two creeks with no history of mining, Spring Creek and Water Creek. Periphyton samples from Rush and Clabber Creek contained mean cadmium concentrations of 436.6 ± 67.3 and 93.38 ± 8.67 µg/kg, respectively. Spring Creek and Water Creek had a mean cadmium concentration of 40.49 ± 3.40 and 41.78 ± 3.99 µg/kg within periphyton. The results indicate increased metal concentrations in algal communities from mined areas. As periphyton is the base of the aquatic food chain, it acts as a conduit for movement of cadmium in the food web.

  17. Development of Database for Accident Analysis in Indian Mines

    NASA Astrophysics Data System (ADS)

    Tripathy, Debi Prasad; Guru Raghavendra Reddy, K.

    2016-10-01

    Mining is a hazardous industry and high accident rates associated with underground mining is a cause of deep concern. Technological developments notwithstanding, rate of fatal accidents and reportable incidents have not shown corresponding levels of decline. This paper argues that adoption of appropriate safety standards by both mine management and the government may result in appreciable reduction in accident frequency. This can be achieved by using the technology in improving the working conditions, sensitising workers and managers about causes and prevention of accidents. Inputs required for a detailed analysis of an accident include information on location, time, type, cost of accident, victim, nature of injury, personal and environmental factors etc. Such information can be generated from data available in the standard coded accident report form. This paper presents a web based application for accident analysis in Indian mines during 2001-2013. An accident database (SafeStat) prototype based on Intranet of the TCP/IP agreement, as developed by the authors, is also discussed.

  18. An appraisal of biological responses and network of environmental interactions in non-mining and mining impacted coastal waters.

    PubMed

    Fernandes, Christabelle E G; Malik, Ashish; Jineesh, V K; Fernandes, Sheryl O; Das, Anindita; Pandey, Sunita S; Kanolkar, Geeta; Sujith, P P; Velip, Dhillan M; Shaikh, Shagufta; Helekar, Samita; Gonsalves, Maria Judith; Nair, Shanta; LokaBharathi, P A

    2015-08-01

    The coastal waters of Goa and Ratnagiri lying on the West coast of India are influenced by terrestrial influx. However, Goa is influenced anthropogenically by iron-ore mining, while Ratnagiri is influenced by deposition of heavy minerals containing iron brought from the hinterlands. We hypothesize that there could be a shift in biological response along with changes in network of interactions between environmental and biological variables in these mining and non-mining impacted regions, lying 160 nmi apart. Biological and environmental parameters were analyzed during pre-monsoon season. Except silicates, the measured parameters were higher at Goa and related significantly, suggesting bacteria centric, detritus-driven region. At Ratnagiri, phytoplankton biomass related positively with silicate suggesting a region dominated by primary producers. This dominance perhaps got reflected as a higher tertiary yield. Thus, even though the regions are geographically proximate, the different biological response could be attributed to the differences in the web of interactions between the measured variables.

  19. A Node Linkage Approach for Sequential Pattern Mining

    PubMed Central

    Navarro, Osvaldo; Cumplido, René; Villaseñor-Pineda, Luis; Feregrino-Uribe, Claudia; Carrasco-Ochoa, Jesús Ariel

    2014-01-01

    Sequential Pattern Mining is a widely addressed problem in data mining, with applications such as analyzing Web usage, examining purchase behavior, and text mining, among others. Nevertheless, with the dramatic increase in data volume, the current approaches prove inefficient when dealing with large input datasets, a large number of different symbols and low minimum supports. In this paper, we propose a new sequential pattern mining algorithm, which follows a pattern-growth scheme to discover sequential patterns. Unlike most pattern growth algorithms, our approach does not build a data structure to represent the input dataset, but instead accesses the required sequences through pseudo-projection databases, achieving better runtime and reducing memory requirements. Our algorithm traverses the search space in a depth-first fashion and only preserves in memory a pattern node linkage and the pseudo-projections required for the branch being explored at the time. Experimental results show that our new approach, the Node Linkage Depth-First Traversal algorithm (NLDFT), has better performance and scalability in comparison with state of the art algorithms. PMID:24933123

  20. Corner-cutting mining assembly

    DOEpatents

    Bradley, J.A.

    1981-07-01

    This invention resulted from a contract with the United States Department of Energy and relates to a mining tool. More particularly, the invention relates to an assembly capable of drilling a hole having a square cross-sectional shape with radiused corners. In mining operations in which conventional auger-type drills are used to form a series of parallel, cylindrical holes in a coal seam, a large amount of coal remains in place in the seam because the shape of the holes leaves thick webs between the holes. A higher percentage of coal can be mined from a seam by a means capable of drilling holes having a substantially square cross section. It is an object of this invention to provide an improved mining apparatus by means of which the amount of coal recovered from a seam deposit can be increased. Another object of the invention is to provide a drilling assembly which cuts corners in a hole having a circular cross section. These objects and other advantages are attained by a preferred embodiment of the invention.

  1. Mercury flow through an Asian rice-based food web.

    PubMed

    Abeysinghe, Kasun S; Qiu, Guangle; Goodale, Eben; Anderson, Christopher W N; Bishop, Kevin; Evers, David C; Goodale, Morgan W; Hintelmann, Holger; Liu, Shengjie; Mammides, Christos; Quan, Rui-Chang; Wang, Jin; Wu, Pianpian; Xu, Xiao-Hang; Yang, Xiao-Dong; Feng, Xinbin

    2017-10-01

    Mercury (Hg) is a globally-distributed pollutant, toxic to humans and animals. Emissions are particularly high in Asia, and the source of exposure for humans there may also be different from other regions, including rice as well as fish consumption, particularly in contaminated areas. Yet the threats Asian wildlife face in rice-based ecosystems are as yet unclear. We sought to understand how Hg flows through rice-based food webs in historic mining and non-mining regions of Guizhou, China. We measured total Hg (THg) and methylmercury (MeHg) in soil, rice, 38 animal species (27 for MeHg) spanning multiple trophic levels, and examined the relationship between stable isotopes and Hg concentrations. Our results confirm biomagnification of THg/MeHg, with a high trophic magnification slope. Invertivorous songbirds had concentrations of THg in their feathers that were 15x and 3x the concentration reported to significantly impair reproduction, at mining and non-mining sites, respectively. High concentrations in specialist rice consumers and in granivorous birds, the later as high as in piscivorous birds, suggest rice is a primary source of exposure. Spiders had the highest THg concentrations among invertebrates and may represent a vector through which Hg is passed to vertebrates, especially songbirds. Our findings suggest there could be significant population level health effects and consequent biodiversity loss in sensitive ecosystems, like agricultural wetlands, across Asia, and invertivorous songbirds would be good subjects for further studies investigating this possibility. Copyright © 2017 Elsevier Ltd. All rights reserved.

  2. How Online Quality Ratings Influence Patients’ Choice of Medical Providers: Controlled Experimental Survey Study

    PubMed Central

    Wang, Weiguang; Gao, Guodong (Gordon); Agarwal, Ritu

    2018-01-01

    Background In recent years, the information environment for patients to learn about physician quality is being rapidly changed by Web-based ratings from both commercial and government efforts. However, little is known about how various types of Web-based ratings affect individuals’ choice of physicians. Objective The objective of this research was to measure the relative importance of Web-based quality ratings from governmental and commercial agencies on individuals’ choice of primary care physicians. Methods In a choice-based conjoint experiment conducted on a sample of 1000 Amazon Mechanical Turk users in October 2016, individuals were asked to choose their preferred primary care physician from pairs of physicians with different ratings in clinical and nonclinical aspects of care provided by governmental and commercial agencies. Results The relative log odds of choosing a physician increases by 1.31 (95% CI 1.26-1.37; P<.001) and 1.32 (95% CI 1.27-1.39; P<.001) units when the government clinical ratings and commercial nonclinical ratings move from 2 to 4 stars, respectively. The relative log odds of choosing a physician increases by 1.12 (95% CI 1.07-1.18; P<.001) units when the commercial clinical ratings move from 2 to 4 stars. The relative log odds of selecting a physician with 4 stars in nonclinical ratings provided by the government is 1.03 (95% CI 0.98-1.09; P<.001) units higher than a physician with 2 stars in this rating. The log odds of selecting a physician with 4 stars in nonclinical government ratings relative to a physician with 2 stars is 0.23 (95% CI 0.13-0.33; P<.001) units higher for females compared with males. Similar star increase in nonclinical commercial ratings increases the relative log odds of selecting the physician by female respondents by 0.15 (95% CI 0.04-0.26; P=.006) units. Conclusions Individuals perceive nonclinical ratings provided by commercial websites as important as clinical ratings provided by government websites when choosing a primary care physician. There are significant gender differences in how the ratings are used. More research is needed on whether patients are making the best use of different types of ratings, as well as the optimal allocation of resources in improving physician ratings from the government’s perspective. PMID:29581091

  3. Utilization of a Web-Based vs Integrated Phone/Web Cessation Program Among 140,000 Tobacco Users: An Evaluation Across 10 Free State Quitlines

    PubMed Central

    Vickerman, Katrina A; Kellogg, Elizabeth S; Zbikowski, Susan M

    2015-01-01

    Background Phone-based tobacco cessation program effectiveness has been established and randomized controlled trials have provided some support for Web-based services. Relatively little is known about who selects different treatment modalities and how they engage with treatments in a real-world setting. Objective This paper describes the characteristics, Web utilization patterns, and return rates of tobacco users who self-selected into a Web-based (Web-Only) versus integrated phone/Web (Phone/Web) cessation program. Methods We examined the demographics, baseline tobacco use, Web utilization patterns, and return rates of 141,429 adult tobacco users who self-selected into a Web-Only or integrated Phone/Web cessation program through 1 of 10 state quitlines from August 2012 through July 2013. For each state, registrants were only included from the timeframe in which both programs were offered to all enrollees. Utilization data were limited to site interactions occurring within 6 months after registration. Results Most participants selected the Phone/Web program (113,019/141,429, 79.91%). After enrollment in Web services, Web-Only were more likely to log in compared to Phone/Web (21,832/28,410, 76.85% vs 23,920/56,892, 42.04%; P<.001), but less likely to return after their initial log-in (8766/21,832, 40.15% vs 13,966/23,920, 58.39%; P<.001). In bivariate and multivariable analyses, those who chose Web-Only were younger, healthier, more highly educated, more likely to be uninsured or commercially insured, more likely to be white non-Hispanic and less likely to be black non-Hispanic, less likely to be highly nicotine-addicted, and more likely to have started their program enrollment online (all P<.001). Among both program populations, participants were more likely to return to Web services if they were women, older, more highly educated, or were sent nicotine replacement therapy (NRT) through their quitline (all P<.001). Phone/Web were also more likely to return if they had completed a coaching call, identified as white non-Hispanic or “other” race, or were commercially insured (all P<.001). Web-Only were less likely to return if they started their enrollment online versus via phone. The interactive Tobacco Tracker, Cost Savings Calculator, and Quitting Plan were the most widely used features overall. Web-Only were more likely than Phone/Web to use most key features (all P<.001), most notably the 5 Quitting Plan behaviors. Among quitlines that offered NRT to both Phone/Web and Web-Only, Web-Only were less likely to have received quitline NRT. Conclusions This paper adds to our understanding of who selects different cessation treatment modalities and how they engage with the program in a real-world setting. Web-Only were younger, healthier smokers of higher socioeconomic status who interacted more intensely with services in a single session, but were less likely to re-engage or access NRT benefits. Further research should examine the efficacy of different engagement techniques and services with different subpopulations of tobacco users. PMID:25673013

  4. Utilization of a Web-based vs integrated phone/Web cessation program among 140,000 tobacco users: an evaluation across 10 free state quitlines.

    PubMed

    Nash, Chelsea M; Vickerman, Katrina A; Kellogg, Elizabeth S; Zbikowski, Susan M

    2015-02-04

    Phone-based tobacco cessation program effectiveness has been established and randomized controlled trials have provided some support for Web-based services. Relatively little is known about who selects different treatment modalities and how they engage with treatments in a real-world setting. This paper describes the characteristics, Web utilization patterns, and return rates of tobacco users who self-selected into a Web-based (Web-Only) versus integrated phone/Web (Phone/Web) cessation program. We examined the demographics, baseline tobacco use, Web utilization patterns, and return rates of 141,429 adult tobacco users who self-selected into a Web-Only or integrated Phone/Web cessation program through 1 of 10 state quitlines from August 2012 through July 2013. For each state, registrants were only included from the timeframe in which both programs were offered to all enrollees. Utilization data were limited to site interactions occurring within 6 months after registration. Most participants selected the Phone/Web program (113,019/141,429, 79.91%). After enrollment in Web services, Web-Only were more likely to log in compared to Phone/Web (21,832/28,410, 76.85% vs 23,920/56,892, 42.04%; P<.001), but less likely to return after their initial log-in (8766/21,832, 40.15% vs 13,966/23,920, 58.39%; P<.001). In bivariate and multivariable analyses, those who chose Web-Only were younger, healthier, more highly educated, more likely to be uninsured or commercially insured, more likely to be white non-Hispanic and less likely to be black non-Hispanic, less likely to be highly nicotine-addicted, and more likely to have started their program enrollment online (all P<.001). Among both program populations, participants were more likely to return to Web services if they were women, older, more highly educated, or were sent nicotine replacement therapy (NRT) through their quitline (all P<.001). Phone/Web were also more likely to return if they had completed a coaching call, identified as white non-Hispanic or "other" race, or were commercially insured (all P<.001). Web-Only were less likely to return if they started their enrollment online versus via phone. The interactive Tobacco Tracker, Cost Savings Calculator, and Quitting Plan were the most widely used features overall. Web-Only were more likely than Phone/Web to use most key features (all P<.001), most notably the 5 Quitting Plan behaviors. Among quitlines that offered NRT to both Phone/Web and Web-Only, Web-Only were less likely to have received quitline NRT. This paper adds to our understanding of who selects different cessation treatment modalities and how they engage with the program in a real-world setting. Web-Only were younger, healthier smokers of higher socioeconomic status who interacted more intensely with services in a single session, but were less likely to re-engage or access NRT benefits. Further research should examine the efficacy of different engagement techniques and services with different subpopulations of tobacco users.

  5. New Science Teachers' Descriptions of Inquiry Enactment

    ERIC Educational Resources Information Center

    Dreon, Oliver, Jr.

    2008-01-01

    This phenomenological study demonstrates the influence that affective factors have on beginning teachers' ability to enact instructional practices. Through narratives shared in interviews and web log postings, two beginning science teachers' emotional engagement with their instructional practices, especially that of implementing inquiry-based…

  6. Log analysis of six boreholes in conjunction with geologic characterization above and on top of the Weeks Island Salt Dome

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Sattler, A.R.

    1996-06-01

    Six boreholes were drilled during the geologic characterization and diagnostics of the Weeks Island sinkhole that is over the two-tiered salt mine which was converted for oil storage by the U.S. Strategic Petroleum Reserve. These holes were drilled to provide for geologic characterization of the Weeks Island Salt Dome and its overburden in the immediate vicinity of the sinkhole (mainly through logs and core); to establish a crosswell configuration for seismic tomography; to establish locations for hydrocarbon detection and tracer injection; and to provide direct observations of sinkhole geometry and material properties. Specific objectives of the logging program were to:more » (1) identify the top of and the physical state of the salt dome; (2) identify the water table; (3) obtain a relative salinity profile in the aquifer within the alluvium, which ranges from the water table directly to the top of the Weeks Island salt dome; and (4) identify a reflecting horizon seen on seismic profiles over this salt dome. Natural gamma, neutron, density, sonic, resistivity and caliper logs were run.« less

  7. Social Web mining and exploitation for serious applications: Technosocial Predictive Analytics and related technologies for public health, environmental and national security surveillance

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Kamel Boulos, Maged; Sanfilippo, Antonio P.; Corley, Courtney D.

    2010-03-17

    This paper explores techno-social predictive analytics (TPA) and related methods for Web “data mining” where users’ posts and queries are garnered from Social Web (“Web 2.0”) tools such as blogs, microblogging and social networking sites to form coherent representations of real-time health events. The paper includes a brief introduction to commonly used Social Web tools such as mashups and aggregators, and maps their exponential growth as an open architecture of participation for the masses and an emerging way to gain insight about people’s collective health status of whole populations. Several health related tool examples are described and demonstrated as practicalmore » means through which health professionals might create clear location specific pictures of epidemiological data such as flu outbreaks.« less

  8. Trophic magnification of PCBs and its relationship to the octanol-water partition coefficient

    USGS Publications Warehouse

    Walters, D.M.; Mills, M.A.; Cade, B.S.; Burkard, L.P.

    2011-01-01

    We investigated polychlorinated biphenyl (PCB) bioaccumulation relative to octanol-water partition coefficient (KOW) and organism trophic position (TP) at the Lake Hartwell Superfund site (South Carolina). We measured PCBs (127 congeners) and stable isotopes (??15N) in sediment, organic matter, phytoplankton, zooplankton, macroinvertebrates, and fish. TP, as calculated from ??15N, was significantly, positively related to PCB concentrations, and food web trophic magnification factors (TMFs) ranged from 1.5-6.6 among congeners. TMFs of individual congeners increased strongly with log KOW, as did the predictive power (r2) of individual TP-PCB regression models used to calculate TMFs. We developed log KOW-TMF models for eight food webs with vastly different environments (freshwater, marine, arctic, temperate) and species composition (cold- vs warmblooded consumers). The effect of KOW on congener TMFs varied strongly across food webs (model slopes 0.0-15.0) because the range of TMFs among studies was also highly variable. We standardized TMFs within studies to mean = 0, standard deviation (SD) = 1 to normalize for scale differences and found a remarkably consistent KOW effect on TMFs (no difference in model slopes among food webs). Our findings underscore the importance of hydrophobicity (as characterized by KOW) in regulating bioaccumulation of recalcitrant compounds in aquatic systems, and demonstrate that relationships between chemical KOW and bioaccumulation from field studies are more generalized than previously recognized. ?? This article not subject to U.S. Copyright. Published 2011 by the American Chemical Society.

  9. Trophic magnification of PCBs and Its relationship to the octanol-water partition coefficient.

    PubMed

    Walters, David M; Mills, Marc A; Cade, Brian S; Burkard, Lawrence P

    2011-05-01

    We investigated polychlorinated biphenyl (PCB) bioaccumulation relative to octanol-water partition coefficient (K(OW)) and organism trophic position (TP) at the Lake Hartwell Superfund site (South Carolina). We measured PCBs (127 congeners) and stable isotopes (δ¹⁵N) in sediment, organic matter, phytoplankton, zooplankton, macroinvertebrates, and fish. TP, as calculated from δ¹⁵N, was significantly, positively related to PCB concentrations, and food web trophic magnification factors (TMFs) ranged from 1.5-6.6 among congeners. TMFs of individual congeners increased strongly with log K(OW), as did the predictive power (r²) of individual TP-PCB regression models used to calculate TMFs. We developed log K(OW)-TMF models for eight food webs with vastly different environments (freshwater, marine, arctic, temperate) and species composition (cold- vs warmblooded consumers). The effect of K(OW) on congener TMFs varied strongly across food webs (model slopes 0.0-15.0) because the range of TMFs among studies was also highly variable. We standardized TMFs within studies to mean = 0, standard deviation (SD) = 1 to normalize for scale differences and found a remarkably consistent K(OW) effect on TMFs (no difference in model slopes among food webs). Our findings underscore the importance of hydrophobicity (as characterized by K(OW)) in regulating bioaccumulation of recalcitrant compounds in aquatic systems, and demonstrate that relationships between chemical K(OW) and bioaccumulation from field studies are more generalized than previously recognized.

  10. Explanation of fields used in the Alaska Resource Data File of mines, prospects, and mineral occurrences in Alaska

    USGS Publications Warehouse

    ,

    1996-01-01

    Descriptions of mines, prospects, and mineral occurrences in the Alaska Resource Data File (ARDF) are published for individual U.S. Geological Survey 1:250,000 scale quadrangles in Alaska (see accompanying map) and are available for downloading from USGS World Wide Web site: http://www-rnrs-ak.wr.usgs.gov/ardf.These descriptions are divided into a number of fields which describe features of each mine, prospect, or mineral occurrence. These descriptions were complied from published literature and from unpublished reports and data from industry, the U.S. Bureau of Mines, and the U.S. Geological Survey and other sources. Compilation of this database is an ongoing process and each report is essentially a progress report. The authors of the individual quadrangle reports would appreciate any corrections or additional information that users may be able to contribute.

  11. Hymenoptera Genome Database: integrating genome annotations in HymenopteraMine

    PubMed Central

    Elsik, Christine G.; Tayal, Aditi; Diesh, Colin M.; Unni, Deepak R.; Emery, Marianne L.; Nguyen, Hung N.; Hagen, Darren E.

    2016-01-01

    We report an update of the Hymenoptera Genome Database (HGD) (http://HymenopteraGenome.org), a model organism database for insect species of the order Hymenoptera (ants, bees and wasps). HGD maintains genomic data for 9 bee species, 10 ant species and 1 wasp, including the versions of genome and annotation data sets published by the genome sequencing consortiums and those provided by NCBI. A new data-mining warehouse, HymenopteraMine, based on the InterMine data warehousing system, integrates the genome data with data from external sources and facilitates cross-species analyses based on orthology. New genome browsers and annotation tools based on JBrowse/WebApollo provide easy genome navigation, and viewing of high throughput sequence data sets and can be used for collaborative genome annotation. All of the genomes and annotation data sets are combined into a single BLAST server that allows users to select and combine sequence data sets to search. PMID:26578564

  12. Using client-side event logging and path tracing to assess and improve the quality of web-based surveys.

    PubMed

    White, Thomas M; Hauan, Michael J

    2002-01-01

    Web-based data collection has considerable appeal. However, the quality of data collected using such instruments is often questionable. There can be systematic problems with the wording of the surveys, and/or the means with which they are deployed. In unsupervised data collection, there are also concerns about whether subjects understand the questions, and wehther they are answering honestly. This paper presents a schema for using client-side timestamps and traces of subjects' paths through instruments to detect problems with the definition of instruments and their deployment. We discuss two large, anonymous, web-based, medical surveys as examples of the utility of this approach.

  13. Students' Navigational Pattern and Performance in an E-Learning Environment: A Case from UP Open University, Philippines

    ERIC Educational Resources Information Center

    Bagarinao, Ricardo T.

    2015-01-01

    The study analyzed the navigational patterns of learners in an online course in Science, Technology, and Society using movement ecological concept. The course site consists of five important pages, namely: home page, resource page, user page, forum page, forum discussion page, and forum add post page. About 11,413 logged data were mined and…

  14. Women Veterans’ Experience With a Web-Based Diabetes Prevention Program: A Qualitative Study to Inform Future Practice

    PubMed Central

    Ertl, Kristyn; Schneider, Jessica; Vasti, Elena; Makki, Fatima; Richardson, Caroline; Havens, Kathryn; Damschroder, Laura

    2015-01-01

    Background Diabetes prevention is a national goal and particularly important in the Veterans Health Administration (VHA) where 1 in 4 veterans has diabetes. There is growing evidence to support the use of Web-based diabetes prevention program (DPP) interventions, shown to be as effective and often more feasible than in-person interventions. Objective Our primary objective was to qualitatively explore women veterans’ early experiences with a Web-based DPP intervention. Our secondary objective was to estimate weight loss, participation, and engagement to provide context for our qualitative findings. Methods We conducted and analyzed semistructured interviews and collected data on weight change, participation, and engagement. A total of 17 women veterans with prediabetes from a Midwest VA Women’s Health Clinic were eligible to participate; 15 completed interviews. Results Participants perceived the DPP program as an appealing way of initiating lifestyle changes and made them feel accountable in achieving their daily goals. The online program was convenient because it could be accessed at any time, and many found that it integrated well into daily life. However, some did not like the logging aspect and some found it to be too impersonal. Participants logged in a mean 76 times, posted a mean 46 group messages, and sent a mean 20.5 private messages to the health coach over 16 weeks. Participants lost 5.24% of baseline weight, and 82% (14/17) of participants completed at least 9 of 16 core modules. Conclusions Women veterans’ early experiences with a Web-based DPP intervention were generally positive. Accountability and convenience were key enabling factors for participation and engagement. A Web-based DPP intervention appears to be a promising means of translating the DPP for women veterans with prediabetes. PMID:26006697

  15. The wired patient: patterns of electronic patient portal use among patients with cardiac disease or diabetes.

    PubMed

    Jones, James Brian; Weiner, Jonathan P; Shah, Nirav R; Stewart, Walter F

    2015-02-20

    As providers develop an electronic health record-based infrastructure, patients are increasingly using Web portals to access their health information and participate electronically in the health care process. Little is known about how such portals are actually used. In this paper, our goal was to describe the types and patterns of portal users in an integrated delivery system. We analyzed 12 months of data from Web server log files on 2282 patients using a Web-based portal to their electronic health record (EHR). We obtained data for patients with cardiovascular disease and/or diabetes who had a Geisinger Clinic primary care provider and were registered "MyGeisinger" Web portal users. Hierarchical cluster analysis was applied to longitudinal data to profile users based on their frequency, intensity, and consistency of use. User types were characterized by basic demographic data from the EHR. We identified eight distinct portal user groups. The two largest groups (41.98%, 948/2258 and 24.84%, 561/2258) logged into the portal infrequently but had markedly different levels of engagement with their medical record. Other distinct groups were characterized by tracking biometric measures (10.54%, 238/2258), sending electronic messages to their provider (9.25%, 209/2258), preparing for an office visit (5.98%, 135/2258), and tracking laboratory results (4.16%, 94/2258). There are naturally occurring groups of EHR Web portal users within a population of adult primary care patients with chronic conditions. More than half of the patient cohort exhibited distinct patterns of portal use linked to key features. These patterns of portal access and interaction provide insight into opportunities for electronic patient engagement strategies.

  16. Characterization of airborne particles in an open pit mining region.

    PubMed

    Huertas, José I; Huertas, María E; Solís, Dora A

    2012-04-15

    We characterized airborne particle samples collected from 15 stations in operation since 2007 in one of the world's largest opencast coal mining regions. Using gravimetric, scanning electron microscopy (SEM-EDS), and X-ray photoelectron spectroscopy (XPS) analysis the samples were characterized in terms of concentration, morphology, particle size distribution (PSD), and elemental composition. All of the total suspended particulate (TSP) samples exhibited a log-normal PSD with a mean of d=5.46 ± 0.32 μm and σ(ln d)=0.61 ± 0.03. Similarly, all particles with an equivalent aerodynamic diameter less than 10 μm (PM(10)) exhibited a log-normal type distribution with a mean of d=3.6 ± 0.38 μm and σ(ln d)=0.55 ± 0.03. XPS analysis indicated that the main elements present in the particles were carbon, oxygen, potassium, and silicon with average mass concentrations of 41.5%, 34.7%, 11.6%, and 5.7% respectively. In SEM micrographs the particles appeared smooth-surfaced and irregular in shape, and tended to agglomerate. The particles were typically clay minerals, including limestone, calcite, quartz, and potassium feldspar. Copyright © 2012 Elsevier B.V. All rights reserved.

  17. ASCOT: a text mining-based web-service for efficient search and assisted creation of clinical trials

    PubMed Central

    2012-01-01

    Clinical trials are mandatory protocols describing medical research on humans and among the most valuable sources of medical practice evidence. Searching for trials relevant to some query is laborious due to the immense number of existing protocols. Apart from search, writing new trials includes composing detailed eligibility criteria, which might be time-consuming, especially for new researchers. In this paper we present ASCOT, an efficient search application customised for clinical trials. ASCOT uses text mining and data mining methods to enrich clinical trials with metadata, that in turn serve as effective tools to narrow down search. In addition, ASCOT integrates a component for recommending eligibility criteria based on a set of selected protocols. PMID:22595088

  18. ASCOT: a text mining-based web-service for efficient search and assisted creation of clinical trials.

    PubMed

    Korkontzelos, Ioannis; Mu, Tingting; Ananiadou, Sophia

    2012-04-30

    Clinical trials are mandatory protocols describing medical research on humans and among the most valuable sources of medical practice evidence. Searching for trials relevant to some query is laborious due to the immense number of existing protocols. Apart from search, writing new trials includes composing detailed eligibility criteria, which might be time-consuming, especially for new researchers. In this paper we present ASCOT, an efficient search application customised for clinical trials. ASCOT uses text mining and data mining methods to enrich clinical trials with metadata, that in turn serve as effective tools to narrow down search. In addition, ASCOT integrates a component for recommending eligibility criteria based on a set of selected protocols.

  19. Can abstract screening workload be reduced using text mining? User experiences of the tool Rayyan.

    PubMed

    Olofsson, Hanna; Brolund, Agneta; Hellberg, Christel; Silverstein, Rebecca; Stenström, Karin; Österberg, Marie; Dagerhamn, Jessica

    2017-09-01

    One time-consuming aspect of conducting systematic reviews is the task of sifting through abstracts to identify relevant studies. One promising approach for reducing this burden uses text mining technology to identify those abstracts that are potentially most relevant for a project, allowing those abstracts to be screened first. To examine the effectiveness of the text mining functionality of the abstract screening tool Rayyan. User experiences were collected. Rayyan was used to screen abstracts for 6 reviews in 2015. After screening 25%, 50%, and 75% of the abstracts, the screeners logged the relevant references identified. A survey was sent to users. After screening half of the search result with Rayyan, 86% to 99% of the references deemed relevant to the study were identified. Of those studies included in the final reports, 96% to 100% were already identified in the first half of the screening process. Users rated Rayyan 4.5 out of 5. The text mining function in Rayyan successfully helped reviewers identify relevant studies early in the screening process. Copyright © 2017 John Wiley & Sons, Ltd.

  20. Hydrologic data for Leviathan Mine and vicinity, Alpine County, California, 1981-83

    USGS Publications Warehouse

    Hammermeister, D.P.; Walmsley, S.J.

    1985-01-01

    The U.S. Geological Survey collected basic hydrologic and water-quality data during 1981-83 to facilitate the geohydrologic evaluation of the Leviathan Mine area and the design of a pollution-abatement project. Surface-water field data included one or more measurements of pH, water temperature, and specific conductance at 45 sites in and adjacent to the mine area. At nine of these sites, daily data on discharge, specific conductance, and water temperature were collected during parts of 1981-82 by using electronic monitor-recorder systems. Ground-water field data included one or more of the water-quality measurements listed above at 71 piezometers in the mine area. Borehole geophysical data included neutron-moisture, neutron-porosity, gamma-gamma density, natural gamma, and temperature logs at three sites. Mineralogic and hydrologic data were obtained for cores taken from nine test holes. One or more surface-water samples from 26 sites were analyzed for major cations, major anions, and a wide range of minor inorganic constituents. Single ground-water samples from 36 piezometers were analyzed for the same array of major and minor constituents. (USGS)

  1. A Novel Framework for Medical Web Information Foraging Using Hybrid ACO and Tabu Search.

    PubMed

    Drias, Yassine; Kechid, Samir; Pasi, Gabriella

    2016-01-01

    We present in this paper a novel approach based on multi-agent technology for Web information foraging. We proposed for this purpose an architecture in which we distinguish two important phases. The first one is a learning process for localizing the most relevant pages that might interest the user. This is performed on a fixed instance of the Web. The second takes into account the openness and dynamicity of the Web. It consists on an incremental learning starting from the result of the first phase and reshaping the outcomes taking into account the changes that undergoes the Web. The system was implemented using a colony of artificial ants hybridized with tabu search in order to achieve more effectiveness and efficiency. To validate our proposal, experiments were conducted on MedlinePlus, a real website dedicated for research in the domain of Health in contrast to other previous works where experiments were performed on web logs datasets. The main results are promising either for those related to strong Web regularities and for the response time, which is very short and hence complies the real time constraint.

  2. Adjusting game difficulty level through Formal Concept Analysis

    NASA Astrophysics Data System (ADS)

    Gómez-Martín, Marco A.; Gómez-Martín, Pedro P.; Gonzâlez-Calero, Pedro A.; Díaz-Agudo, Belén

    In order to reach as many players as possible, videogames usually allow the user to choose the difficulty level. To do it, game designers have to decide the values that some game parameters will have depending on that decision. In simple videogames this is almost trivial: minesweeper is harder with longer board sizes and number of mines. In more complex games, game designers may take advantage of data mining to establish which of all the possible parameters will affect positively to the player experience. This paper describes the use of Formal Concept Analysis to help to balance the game using the logs obtained in the tests made prior the release of the game.

  3. Public health, GIS, and the internet.

    PubMed

    Croner, Charles M

    2003-01-01

    Internet access and use of georeferenced public health information for GIS application will be an important and exciting development for the nation's Department of Health and Human Services and other health agencies in this new millennium. Technological progress toward public health geospatial data integration, analysis, and visualization of space-time events using the Web portends eventual robust use of GIS by public health and other sectors of the economy. Increasing Web resources from distributed spatial data portals and global geospatial libraries, and a growing suite of Web integration tools, will provide new opportunities to advance disease surveillance, control, and prevention, and insure public access and community empowerment in public health decision making. Emerging supercomputing, data mining, compression, and transmission technologies will play increasingly critical roles in national emergency, catastrophic planning and response, and risk management. Web-enabled public health GIS will be guided by Federal Geographic Data Committee spatial metadata, OpenGIS Web interoperability, and GML/XML geospatial Web content standards. Public health will become a responsive and integral part of the National Spatial Data Infrastructure.

  4. 77 FR 63415 - BMW of North America, LLC, a Subsidiary of BMW AG, Receipt of Petition for Decision of...

    Federal Register 2010, 2011, 2012, 2013, 2014

    2012-10-16

    ... imports motor vehicles. \\2\\ BMW AG, is a German company that manufactures motor vehicles. Pursuant to 49 U.... Electronically: by logging onto the Federal Docket Management System (FDMS) Web site at http://www.regulations...

  5. The Voice of Chinese Health Consumers: A Text Mining Approach to Web-Based Physician Reviews

    PubMed Central

    Zhang, Kunpeng

    2016-01-01

    Background Many Web-based health care platforms allow patients to evaluate physicians by posting open-end textual reviews based on their experiences. These reviews are helpful resources for other patients to choose high-quality doctors, especially in countries like China where no doctor referral systems exist. Analyzing such a large amount of user-generated content to understand the voice of health consumers has attracted much attention from health care providers and health care researchers. Objective The aim of this paper is to automatically extract hidden topics from Web-based physician reviews using text-mining techniques to examine what Chinese patients have said about their doctors and whether these topics differ across various specialties. This knowledge will help health care consumers, providers, and researchers better understand this information. Methods We conducted two-fold analyses on the data collected from the “Good Doctor Online” platform, the largest online health community in China. First, we explored all reviews from 2006-2014 using descriptive statistics. Second, we applied the well-known topic extraction algorithm Latent Dirichlet Allocation to more than 500,000 textual reviews from over 75,000 Chinese doctors across four major specialty areas to understand what Chinese health consumers said online about their doctor visits. Results On the “Good Doctor Online” platform, 112,873 out of 314,624 doctors had been reviewed at least once by April 11, 2014. Among the 772,979 textual reviews, we chose to focus on four major specialty areas that received the most reviews: Internal Medicine, Surgery, Obstetrics/Gynecology and Pediatrics, and Chinese Traditional Medicine. Among the doctors who received reviews from those four medical specialties, two-thirds of them received more than two reviews and in a few extreme cases, some doctors received more than 500 reviews. Across the four major areas, the most popular topics reviewers found were the experience of finding doctors, doctors’ technical skills and bedside manner, general appreciation from patients, and description of various symptoms. Conclusions To the best of our knowledge, our work is the first study using an automated text-mining approach to analyze a large amount of unstructured textual data of Web-based physician reviews in China. Based on our analysis, we found that Chinese reviewers mainly concentrate on a few popular topics. This is consistent with the goal of Chinese online health platforms and demonstrates the health care focus in China’s health care system. Our text-mining approach reveals a new research area on how to use big data to help health care providers, health care administrators, and policy makers hear patient voices, target patient concerns, and improve the quality of care in this age of patient-centered care. Also, on the health care consumer side, our text mining technique helps patients make more informed decisions about which specialists to see without reading thousands of reviews, which is simply not feasible. In addition, our comparison analysis of Web-based physician reviews in China and the United States also indicates some cultural differences. PMID:27165558

  6. The Voice of Chinese Health Consumers: A Text Mining Approach to Web-Based Physician Reviews.

    PubMed

    Hao, Haijing; Zhang, Kunpeng

    2016-05-10

    Many Web-based health care platforms allow patients to evaluate physicians by posting open-end textual reviews based on their experiences. These reviews are helpful resources for other patients to choose high-quality doctors, especially in countries like China where no doctor referral systems exist. Analyzing such a large amount of user-generated content to understand the voice of health consumers has attracted much attention from health care providers and health care researchers. The aim of this paper is to automatically extract hidden topics from Web-based physician reviews using text-mining techniques to examine what Chinese patients have said about their doctors and whether these topics differ across various specialties. This knowledge will help health care consumers, providers, and researchers better understand this information. We conducted two-fold analyses on the data collected from the "Good Doctor Online" platform, the largest online health community in China. First, we explored all reviews from 2006-2014 using descriptive statistics. Second, we applied the well-known topic extraction algorithm Latent Dirichlet Allocation to more than 500,000 textual reviews from over 75,000 Chinese doctors across four major specialty areas to understand what Chinese health consumers said online about their doctor visits. On the "Good Doctor Online" platform, 112,873 out of 314,624 doctors had been reviewed at least once by April 11, 2014. Among the 772,979 textual reviews, we chose to focus on four major specialty areas that received the most reviews: Internal Medicine, Surgery, Obstetrics/Gynecology and Pediatrics, and Chinese Traditional Medicine. Among the doctors who received reviews from those four medical specialties, two-thirds of them received more than two reviews and in a few extreme cases, some doctors received more than 500 reviews. Across the four major areas, the most popular topics reviewers found were the experience of finding doctors, doctors' technical skills and bedside manner, general appreciation from patients, and description of various symptoms. To the best of our knowledge, our work is the first study using an automated text-mining approach to analyze a large amount of unstructured textual data of Web-based physician reviews in China. Based on our analysis, we found that Chinese reviewers mainly concentrate on a few popular topics. This is consistent with the goal of Chinese online health platforms and demonstrates the health care focus in China's health care system. Our text-mining approach reveals a new research area on how to use big data to help health care providers, health care administrators, and policy makers hear patient voices, target patient concerns, and improve the quality of care in this age of patient-centered care. Also, on the health care consumer side, our text mining technique helps patients make more informed decisions about which specialists to see without reading thousands of reviews, which is simply not feasible. In addition, our comparison analysis of Web-based physician reviews in China and the United States also indicates some cultural differences.

  7. Open Clients for Distributed Databases

    NASA Astrophysics Data System (ADS)

    Chayes, D. N.; Arko, R. A.

    2001-12-01

    We are actively developing a collection of open source example clients that demonstrate use of our "back end" data management infrastructure. The data management system is reported elsewhere at this meeting (Arko and Chayes: A Scaleable Database Infrastructure). In addition to their primary goal of being examples for others to build upon, some of these clients may have limited utility in them selves. More information about the clients and the data infrastructure is available on line at http://data.ldeo.columbia.edu. The available examples to be demonstrated include several web-based clients including those developed for the Community Review System of the Digital Library for Earth System Education, a real-time watch standers log book, an offline interface to use log book entries, a simple client to search on multibeam metadata and others are Internet enabled and generally web-based front ends that support searches against one or more relational databases using industry standard SQL queries. In addition to the web based clients, simple SQL searches from within Excel and similar applications will be demonstrated. By defining, documenting and publishing a clear interface to the fully searchable databases, it becomes relatively easy to construct client interfaces that are optimized for specific applications in comparison to building a monolithic data and user interface system.

  8. Bioaccumulation and trophic transfer of short chain chlorinated paraffins in a marine food web from Liaodong Bay, North China.

    PubMed

    Ma, Xindong; Zhang, Haijun; Wang, Zhen; Yao, Ziwei; Chen, Jingwen; Chen, Jiping

    2014-05-20

    Short chain chlorinated paraffins (SCCPs) are under the evaluation for inclusion into the Stockholm Convention on persistent organic pollutants. However, information on their bioconcentration and biomagnification in marine ecosystems is unavailable, limiting the evaluation of their ecological risks. In this study, seawater, sediment, zooplankton, invertebrates, and fishes collected from Liaodong Bay, Bohai Sea, North China were analyzed to investigate the residual level, congener group profile, bioaccumulation, and trophic transfer of SCCPs in a marine food web. The total concentrations of SCCPs ranged from 4.1 to 13.1 ng L(-1) in seawater, 65 to 541 ng g(-1) (dw) in sediment, and 86 to 4400 ng g(-1) (ww) in organisms. Correspondence analysis indicated the relative enrichment of C10Cl5 and C11Cl5 formula groups in most aquatic organisms. Both the logarithm bioaccumulation factors (log BAFs: 4.1-6.7) and biota-sediment accumulation factors (BSAFs: 0.1-7.3) of individual congeners implied the bioaccumulation of SCCPs. The trophic magnification factor (TMF) of ∑SCCPs was determined to be 2.38 in the zooplankton-shrimp-fish food web, indicating biomagnification potential of SCCPs in the marine ecosystem. The TMF values of individual congener groups significantly correlated with their log KOW values.

  9. Mining-related metals in terrestrial food webs of the upper Clark Fork River basin

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Pastorok, R.A.; LaTier, A.J.; Butcher, M.K.

    1994-12-31

    Fluvial deposits of tailings and other mining-related waste in selected riparian habitats of the Upper Clark Fork River basin (Montana) have resulted in metals enriched soils. The significance of metals exposure to selected wildlife species was evaluated by measuring tissue residues of metals (arsenic, cadmium, copper, lead, zinc) in key dietary species, including dominant grasses (tufted hair grass and redtop), willows, alfalfa, barley, invertebrates (grasshoppers, spiders, and beetles), and deer mice. Average metals concentrations in grasses, invertebrates, and deer mice collected from tailings-affected sites were elevated relative to reference to reference levels. Soil-tissue bioconcentration factors for grasses and invertebrates weremore » generally lower than expected based on the range of values in the literature, indicating the reduced bioavailability of metals from mining waste. In general, metals concentrations in willows, alfalfa, and barley were not elevated above reference levels. Using these data and plausible assumptions for other exposure parameters for white-tailed deer, red fox, and American kestrel, metals intake was estimated for soil and diet ingestion pathways. Comparisons of exposure estimates with toxicity reference values indicated that the elevated concentrations of metals in key food web species do not pose a significant risk to wildlife.« less

  10. CrosstalkNet: A Visualization Tool for Differential Co-expression Networks and Communities.

    PubMed

    Manem, Venkata; Adam, George Alexandru; Gruosso, Tina; Gigoux, Mathieu; Bertos, Nicholas; Park, Morag; Haibe-Kains, Benjamin

    2018-04-15

    Variations in physiological conditions can rewire molecular interactions between biological compartments, which can yield novel insights into gain or loss of interactions specific to perturbations of interest. Networks are a promising tool to elucidate intercellular interactions, yet exploration of these large-scale networks remains a challenge due to their high dimensionality. To retrieve and mine interactions, we developed CrosstalkNet, a user friendly, web-based network visualization tool that provides a statistical framework to infer condition-specific interactions coupled with a community detection algorithm for bipartite graphs to identify significantly dense subnetworks. As a case study, we used CrosstalkNet to mine a set of 54 and 22 gene-expression profiles from breast tumor and normal samples, respectively, with epithelial and stromal compartments extracted via laser microdissection. We show how CrosstalkNet can be used to explore large-scale co-expression networks and to obtain insights into the biological processes that govern cross-talk between different tumor compartments. Significance: This web application enables researchers to mine complex networks and to decipher novel biological processes in tumor epithelial-stroma cross-talk as well as in other studies of intercompartmental interactions. Cancer Res; 78(8); 2140-3. ©2018 AACR . ©2018 American Association for Cancer Research.

  11. Expert Coaching in Weight Loss: Retrospective Analysis

    PubMed Central

    Kushner, Robert F; Hill, James O; Lindquist, Richard; Brunning, Scott; Margulies, Amy

    2018-01-01

    Background Providing coaches as part of a weight management program is a common practice to increase participant engagement and weight loss success. Understanding coach and participant interactions and how these interactions impact weight loss success needs to be further explored for coaching best practices. Objective The purpose of this study was to analyze the coach and participant interaction in a 6-month weight loss intervention administered by Retrofit, a personalized weight management and Web-based disease prevention solution. The study specifically examined the association between different methods of coach-participant interaction and weight loss and tried to understand the level of coaching impact on weight loss outcome. Methods A retrospective analysis was performed using 1432 participants enrolled from 2011 to 2016 in the Retrofit weight loss program. Participants were males and females aged 18 years or older with a baseline body mass index of ≥25 kg/m², who also provided at least one weight measurement beyond baseline. First, a detailed analysis of different coach-participant interaction was performed using both intent-to-treat and completer populations. Next, a multiple regression analysis was performed using all measures associated with coach-participant interactions involving expert coaching sessions, live weekly expert-led Web-based classes, and electronic messaging and feedback. Finally, 3 significant predictors (P<.001) were analyzed in depth to reveal the impact on weight loss outcome. Results Participants in the Retrofit weight loss program lost a mean 5.14% (SE 0.14) of their baseline weight, with 44% (SE 0.01) of participants losing at least 5% of their baseline weight. Multiple regression model (R2=.158, P<.001) identified the following top 3 measures as significant predictors of weight loss at 6 months: expert coaching session attendance (P<.001), live weekly Web-based class attendance (P<.001), and food log feedback days per week (P<.001). Attending 80% of expert coaching sessions, attending 60% of live weekly Web-based classes, and receiving a minimum of 1 food log feedback day per week were associated with clinically significant weight loss. Conclusions Participant’s one-on-one expert coaching session attendance, live weekly expert-led interactive Web-based class attendance, and the number of food log feedback days per week from expert coach were significant predictors of weight loss in a 6-month intervention. PMID:29535082

  12. Web processing service for landslide hazard assessment

    NASA Astrophysics Data System (ADS)

    Sandric, I.; Ursaru, P.; Chitu, D.; Mihai, B.; Savulescu, I.

    2012-04-01

    Hazard analysis requires heavy computation and specialized software. Web processing services can offer complex solutions that can be accessed through a light client (web or desktop). This paper presents a web processing service (both WPS and Esri Geoprocessing Service) for landslides hazard assessment. The web processing service was build with Esri ArcGIS Server solution and Python, developed using ArcPy, GDAL Python and NumPy. A complex model for landslide hazard analysis using both predisposing and triggering factors combined into a Bayesian temporal network with uncertainty propagation was build and published as WPS and Geoprocessing service using ArcGIS Standard Enterprise 10.1. The model uses as predisposing factors the first and second derivatives from DEM, the effective precipitations, runoff, lithology and land use. All these parameters can be served by the client from other WFS services or by uploading and processing the data on the server. The user can select the option of creating the first and second derivatives from the DEM automatically on the server or to upload the data already calculated. One of the main dynamic factors from the landslide analysis model is leaf area index. The LAI offers the advantage of modelling not just the changes from different time periods expressed in years, but also the seasonal changes in land use throughout a year. The LAI index can be derived from various satellite images or downloaded as a product. The upload of such data (time series) is possible using a NetCDF file format. The model is run in a monthly time step and for each time step all the parameters values, a-priory, conditional and posterior probability are obtained and stored in a log file. The validation process uses landslides that have occurred during the period up to the active time step and checks the records of the probabilities and parameters values for those times steps with the values of the active time step. Each time a landslide has been positive identified new a-priory probabilities are recorded for each parameter. A complete log for the entire model is saved and used for statistical analysis and a NETCDF file is created and it can be downloaded from the server with the log file

  13. An Extraction Method of an Informative DOM Node from a Web Page by Using Layout Information

    NASA Astrophysics Data System (ADS)

    Tsuruta, Masanobu; Masuyama, Shigeru

    We propose an informative DOM node extraction method from a Web page for preprocessing of Web content mining. Our proposed method LM uses layout data of DOM nodes generated by a generic Web browser, and the learning set consists of hundreds of Web pages and the annotations of informative DOM nodes of those Web pages. Our method does not require large scale crawling of the whole Web site to which the target Web page belongs. We design LM so that it uses the information of the learning set more efficiently in comparison to the existing method that uses the same learning set. By experiments, we evaluate the methods obtained by combining one that consists of the method for extracting the informative DOM node both the proposed method and the existing methods, and the existing noise elimination methods: Heur removes advertisements and link-lists by some heuristics and CE removes the DOM nodes existing in the Web pages in the same Web site to which the target Web page belongs. Experimental results show that 1) LM outperforms other methods for extracting the informative DOM node, 2) the combination method (LM, {CE(10), Heur}) based on LM (precision: 0.755, recall: 0.826, F-measure: 0.746) outperforms other combination methods.

  14. Mining Specific and General Features in Both Positive and Negative Relevance Feedback. QUT E-Discovery Lab at the TREC󈧍 Relevance Feedback Track

    DTIC Science & Technology

    2009-11-01

    relevance feedback algo- rithm. Four methods, εMap [1], MapA , P10A, and StatAP [2], were used in the track to measure the performance of Phase 2 runs...εMap and StatAP were applied to the runs us- ing the testing set of only ClueWeb09 Category-B, whereas MapA and P10A were applied to those using the...whole ClueWeb09 English set. Because our experiments were based on only ClueWeb09 Category-B, measuring our per- formance by MapA and P10A might not

  15. The Comprehensive Microbial Resource

    PubMed Central

    Peterson, Jeremy D.; Umayam, Lowell A.; Dickinson, Tanja; Hickey, Erin K.; White, Owen

    2001-01-01

    One challenge presented by large-scale genome sequencing efforts is effective display of uniform information to the scientific community. The Comprehensive Microbial Resource (CMR) contains robust annotation of all complete microbial genomes and allows for a wide variety of data retrievals. The bacterial information has been placed on the Web at http://www.tigr.org/CMR for retrieval using standard web browsing technology. Retrievals can be based on protein properties such as molecular weight or hydrophobicity, GC-content, functional role assignments and taxonomy. The CMR also has special web-based tools to allow data mining using pre-run homology searches, whole genome dot-plots, batch downloading and traversal across genomes using a variety of datatypes. PMID:11125067

  16. Improving entrepreneurial opportunity recognition through web content analytics

    NASA Astrophysics Data System (ADS)

    Bakar, Muhamad Shahbani Abu; Azmi, Azwiyati

    2017-10-01

    The ability to recognize and develop an opportunity into a venture defines an entrepreneur. Research in opportunity recognition has been robust and focuses more on explaining the processes involved in opportunity recognition. Factors such as prior knowledge, cognitive and creative capabilities are shown to affect opportunity recognition in entrepreneurs. Prior knowledge in areas such as customer problems, ways to serve the market, and technology has been shows in various studies to be a factor that facilitates entrepreneurs to identify and recognize opportunities. Findings from research also shows that experienced entrepreneurs search and scan for information to discover opportunities. Searching and scanning for information has also been shown to help novice entrepreneurs who lack prior knowledge to narrow this gap and enable them to better identify and recognize opportunities. There is less focus in research on finding empirically proven techniques and methods to develop and enhance opportunity recognition in student entrepreneurs. This is important as the country pushes for more graduate entrepreneurs that can drive the economy. This paper aims to discuss Opportunity Recognition Support System (ORSS), an information support system to help especially student entrepreneurs in identifying and recognizing business opportunities. The ORSS aims to provide the necessary knowledge to student entrepreneurs to be able to better identify and recognize opportunities. Applying design research, theories in opportunity recognition are applied to identify the requirements for the support system and the requirements in turn dictate the design of the support system. The paper proposes the use of web content mining and analytics as two core components and techniques for the support system. Web content mining can mine the vast knowledge repositories available on the internet and analytics can provide entrepreneurs with further insights into the information needed to recognize opportunities in a given market or industry.

  17. DOE Office of Scientific and Technical Information (OSTI.GOV)

    Thompson, R.L.; Wade, G.L.; Straw, R.A.

    A descriptive study of the naturally invading and planted flora was conducted during 1984-1985 on a 14- and 21-year-old contour surface mine the 14.2 ha Log Mountain Demonstration Area (LMDA), in Bell County, Kentucky. Six habitats are designated from areas created from coal mining; the 1963 bench, 1970 bench, bench highwalls, mine outslopes, mine seeps, and coal haul-telephone microwave tower road. Twenty-four of 25 woody and herbaceous species (11 indigenous, 13 non-indigenous) have persisted from plantings by personnel of the Northeastern Forest Experiment Station, USDA Forest Service. We recommend 11 native and exotic woody and herbaceous species for planting onmore » coal surface-mined areas. An annotated list of vascular plants comprises 360 taxa (286 indigenous, 74 non-indigenous) in 224 genera from 82 families. Taxa consist of 1 Lycopodiophyta, 1 Equisetophyta, 8 Polypodiophyta, 7 Pinophyta, and 343 Magnoliophyta. The most species-rich families are the Asteraceae (64), Poaceae (39), Fabaceae (20), Cyperaceae (16), Rosaceae (13), and Lamiaceae (11). A total of 155 Bell County distribution records were documented. Three threatened Kentucky species (Gentiana decora, Liparis loeselii, Silene ovata) were present in refugial habitats created by surface mining. The high species richness has resulted from native and naturalized invading species from the environs, native and exotic planted species, and species from the remnant seed bank. Forest vegetation is a complex mosaic of natural and semi-natural plant communities on the unplanted and planted areas of LMDA.« less

  18. NewsWire, 2002.

    ERIC Educational Resources Information Center

    Byrom, Elizabeth, Ed.; Bingham, Margaret, Ed.; Bowman, Gloria, Ed.; Shoemaker, Dan, Ed.

    2002-01-01

    This document presents the 3 2002 issues of the newsletter "NewsWire," (volume 5). Issue Number One focuses on collaborative Web projects. This issue begins with descriptions of four individual projects: "iEARN"; "Operation RubyThroat"; "Follow the Polar Huskies!"; and "Log in Your Animal Roadkill!" Features that follow include: "Bringing the…

  19. Agricultural Products | National Agricultural Library

    Science.gov Websites

    Skip to main content Home National Agricultural Library United States Department of Agriculture Ag News Contact Us Search  Log inRegister Home Home Agricultural Products NEWT: National Extension Web , tables, graphs), Agricultural Products html National Animal Nutrition Program (NANP) Feed Composition

  20. Data mining learning bootstrap through semantic thumbnail analysis

    NASA Astrophysics Data System (ADS)

    Battiato, Sebastiano; Farinella, Giovanni Maria; Giuffrida, Giovanni; Tribulato, Giuseppe

    2007-01-01

    The rapid increase of technological innovations in the mobile phone industry induces the research community to develop new and advanced systems to optimize services offered by mobile phones operators (telcos) to maximize their effectiveness and improve their business. Data mining algorithms can run over data produced by mobile phones usage (e.g. image, video, text and logs files) to discover user's preferences and predict the most likely (to be purchased) offer for each individual customer. One of the main challenges is the reduction of the learning time and cost of these automatic tasks. In this paper we discuss an experiment where a commercial offer is composed by a small picture augmented with a short text describing the offer itself. Each customer's purchase is properly logged with all relevant information. Upon arrival of new items we need to learn who the best customers (prospects) for each item are, that is, the ones most likely to be interested in purchasing that specific item. Such learning activity is time consuming and, in our specific case, is not applicable given the large number of new items arriving every day. Basically, given the current customer base we are not able to learn on all new items. Thus, we need somehow to select among those new items to identify the best candidates. We do so by using a joint analysis between visual features and text to estimate how good each new item could be, that is, whether or not is worth to learn on it. Preliminary results show the effectiveness of the proposed approach to improve classical data mining techniques.

  1. Optimizing Crawler4j using MapReduce Programming Model

    NASA Astrophysics Data System (ADS)

    Siddesh, G. M.; Suresh, Kavya; Madhuri, K. Y.; Nijagal, Madhushree; Rakshitha, B. R.; Srinivasa, K. G.

    2017-06-01

    World wide web is a decentralized system that consists of a repository of information on the basis of web pages. These web pages act as a source of information or data in the present analytics world. Web crawlers are used for extracting useful information from web pages for different purposes. Firstly, it is used in web search engines where the web pages are indexed to form a corpus of information and allows the users to query on the web pages. Secondly, it is used for web archiving where the web pages are stored for later analysis phases. Thirdly, it can be used for web mining where the web pages are monitored for copyright purposes. The amount of information processed by the web crawler needs to be improved by using the capabilities of modern parallel processing technologies. In order to solve the problem of parallelism and the throughput of crawling this work proposes to optimize the Crawler4j using the Hadoop MapReduce programming model by parallelizing the processing of large input data. Crawler4j is a web crawler that retrieves useful information about the pages that it visits. The crawler Crawler4j coupled with data and computational parallelism of Hadoop MapReduce programming model improves the throughput and accuracy of web crawling. The experimental results demonstrate that the proposed solution achieves significant improvements with respect to performance and throughput. Hence the proposed approach intends to carve out a new methodology towards optimizing web crawling by achieving significant performance gain.

  2. Large-Strain Monitoring Above a Longwall Coal Mine With GPS and Seismic Measurements

    NASA Astrophysics Data System (ADS)

    Swanson, P. L.; Andreatta, V.; Meertens, C. M.; Krahenbuhl, T.; Kenner, B.

    2001-12-01

    As part of an effort to evaluate continuous GPS measurements for use in mine safety studies, a joint GPS-seismic experiment was conducted at an underground longwall coal mine near Paonia, Colorado in June, 2001. Seismic and deformation signals were measured using prototype low-cost monitoring systems as a longwall panel was excavated 150 m beneath the site. Data from both seismic and GPS instruments were logged onto low-power PC-104 Linux computers which were networked using a wireless LAN. The seismic system under development at NIOSH/SRL is based on multiple distributed 8-channel 24-bit A/D converters. The GPS system uses a serial single-frequency (L1) receiver and UNAVCO's "Jstream" Java data logging software. For this experiment, a continuously operating dual-frequency GPS receiver was installed 2.4 km away to serve as a reference site. In addition to the continuously operating sites, 10 benchmarks were surveyed daily with short "rapid-static" occupations in order to provide greater spatial sampling. Two single-frequency sites were located 35 meters apart on a relatively steep north-facing slope. As mining progressed from the east, net displacements of 1.2 meters to the north and 1.65 meters of subsidence were observed over a period of 6 days. The east component exhibited up to 0.45 meters of eastward displacement (toward the excavation) followed by reverse movement to the west. This cycle, observed approximately two days earlier at the eastern L1 site, is consistent with a change in surface strain from tension to compression as the excavation front passed underneath. As this strain "wave" propagated across the field site, surface deformation underwent a cycle of tension crack nucleation, crack opening (up to 15 cm normal displacements), subsequent crack closure, and production of low-angle-thrust compressional deformation features. Analysis of seismic results, surface deformation, and additional survey results are presented.

  3. Preservation Benefits Geoscientific Investigations Across the Nation

    NASA Astrophysics Data System (ADS)

    Powers, L. A.; Latysh, N.

    2017-12-01

    Since 2005, the National Geological and Geophysical Data Preservation Program (NGGDPP) of the U.S. Geological Survey (USGS) has distributed financial grants to state geological surveys to preserve, archive, and make available valuable geoscientific samples and data to researchers and the public. States have cataloged and preserved materials that include geophysical logs, geotechnical reports, fragile historical documents, maps, geologic samples, and legacy aerial and field-investigation photographs. Approximately 3 million metadata records describing preserved data and artifacts are cataloged in the National Digital Catalog, a component of the USGS ScienceBase data management infrastructure. Providing a centralized domain in the National Digital Catalog for uniformly described records has enabled discovery of important geoscientific assets across the Nation. Scientific investigations continue to be informed by preserved materials and data. Tennessee Geological Survey's preserved collection of historical documents describing coal mining activities in the State was used to identify vulnerable areas overlying abandoned underground coal mines, which caused surface collapses and sinkholes in populated areas. Missouri Geological Survey's preserved collection of legacy field notebooks was used to identify thousands of abandoned mines, many of which have significant soil or groundwater lead contamination and are located in areas that now have residential development. The information enabled the evaluation of risk to human health, environment, and infrastructure and identification of needed remedial actions. Information in the field notebooks also assisted the Missouri Department of Transportation responding to highway collapses and assessing collapse potential in abandoned coal mining lands. Digitization of natural gamma ray logs allowed Minnesota Geological Survey staff to directly access well data in the field, accelerating the ability to address geoscientific questions related to aquifer studies, contaminant transport, and geologic mapping and characterization. Digitization and preservation of materials and data, which would otherwise be prohibitively expensive or impossible to reproduce, are a nominal cost compared to the return in societal value that they provide.

  4. Make Mine a Metasearcher, Please!

    ERIC Educational Resources Information Center

    Repman, Judi; Carlson, Randal D.

    2000-01-01

    Describes metasearch tools and explains their value in helping library media centers improve students' Web searches. Discusses Boolean queries and the emphasis on speed at the expense of comprehensiveness; and compares four metasearch tools, including the number of search engines consulted, user control, and databases included. (LRW)

  5. DATAFERRETT AND DATAWEB

    EPA Science Inventory

    DataFerrett is a data extraction software and a data mining tool that accesses data stored in TheDataWeb through the Internet. It can be installed as an application on your desktop or use a java applet with an Internet browser. Census Bureau and Bureau of Labor Statistics release...

  6. The Umbra Simulation and Integration Framework Applied to Emergency Response Training

    NASA Technical Reports Server (NTRS)

    Hamilton, Paul Lawrence; Britain, Robert

    2010-01-01

    The Mine Emergency Response Interactive Training Simulation (MERITS) is intended to prepare personnel to manage an emergency in an underground coal mine. The creation of an effective training environment required realistic emergent behavior in response to simulation events and trainee interventions, exploratory modification of miner behavior rules, realistic physics, and incorporation of legacy code. It also required the ability to add rich media to the simulation without conflicting with normal desktop security settings. Our Umbra Simulation and Integration Framework facilitated agent-based modeling of miners and rescuers and made it possible to work with subject matter experts to quickly adjust behavior through script editing, rather than through lengthy programming and recompilation. Integration of Umbra code with the WebKit browser engine allowed the use of JavaScript-enabled local web pages for media support. This project greatly extended the capabilities of Umbra in support of training simulations and has implications for simulations that combine human behavior, physics, and rich media.

  7. A web server for mining Comparative Genomic Hybridization (CGH) data

    NASA Astrophysics Data System (ADS)

    Liu, Jun; Ranka, Sanjay; Kahveci, Tamer

    2007-11-01

    Advances in cytogenetics and molecular biology has established that chromosomal alterations are critical in the pathogenesis of human cancer. Recurrent chromosomal alterations provide cytological and molecular markers for the diagnosis and prognosis of disease. They also facilitate the identification of genes that are important in carcinogenesis, which in the future may help in the development of targeted therapy. A large amount of publicly available cancer genetic data is now available and it is growing. There is a need for public domain tools that allow users to analyze their data and visualize the results. This chapter describes a web based software tool that will allow researchers to analyze and visualize Comparative Genomic Hybridization (CGH) datasets. It employs novel data mining methodologies for clustering and classification of CGH datasets as well as algorithms for identifying important markers (small set of genomic intervals with aberrations) that are potentially cancer signatures. The developed software will help in understanding the relationships between genomic aberrations and cancer types.

  8. Design and Implementation of a Comprehensive Web-based Survey for Ovarian Cancer Survivorship with an Analysis of Prediagnosis Symptoms via Text Mining

    PubMed Central

    Sun, Jiayang; Bogie, Kath M; Teagno, Joe; Sun, Yu-Hsiang (Sam); Carter, Rebecca R; Cui, Licong; Zhang, Guo-Qiang

    2014-01-01

    Ovarian cancer (OvCa) is the most lethal gynecologic disease in the United States, with an overall 5-year survival rate of 44.5%, about half of the 89.2% for all breast cancer patients. To identify factors that possibly contribute to the long-term survivorship of women with OvCa, we conducted a comprehensive online Ovarian Cancer Survivorship Survey from 2009 to 2013. This paper presents the design and implementation of our survey, introduces its resulting data source, the OVA-CRADLE™ (Clinical Research Analytics and Data Lifecycle Environment), and illustrates a sample application of the survey and data by an analysis of prediagnosis symptoms, using text mining and statistics. The OVA-CRADLE™ is an application of our patented Physio-MIMI technology, facilitating Web-based access, online query and exploration of data. The prediagnostic symptoms and association of early-stage OvCa diagnosis with endometriosis provide potentially important indicators for future studies in this field. PMID:25861211

  9. NCBI GEO: mining millions of expression profiles--database and tools.

    PubMed

    Barrett, Tanya; Suzek, Tugba O; Troup, Dennis B; Wilhite, Stephen E; Ngau, Wing-Chi; Ledoux, Pierre; Rudnev, Dmitry; Lash, Alex E; Fujibuchi, Wataru; Edgar, Ron

    2005-01-01

    The Gene Expression Omnibus (GEO) at the National Center for Biotechnology Information (NCBI) is the largest fully public repository for high-throughput molecular abundance data, primarily gene expression data. The database has a flexible and open design that allows the submission, storage and retrieval of many data types. These data include microarray-based experiments measuring the abundance of mRNA, genomic DNA and protein molecules, as well as non-array-based technologies such as serial analysis of gene expression (SAGE) and mass spectrometry proteomic technology. GEO currently holds over 30,000 submissions representing approximately half a billion individual molecular abundance measurements, for over 100 organisms. Here, we describe recent database developments that facilitate effective mining and visualization of these data. Features are provided to examine data from both experiment- and gene-centric perspectives using user-friendly Web-based interfaces accessible to those without computational or microarray-related analytical expertise. The GEO database is publicly accessible through the World Wide Web at http://www.ncbi.nlm.nih.gov/geo.

  10. GenCLiP 2.0: a web server for functional clustering of genes and construction of molecular networks based on free terms.

    PubMed

    Wang, Jia-Hong; Zhao, Ling-Feng; Lin, Pei; Su, Xiao-Rong; Chen, Shi-Jun; Huang, Li-Qiang; Wang, Hua-Feng; Zhang, Hai; Hu, Zhen-Fu; Yao, Kai-Tai; Huang, Zhong-Xi

    2014-09-01

    Identifying biological functions and molecular networks in a gene list and how the genes may relate to various topics is of considerable value to biomedical researchers. Here, we present a web-based text-mining server, GenCLiP 2.0, which can analyze human genes with enriched keywords and molecular interactions. Compared with other similar tools, GenCLiP 2.0 offers two unique features: (i) analysis of gene functions with free terms (i.e. any terms in the literature) generated by literature mining or provided by the user and (ii) accurate identification and integration of comprehensive molecular interactions from Medline abstracts, to construct molecular networks and subnetworks related to the free terms. http://ci.smu.edu.cn. Supplementary data are available at Bioinformatics online. © The Author 2014. Published by Oxford University Press. All rights reserved. For Permissions, please e-mail: journals.permissions@oup.com.

  11. String Mining in Bioinformatics

    NASA Astrophysics Data System (ADS)

    Abouelhoda, Mohamed; Ghanem, Moustafa

    Sequence analysis is a major area in bioinformatics encompassing the methods and techniques for studying the biological sequences, DNA, RNA, and proteins, on the linear structure level. The focus of this area is generally on the identification of intra- and inter-molecular similarities. Identifying intra-molecular similarities boils down to detecting repeated segments within a given sequence, while identifying inter-molecular similarities amounts to spotting common segments among two or multiple sequences. From a data mining point of view, sequence analysis is nothing but string- or pattern mining specific to biological strings. For a long time, this point of view, however, has not been explicitly embraced neither in the data mining nor in the sequence analysis text books, which may be attributed to the co-evolution of the two apparently independent fields. In other words, although the word "data-mining" is almost missing in the sequence analysis literature, its basic concepts have been implicitly applied. Interestingly, recent research in biological sequence analysis introduced efficient solutions to many problems in data mining, such as querying and analyzing time series [49,53], extracting information from web pages [20], fighting spam mails [50], detecting plagiarism [22], and spotting duplications in software systems [14].

  12. "Nehiyawewin Askihk": Cree Language on the Land--Language Planning through Consultation in the Loon River Cree First Nation

    ERIC Educational Resources Information Center

    Schreyer, Christine

    2008-01-01

    This article examines the parallel development of language planning and land planning within the Loon River Cree First Nation. Loon River Cree territory, located in north-central Alberta, Canada, is an area where major oil and gas industry, as well as logging and mining are constantly encroaching. The community, who still use Cree in their daily…

  13. Techniques for estimating flood-peak discharges of rural, unregulated streams in Ohio

    USGS Publications Warehouse

    Koltun, G.F.; Roberts, J.W.

    1990-01-01

    Multiple-regression equations are presented for estimating flood-peak discharges having recurrence intervals of 2, 5, 10, 25, 50, and 100 years at ungaged sites on rural, unregulated streams in Ohio. The average standard errors of prediction for the equations range from 33.4% to 41.4%. Peak discharge estimates determined by log-Pearson Type III analysis using data collected through the 1987 water year are reported for 275 streamflow-gaging stations. Ordinary least-squares multiple-regression techniques were used to divide the State into three regions and to identify a set of basin characteristics that help explain station-to- station variation in the log-Pearson estimates. Contributing drainage area, main-channel slope, and storage area were identified as suitable explanatory variables. Generalized least-square procedures, which include historical flow data and account for differences in the variance of flows at different gaging stations, spatial correlation among gaging station records, and variable lengths of station record were used to estimate the regression parameters. Weighted peak-discharge estimates computed as a function of the log-Pearson Type III and regression estimates are reported for each station. A method is provided to adjust regression estimates for ungaged sites by use of weighted and regression estimates for a gaged site located on the same stream. Limitations and shortcomings cited in an earlier report on the magnitude and frequency of floods in Ohio are addressed in this study. Geographic bias is no longer evident for the Maumee River basin of northwestern Ohio. No bias is found to be associated with the forested-area characteristic for the range used in the regression analysis (0.0 to 99.0%), nor is this characteristic significant in explaining peak discharges. Surface-mined area likewise is not significant in explaining peak discharges, and the regression equations are not biased when applied to basins having approximately 30% or less surface-mined area. Analyses of residuals indicate that the equations tend to overestimate flood-peak discharges for basins having approximately 30% or more surface-mined area. (USGS)

  14. A comprehensive review on privacy preserving data mining.

    PubMed

    Aldeen, Yousra Abdul Alsahib S; Salleh, Mazleena; Razzaque, Mohammad Abdur

    2015-01-01

    Preservation of privacy in data mining has emerged as an absolute prerequisite for exchanging confidential information in terms of data analysis, validation, and publishing. Ever-escalating internet phishing posed severe threat on widespread propagation of sensitive information over the web. Conversely, the dubious feelings and contentions mediated unwillingness of various information providers towards the reliability protection of data from disclosure often results utter rejection in data sharing or incorrect information sharing. This article provides a panoramic overview on new perspective and systematic interpretation of a list published literatures via their meticulous organization in subcategories. The fundamental notions of the existing privacy preserving data mining methods, their merits, and shortcomings are presented. The current privacy preserving data mining techniques are classified based on distortion, association rule, hide association rule, taxonomy, clustering, associative classification, outsourced data mining, distributed, and k-anonymity, where their notable advantages and disadvantages are emphasized. This careful scrutiny reveals the past development, present research challenges, future trends, the gaps and weaknesses. Further significant enhancements for more robust privacy protection and preservation are affirmed to be mandatory.

  15. DISEASES: text mining and data integration of disease-gene associations.

    PubMed

    Pletscher-Frankild, Sune; Pallejà, Albert; Tsafou, Kalliopi; Binder, Janos X; Jensen, Lars Juhl

    2015-03-01

    Text mining is a flexible technology that can be applied to numerous different tasks in biology and medicine. We present a system for extracting disease-gene associations from biomedical abstracts. The system consists of a highly efficient dictionary-based tagger for named entity recognition of human genes and diseases, which we combine with a scoring scheme that takes into account co-occurrences both within and between sentences. We show that this approach is able to extract half of all manually curated associations with a false positive rate of only 0.16%. Nonetheless, text mining should not stand alone, but be combined with other types of evidence. For this reason, we have developed the DISEASES resource, which integrates the results from text mining with manually curated disease-gene associations, cancer mutation data, and genome-wide association studies from existing databases. The DISEASES resource is accessible through a web interface at http://diseases.jensenlab.org/, where the text-mining software and all associations are also freely available for download. Copyright © 2014 The Authors. Published by Elsevier Inc. All rights reserved.

  16. MetaRanker 2.0: a web server for prioritization of genetic variation data

    PubMed Central

    Pers, Tune H.; Dworzyński, Piotr; Thomas, Cecilia Engel; Lage, Kasper; Brunak, Søren

    2013-01-01

    MetaRanker 2.0 is a web server for prioritization of common and rare frequency genetic variation data. Based on heterogeneous data sets including genetic association data, protein–protein interactions, large-scale text-mining data, copy number variation data and gene expression experiments, MetaRanker 2.0 prioritizes the protein-coding part of the human genome to shortlist candidate genes for targeted follow-up studies. MetaRanker 2.0 is made freely available at www.cbs.dtu.dk/services/MetaRanker-2.0. PMID:23703204

  17. MetaRanker 2.0: a web server for prioritization of genetic variation data.

    PubMed

    Pers, Tune H; Dworzyński, Piotr; Thomas, Cecilia Engel; Lage, Kasper; Brunak, Søren

    2013-07-01

    MetaRanker 2.0 is a web server for prioritization of common and rare frequency genetic variation data. Based on heterogeneous data sets including genetic association data, protein-protein interactions, large-scale text-mining data, copy number variation data and gene expression experiments, MetaRanker 2.0 prioritizes the protein-coding part of the human genome to shortlist candidate genes for targeted follow-up studies. MetaRanker 2.0 is made freely available at www.cbs.dtu.dk/services/MetaRanker-2.0.

  18. Saada: A Generator of Astronomical Database

    NASA Astrophysics Data System (ADS)

    Michel, L.

    2011-11-01

    Saada transforms a set of heterogeneous FITS files or VOtables of various categories (images, tables, spectra, etc.) in a powerful database deployed on the Web. Databases are located on your host and stay independent of any external server. This job doesn’t require writing code. Saada can mix data of various categories in multiple collections. Data collections can be linked each to others making relevant browsing paths and allowing data-mining oriented queries. Saada supports 4 VO services (Spectra, images, sources and TAP) . Data collections can be published immediately after the deployment of the Web interface.

  19. 78 FR 32270 - U.S. Extractive Industries Transparency Initiative Multi-Stakeholder Group (USEITI MSG) Advisory...

    Federal Register 2010, 2011, 2012, 2013, 2014

    2013-05-29

    ..., consideration of sub- national payments, and discussions on scope and materiality. The agenda for the July 23-24... lines. Please plan to dial into the meeting and/or log- in to WebEx at least 10-15 minutes prior to the...

  20. Publications - GMC 376 | Alaska Division of Geological & Geophysical

    Science.gov Websites

    Alaska's Mineral Industry Reports AKGeology.info Rare Earth Elements WebGeochem Engineering Geology Alaska DGGS GMC 376 Publication Details Title: NWE Drill Logs for the Orange Hill Property, Nabesna Quadrangle , Alaska: 1973 and 1974 Drill holes No. 112 through No. 123 Authors: Northwest Explorations Publication

  1. Publications - GMC 389 | Alaska Division of Geological & Geophysical

    Science.gov Websites

    Alaska's Mineral Industry Reports AKGeology.info Rare Earth Elements WebGeochem Engineering Geology Alaska DGGS GMC 389 Publication Details Title: Core photographs, assay results, and 1988 drill logs from the Cominco DDH-1 through DDH-4 boreholes, Shadow Prospect, Tyonek Quadrangle, Alaska Authors: Millrock

  2. A Query Analysis of Consumer Health Information Retrieval

    PubMed Central

    Hong, Yi; de la Cruz, Norberto; Barnas, Gary; Early, Eileen; Gillis, Rick

    2002-01-01

    The log files of MCW HealthLink web site were analyzed to study users' needs for consumer health information and get a better understanding of the health topics users are searching for, the paths users usually take to find consumer health information and the way to improve search effectiveness.

  3. Digital Workflows for a 3d Semantic Representation of AN Ancient Mining Landscape

    NASA Astrophysics Data System (ADS)

    Hiebel, G.; Hanke, K.

    2017-08-01

    The ancient mining landscape of Schwaz/Brixlegg in the Tyrol, Austria witnessed mining from prehistoric times to modern times creating a first order cultural landscape when it comes to one of the most important inventions in human history: the production of metal. In 1991 a part of this landscape was lost due to an enormous landslide that reshaped part of the mountain. With our work we want to propose a digital workflow to create a 3D semantic representation of this ancient mining landscape with its mining structures to preserve it for posterity. First, we define a conceptual model to integrate the data. It is based on the CIDOC CRM ontology and CRMgeo for geometric data. To transform our information sources to a formal representation of the classes and properties of the ontology we applied semantic web technologies and created a knowledge graph in RDF (Resource Description Framework). Through the CRMgeo extension coordinate information of mining features can be integrated into the RDF graph and thus related to the detailed digital elevation model that may be visualized together with the mining structures using Geoinformation systems or 3D visualization tools. The RDF network of the triple store can be queried using the SPARQL query language. We created a snapshot of mining, settlement and burial sites in the Bronze Age. The results of the query were loaded into a Geoinformation system and a visualization of known bronze age sites related to mining, settlement and burial activities was created.

  4. Phytostabilization of mine tailings in arid and semiarid environments--an emerging remediation technology.

    PubMed

    Mendez, Monica O; Maier, Raina M

    2008-03-01

    Unreclaimed mine tailings sites are a worldwide problem, with thousands of unvegetated, exposed tailings piles presenting a source of contamination for nearby communities. Tailings disposal sites in arid and semiarid environments are especially subject to eolian dispersion and water erosion. Phytostabilization, the use of plants for in situ stabilization of tailings and metal contaminants, is a feasible alternative to costly remediation practices. In this review we emphasize considerations for phytostabilization of mine tailings in arid and semiarid environments, as well as issues impeding its long-term success. We reviewed literature addressing mine closures and revegetation of mine tailings, along with publications evaluating plant ecology, microbial ecology, and soil properties of mine tailings. Data were extracted from peer-reviewed articles and books identified in Web of Science and Agricola databases, and publications available through the U.S. Department of Agriculture, U.S. Environmental Protection Agency, and the United Nations Environment Programme. Harsh climatic conditions in arid and semiarid environments along with the innate properties of mine tailings require specific considerations. Plants suitable for phytostabilization must be native, be drought-, salt-, and metal-tolerant, and should limit shoot metal accumulation. Factors for evaluating metal accumulation and toxicity issues are presented. Also reviewed are aspects of implementing phytostabilization, including plant growth stage, amendments, irrigation, and evaluation. Phytostabilization of mine tailings is a promising remedial technology but requires further research to identify factors affecting its long-term success by expanding knowledge of suitable plant species and mine tailings chemistry in ongoing field trials.

  5. Tracking delays in report availability caused by incorrect exam status with Web-based issue tracking: a quality initiative.

    PubMed

    Awan, Omer Abdulrehman; van Wagenberg, Frans; Daly, Mark; Safdar, Nabile; Nagy, Paul

    2011-04-01

    Many radiology information systems (RIS) cannot accept a final report from a dictation reporting system before the exam has been completed in the RIS by a technologist. A radiologist can still render a report in a reporting system once images are available, but the RIS and ancillary systems may not get the results because of the study's uncompleted status. This delay in completing the study caused an alarming number of delayed reports and was undetected by conventional RIS reporting techniques. We developed a Web-based reporting tool to monitor uncompleted exams and automatically page section supervisors when a report was being delayed by its incomplete status in the RIS. Institutional Review Board exemption was obtained. At four imaging centers, a Python script was developed to poll the dictation system every 10 min for exams in five different modalities that were signed by the radiologist but could not be sent to the RIS. This script logged the exams into an existing Web-based tracking tool using PHP and a MySQL database. The script also text-paged the modality supervisor. The script logged the time at which the report was finally sent, and statistics were aggregated onto a separate Web-based reporting tool. Over a 1-year period, the average number of uncompleted exams per month and time to problem resolution decreased at every imaging center and in almost every imaging modality. Automated feedback provides a vital link in improving technologist performance and patient care without assigning a human resource to manage report queues.

  6. Secure, web-accessible call rosters for academic radiology departments.

    PubMed

    Nguyen, A V; Tellis, W M; Avrin, D E

    2000-05-01

    Traditionally, radiology department call rosters have been posted via paper and bulletin boards. Frequently, changes to these lists are made by multiple people independently, but often not synchronized, resulting in confusion among the house staff and technical staff as to who is on call and when. In addition, multiple and disparate copies exist in different sections of the department, and changes made would not be propagated to all the schedules. To eliminate such difficulties, a paperless call scheduling application was developed. Our call scheduling program allowed Java-enabled web access to a database by designated personnel from each radiology section who have privileges to make the necessary changes. Once a person made a change, everyone accessing the database would see the modification. This eliminates the chaos resulting from people swapping shifts at the last minute and not having the time to record or broadcast the change. Furthermore, all changes to the database were logged. Users are given a log-in name and password and can only edit their section; however, all personnel have access to all sections' schedules. Our applet was written in Java 2 using the latest technology in database access. We access our Interbase database through the DataExpress and DB Swing (Borland, Scotts Valley, CA) components. The result is secure access to the call rosters via the web. There are many advantages to the web-enabled access, mainly the ability for people to make changes and have the changes recorded and propagated in a single virtual location and available to all who need to know.

  7. 76 FR 6110 - Mine Safety Disclosure

    Federal Register 2010, 2011, 2012, 2013, 2014

    2011-02-03

    ... Comments Use the Commission's Internet comment form ( http://www.sec.gov/rules/proposed.shtml ); Send an e... all comments on the Commission's Internet Web site ( http://www.sec.gov/rules/proposed.shtml... on the proposal to, among other things, allow for the collection of information and improve the...

  8. Effects of microhabitat and large-scale land use on stream salamander occupancy in the coalfields of Central Appalachia

    USGS Publications Warehouse

    Sweeten, Sara E.; Ford, W. Mark

    2016-01-01

    Large-scale coal mining practices, particularly surface coal extraction and associated valley fills as well as residential wastewater discharge, are of ecological concern for aquatic systems in central Appalachia. Identifying and quantifying alterations to ecosystems along a gradient of spatial scales is a necessary first-step to aid in mitigation of negative consequences to aquatic biota. In central Appalachian headwater streams, apart from fish, salamanders are the most abundant vertebrate predator that provide a significant intermediate trophic role linking aquatic and terrestrial food webs. Stream salamander species are considered to be sensitive to aquatic stressors and environmental alterations, as past research has shown linkages among microhabitat parameters, large-scale land use such as urbanization and logging, and salamander abundances. However, there is little information examining these relationships between environmental conditions and salamander occupancy in the coalfields of central Appalachia. In the summer of 2013, 70 sites (sampled two to three times each) in the southwest Virginia coalfields were visited to collect salamanders and quantify stream and riparian microhabitat parameters. Using an information-theoretic framework, effects of microhabitat and large-scale land use on stream salamander occupancy were compared. The findings indicate that Desmognathus spp. occupancy rates are more correlated to microhabitat parameters such as canopy cover than to large-scale land uses. However, Eurycea spp. occupancy rates had a strong association with large-scale land uses, particularly recent mining and forest cover within the watershed. These findings suggest that protection of riparian habitats is an important consideration for maintaining aquatic systems in central Appalachia. If this is not possible, restoration riparian areas should follow guidelines using quick-growing tree species that are native to Appalachian riparian areas. These types of trees would rapidly establish a canopy cover, stabilize the soil, and impede invasive plant species which would, in turn, provide high-quality refuges for stream salamanders.

  9. Intelligent web image retrieval system

    NASA Astrophysics Data System (ADS)

    Hong, Sungyong; Lee, Chungwoo; Nah, Yunmook

    2001-07-01

    Recently, the web sites such as e-business sites and shopping mall sites deal with lots of image information. To find a specific image from these image sources, we usually use web search engines or image database engines which rely on keyword only retrievals or color based retrievals with limited search capabilities. This paper presents an intelligent web image retrieval system. We propose the system architecture, the texture and color based image classification and indexing techniques, and representation schemes of user usage patterns. The query can be given by providing keywords, by selecting one or more sample texture patterns, by assigning color values within positional color blocks, or by combining some or all of these factors. The system keeps track of user's preferences by generating user query logs and automatically add more search information to subsequent user queries. To show the usefulness of the proposed system, some experimental results showing recall and precision are also explained.

  10. How Online Quality Ratings Influence Patients' Choice of Medical Providers: Controlled Experimental Survey Study.

    PubMed

    Yaraghi, Niam; Wang, Weiguang; Gao, Guodong Gordon; Agarwal, Ritu

    2018-03-26

    In recent years, the information environment for patients to learn about physician quality is being rapidly changed by Web-based ratings from both commercial and government efforts. However, little is known about how various types of Web-based ratings affect individuals' choice of physicians. The objective of this research was to measure the relative importance of Web-based quality ratings from governmental and commercial agencies on individuals' choice of primary care physicians. In a choice-based conjoint experiment conducted on a sample of 1000 Amazon Mechanical Turk users in October 2016, individuals were asked to choose their preferred primary care physician from pairs of physicians with different ratings in clinical and nonclinical aspects of care provided by governmental and commercial agencies. The relative log odds of choosing a physician increases by 1.31 (95% CI 1.26-1.37; P<.001) and 1.32 (95% CI 1.27-1.39; P<.001) units when the government clinical ratings and commercial nonclinical ratings move from 2 to 4 stars, respectively. The relative log odds of choosing a physician increases by 1.12 (95% CI 1.07-1.18; P<.001) units when the commercial clinical ratings move from 2 to 4 stars. The relative log odds of selecting a physician with 4 stars in nonclinical ratings provided by the government is 1.03 (95% CI 0.98-1.09; P<.001) units higher than a physician with 2 stars in this rating. The log odds of selecting a physician with 4 stars in nonclinical government ratings relative to a physician with 2 stars is 0.23 (95% CI 0.13-0.33; P<.001) units higher for females compared with males. Similar star increase in nonclinical commercial ratings increases the relative log odds of selecting the physician by female respondents by 0.15 (95% CI 0.04-0.26; P=.006) units. Individuals perceive nonclinical ratings provided by commercial websites as important as clinical ratings provided by government websites when choosing a primary care physician. There are significant gender differences in how the ratings are used. More research is needed on whether patients are making the best use of different types of ratings, as well as the optimal allocation of resources in improving physician ratings from the government's perspective. ©Niam Yaraghi, Weiguang Wang, Guodong (Gordon) Gao, Ritu Agarwal. Originally published in the Journal of Medical Internet Research (http://www.jmir.org), 26.03.2018.

  11. PubMed-EX: a web browser extension to enhance PubMed search with text mining features.

    PubMed

    Tsai, Richard Tzong-Han; Dai, Hong-Jie; Lai, Po-Ting; Huang, Chi-Hsin

    2009-11-15

    PubMed-EX is a browser extension that marks up PubMed search results with additional text-mining information. PubMed-EX's page mark-up, which includes section categorization and gene/disease and relation mark-up, can help researchers to quickly focus on key terms and provide additional information on them. All text processing is performed server-side, freeing up user resources. PubMed-EX is freely available at http://bws.iis.sinica.edu.tw/PubMed-EX and http://iisr.cse.yzu.edu.tw:8000/PubMed-EX/.

  12. Two year experience with Web connectivity to PACS at a community-based hospital

    NASA Astrophysics Data System (ADS)

    Wadley, Brian D.; Hayward, Ulrike; Trambert, Michael; Kywi, Alberto; Hartzman, Steven

    2002-05-01

    Referring physician web based access to a PACS is evaluated at a community-based hospital. Survey results show 100% perceived improvement in referring physician productivity, and 97% perceived improvement in patient care. Overall satisfaction and perception of ease of use is greater than 90%.Web connectivity to PACS is perceived by 97% to result in decreased visits and calls to the radiology department due to the availability of virtual results, with high value placed on the voice clip summary results by 94% of respondents (the voice clip is a recording made by the radiologist at the time of reading the exam, which summarizes the results of the study).The availability of exams and virtual results via web-based connectivity is perceived to result in increase referrals by 58% of respondents. Web log audit trails are analyzed to evaluate the use of the system by referring physicians. The gross number of cases retrieved over the web for a two-week period is assessed. The number of times a voice clip was played is also evaluated. The results overall show a very positive experience with web connectivity to PACS.

  13. Analytical Fingerprint of Wolframite Ore Concentrates.

    PubMed

    Gäbler, Hans-Eike; Schink, Wilhelm; Goldmann, Simon; Bahr, Andreas; Gawronski, Timo

    2017-07-01

    Ongoing violent conflicts in Central Africa are fueled by illegal mining and trading of tantalum, tin, and tungsten ores. The credibility of document-based traceability systems can be improved by an analytical fingerprint applied as an independent method to confirm or doubt the documented origin of ore minerals. Wolframite (Fe,Mn)WO 4 is the most important ore mineral for tungsten and is subject to artisanal mining in Central Africa. Element concentrations of wolframite grains analyzed by laser ablation-inductively coupled plasma-mass spectrometry are used to establish the analytical fingerprint. The data from ore concentrate samples are multivariate, not normal or log-normal distributed. The samples cannot be regarded as representative aliquots of a population. Based on the Kolmogorov-Smirnov distance, a measure of similarity between a sample in question and reference samples from a database is determined. A decision criterion is deduced to recognize samples which do not originate from the declared mine site. © 2017 American Academy of Forensic Sciences.

  14. A New MI-Based Visualization Aided Validation Index for Mining Big Longitudinal Web Trial Data

    PubMed Central

    Zhang, Zhaoyang; Fang, Hua; Wang, Honggang

    2016-01-01

    Web-delivered clinical trials generate big complex data. To help untangle the heterogeneity of treatment effects, unsupervised learning methods have been widely applied. However, identifying valid patterns is a priority but challenging issue for these methods. This paper, built upon our previous research on multiple imputation (MI)-based fuzzy clustering and validation, proposes a new MI-based Visualization-aided validation index (MIVOOS) to determine the optimal number of clusters for big incomplete longitudinal Web-trial data with inflated zeros. Different from a recently developed fuzzy clustering validation index, MIVOOS uses a more suitable overlap and separation measures for Web-trial data but does not depend on the choice of fuzzifiers as the widely used Xie and Beni (XB) index. Through optimizing the view angles of 3-D projections using Sammon mapping, the optimal 2-D projection-guided MIVOOS is obtained to better visualize and verify the patterns in conjunction with trajectory patterns. Compared with XB and VOS, our newly proposed MIVOOS shows its robustness in validating big Web-trial data under different missing data mechanisms using real and simulated Web-trial data. PMID:27482473

  15. Hymenoptera Genome Database: integrating genome annotations in HymenopteraMine.

    PubMed

    Elsik, Christine G; Tayal, Aditi; Diesh, Colin M; Unni, Deepak R; Emery, Marianne L; Nguyen, Hung N; Hagen, Darren E

    2016-01-04

    We report an update of the Hymenoptera Genome Database (HGD) (http://HymenopteraGenome.org), a model organism database for insect species of the order Hymenoptera (ants, bees and wasps). HGD maintains genomic data for 9 bee species, 10 ant species and 1 wasp, including the versions of genome and annotation data sets published by the genome sequencing consortiums and those provided by NCBI. A new data-mining warehouse, HymenopteraMine, based on the InterMine data warehousing system, integrates the genome data with data from external sources and facilitates cross-species analyses based on orthology. New genome browsers and annotation tools based on JBrowse/WebApollo provide easy genome navigation, and viewing of high throughput sequence data sets and can be used for collaborative genome annotation. All of the genomes and annotation data sets are combined into a single BLAST server that allows users to select and combine sequence data sets to search. © The Author(s) 2015. Published by Oxford University Press on behalf of Nucleic Acids Research.

  16. CottonGen: a genomics, genetics and breeding database for cotton research

    USDA-ARS?s Scientific Manuscript database

    CottonGen (http://www.cottongen.org) is a curated and integrated web-based relational database providing access to publicly available genomic, genetic and breeding data for cotton. CottonGen supercedes CottonDB and the Cotton Marker Database, with enhanced tools for easier data sharing, mining, vis...

  17. Automatic Recommendations for E-Learning Personalization Based on Web Usage Mining Techniques and Information Retrieval

    ERIC Educational Resources Information Center

    Khribi, Mohamed Koutheair; Jemni, Mohamed; Nasraoui, Olfa

    2009-01-01

    In this paper, we describe an automatic personalization approach aiming to provide online automatic recommendations for active learners without requiring their explicit feedback. Recommended learning resources are computed based on the current learner's recent navigation history, as well as exploiting similarities and dissimilarities among…

  18. 1872 vs 2004: Mining claim meets the World Wide Web

    Treesearch

    Edward Russell

    2006-01-01

    Inappropriate development or land use on private inholdings in a matrix of predominantly public land have the potential to profoundly impact backcountry landscapes. Beyond damage to natural systems and cultural resources, ramifications include impacts on neighboring communities dependent upon tourism and backcountry recreation for their economic vitality. Limited...

  19. Topic Models for Link Prediction in Document Networks

    ERIC Educational Resources Information Center

    Kataria, Saurabh

    2012-01-01

    Recent explosive growth of interconnected document collections such as citation networks, network of web pages, content generated by crowd-sourcing in collaborative environments, etc., has posed several challenging problems for data mining and machine learning community. One central problem in the domain of document networks is that of "link…

  20. A Semantic Web-based System for Mining Genetic Mutations in Cancer Clinical Trials.

    PubMed

    Priya, Sambhawa; Jiang, Guoqian; Dasari, Surendra; Zimmermann, Michael T; Wang, Chen; Heflin, Jeff; Chute, Christopher G

    2015-01-01

    Textual eligibility criteria in clinical trial protocols contain important information about potential clinically relevant pharmacogenomic events. Manual curation for harvesting this evidence is intractable as it is error prone and time consuming. In this paper, we develop and evaluate a Semantic Web-based system that captures and manages mutation evidences and related contextual information from cancer clinical trials. The system has 2 main components: an NLP-based annotator and a Semantic Web ontology-based annotation manager. We evaluated the performance of the annotator in terms of precision and recall. We demonstrated the usefulness of the system by conducting case studies in retrieving relevant clinical trials using a collection of mutations identified from TCGA Leukemia patients and Atlas of Genetics and Cytogenetics in Oncology and Haematology. In conclusion, our system using Semantic Web technologies provides an effective framework for extraction, annotation, standardization and management of genetic mutations in cancer clinical trials.

  1. Modeling CO 2 Sequestration in Saline Aquifer and Depleted Oil Reservoir To Evaluate Regional CO 2 Sequestration Potential of Ozark Plateau Aquifer System, South-Central Kansas

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Watney, W. Lynn

    2014-09-30

    1. Drilled, cored, and logged three wells to the basement and collecting more than 2,700 ft of conventional core; obtained 20 m i2 of multicomponent 3D seismic imaging and merged and reprocessed more than 125 mi 2 of existing 3D seismic data for use in modeling CO 2- EOR oil recovery and CO 2 storage in five oil fields in southern Kansas. 2. Determined the technical feasibility of injecting and sequestering CO 2 in a set of four depleted oil reservoirs in the Cutter, Pleasant Prairie South, Eubank, and Shuck fields in southwest Kansas; of concurrently recovering oil from thosemore » fields; and of quantifying the volumes of CO2 sequestered and oil recovered during the process. 3. Formed a consortium of six oil operating companies, five of which own and operate the four fields. The consortium became part of the Southwest Kansas CO 2-EOR Initiative for the purpose of sharing data, knowledge, and interest in understanding the potential for CO 2-EOR in Kansas. 4. Built a regional well database covering 30,000 mi 2 and containing stratigraphic tops from ~90,000 wells; correlated 30 major stratigraphic horizons; digitized key wells, including wireline logs and sample logs; and analyzed more than 3,000 drill stem tests to establish that fluid levels in deep aquifers below the Permian evaporites are not connected to the surface and therefore pressures are not hydrostatic. Connectivity with the surface aquifers is lacking because shale aquitards and impermeable evaporite layers consist of both halite and anhydrite. 5. Developed extensive web applications and an interactive mapping system that do the following: a. Facilitate access to a wide array of data obtained in the study, including core descriptions and analyses, sample logs, digital (LAS) well logs, seismic data, gravity and magnetics maps, structural and stratigraphic maps, inferred fault traces, earthquakes, Class I and II disposal wells, and surface lineaments. b. Provide real-time analysis of the project dataset, including automated integration and viewing of well logs, core, core analyses, brine chemistry, and stratigraphy using the Java Profile app. A cross-section app allows for the display of log data for up to four wells at a time. 6. Integrated interpretations from the project’s interactive web-based mapping system to gain insights to aid in assessing the efficacy of geologic CO 2 storage in Kansas and insights toward understanding recent seismicity to aid in evaluating induced vs. naturally occurring earthquakes. 7. Developed a digital type-log system, including web-based software to modify and refine stratigraphic nomenclature to provide stakeholders a common means for communication about the subsurface. 8. Contracted use of a nuclear magnetic resonance (NMR) log and ran it slowly to capture response and characterize larger pores common for carbonate reservoirs. Used NMR to extend core analyses to apply permeability, relative permeability to CO 2, and capillary pressure to the major rock types, each uniquely expressed as a reservoir quality index (RQI), present in the Mississippian and Arbuckle rocks. 9. Characterized and evaluated the possible role of microbes in dense brines. Used microbes to compliment H/O stable isotopes to fingerprint brine systems. Used perforation/swabbing to obtain samples from multiple hydrostratigraphic units and confirmed equivalent results using less expensive drill stem tests (DST). 10. Used an integrated approach from whole core, logs, tests, and seismic to verify and quantify properties of vuggy, brecciated, and fractured carbonate intervals. 11. Used complex geocellular static and dynamic models to evaluate regional storage capacity using large parallel processing. 12. Carbonates are complex reservoirs and CO 2-EOR needs to move to the next generation to increase effectiveness of CO 2 and efficiency and safety of the injection.« less

  2. Discovering and visualizing indirect associations between biomedical concepts

    PubMed Central

    Tsuruoka, Yoshimasa; Miwa, Makoto; Hamamoto, Kaisei; Tsujii, Jun'ichi; Ananiadou, Sophia

    2011-01-01

    Motivation: Discovering useful associations between biomedical concepts has been one of the main goals in biomedical text-mining, and understanding their biomedical contexts is crucial in the discovery process. Hence, we need a text-mining system that helps users explore various types of (possibly hidden) associations in an easy and comprehensible manner. Results: This article describes FACTA+, a real-time text-mining system for finding and visualizing indirect associations between biomedical concepts from MEDLINE abstracts. The system can be used as a text search engine like PubMed with additional features to help users discover and visualize indirect associations between important biomedical concepts such as genes, diseases and chemical compounds. FACTA+ inherits all functionality from its predecessor, FACTA, and extends it by incorporating three new features: (i) detecting biomolecular events in text using a machine learning model, (ii) discovering hidden associations using co-occurrence statistics between concepts, and (iii) visualizing associations to improve the interpretability of the output. To the best of our knowledge, FACTA+ is the first real-time web application that offers the functionality of finding concepts involving biomolecular events and visualizing indirect associations of concepts with both their categories and importance. Availability: FACTA+ is available as a web application at http://refine1-nactem.mc.man.ac.uk/facta/, and its visualizer is available at http://refine1-nactem.mc.man.ac.uk/facta-visualizer/. Contact: tsuruoka@jaist.ac.jp PMID:21685059

  3. Evaluating the Tradeoffs Between Dollars Spent and Lives saved in Military Settings

    DTIC Science & Technology

    2013-11-18

    data needed , and completing and reviewing the collection of information. Send comments regarding this burden estimate or any other aspect of this...Programs that improve workplace safety have costs. Firms may need to purchase additional equipment or protective devices, install machine guards...employers the costs may be slight. Because of the inherent dangers in production, firms in mining, logging, fishing, and construction will need to

  4. Conservation performance of different conservation governance regimes in the Peruvian Amazon.

    PubMed

    Schleicher, Judith; Peres, Carlos A; Amano, Tatsuya; Llactayo, William; Leader-Williams, Nigel

    2017-09-12

    State-controlled protected areas (PAs) have dominated conservation strategies globally, yet their performance relative to other governance regimes is rarely assessed comprehensively. Furthermore, performance indicators of forest PAs are typically restricted to deforestation, although the extent of forest degradation is greater. We address these shortfalls through an empirical impact evaluation of state PAs, Indigenous Territories (ITs), and civil society and private Conservation Concessions (CCs) on deforestation and degradation throughout the Peruvian Amazon. We integrated remote-sensing data with environmental and socio-economic datasets, and used propensity-score matching to assess: (i) how deforestation and degradation varied across governance regimes between 2006-2011; (ii) their proximate drivers; and (iii) whether state PAs, CCs and ITs avoided deforestation and degradation compared with logging and mining concessions, and the unprotected landscape. CCs, state PAs, and ITs all avoided deforestation and degradation compared to analogous areas in the unprotected landscape. CCs and ITs were on average more effective in this respect than state PAs, showing that local governance can be equally or more effective than centralized state regimes. However, there were no consistent differences between conservation governance regimes when matched to logging and mining concessions. Future impact assessments would therefore benefit from further disentangling governance regimes across unprotected land.

  5. Dive and discover: Expeditions to the seafloor

    NASA Astrophysics Data System (ADS)

    Lawrence, Lisa Ayers

    The Dive and Discover Web site is a virtual treasure chest of deep sea science and classroom resources. The goals of Dive and Discover are to engage students, teachers, and the general public in the excitement of ocean disco very through an interactive educational Web site. You can follow scientists on oceanographic research cruises by reading their daily cruise logs, viewing photos and video clips of the discoveries, and even e-mailing questions to the scientists and crew. WHOI has also included an “Educator's Companion” section with teaching strategies, activities, and assessments, making Dive and Discover an excellent resource for the classroom.

  6. Dive and discover: Expeditions to the seafloor

    NASA Astrophysics Data System (ADS)

    Ayers Lawrence, Lisa

    The Dive and Discover Web site is a virtual treasure chest of deep sea science and classroom resources. The goals of Dive and Discover are to engage students, teachers, and the general public in the excitement of ocean disco very through an interactive educational Web site. You can follow scientists on oceanographic research cruises by reading their daily cruise logs, viewing photos and video clips of the discoveries, and even e-mailing questions to the scientists and crew. WHOI has also included an "Educator's Companion" section with teaching strategies, activities, and assessments, making Dive and Discover an excellent resource for the classroom.

  7. Reviews

    NASA Astrophysics Data System (ADS)

    2007-09-01

    WE RECOMMEND Energy Foresight Valuable and original GCSE curriculum support on DVD Developing Scientific Literacy: Using News Media in the Classroom This book helpfully evaluates science stories in today's media Radioactivity Explained and Electricity Explained Interactive software ideal for classroom use TEP Generator Wind-up generator specially designed for schools SEP Energymeter A joule meter with more uses than its appearance suggests Into the Cool: Energy Flow, Thermodynamics and Life This book explores the physics behind biology CmapTools Handy software for mapping knowledge and resources LogIT Black Box This hub contains multiple sensors for endless experimental fun WEB WATCH Water Web 2.0

  8. Monitoring the performance of the Southern African Large Telescope

    NASA Astrophysics Data System (ADS)

    Hettlage, Christian; Coetzee, Chris; Väisänen, Petri; Romero Colmenero, Encarni; Crawford, Steven M.; Kotze, Paul; Rabe, Paul; Hulme, Stephen; Brink, Janus; Maartens, Deneys; Browne, Keith; Strydom, Ockert; De Bruyn, David

    2016-07-01

    The efficient operation of a telescope requires awareness of its performance on a daily and long-term basis. This paper outlines the Fault Tracker, WebSAMMI and the Dashboard used by the Southern African Large Telescope (SALT) to achieve this aim. Faults are mostly logged automatically, but the Fault Tracker allows users to add and edit faults. The SALT Astronomer and SALT Operator record weather conditions and telescope usage with WebSAMMI. Various efficiency metrics are shown for different time periods on the Dashboard. A kiosk mode for displaying on a public screen is included. Possible applications for other telescopes are discussed.

  9. The Diesel Exhaust in Miners Study: III. Interrelations between respirable elemental carbon and gaseous and particulate components of diesel exhaust derived from area sampling in underground non-metal mining facilities.

    PubMed

    Vermeulen, Roel; Coble, Joseph B; Yereb, Daniel; Lubin, Jay H; Blair, Aaron; Portengen, Lützen; Stewart, Patricia A; Attfield, Michael; Silverman, Debra T

    2010-10-01

    Diesel exhaust (DE) has been implicated as a potential lung carcinogen. However, the exact components of DE that might be involved have not been clearly identified. In the past, nitrogen oxides (NO(x)) and carbon oxides (CO(x)) were measured most frequently to estimate DE, but since the 1990s, the most commonly accepted surrogate for DE has been elemental carbon (EC). We developed quantitative estimates of historical exposure levels of respirable elemental carbon (REC) for an epidemiologic study of mortality, particularly lung cancer, among diesel-exposed miners by back-extrapolating 1998-2001 REC exposure levels using historical measurements of carbon monoxide (CO). The choice of CO was based on the availability of historical measurement data. Here, we evaluated the relationship of REC with CO and other current and historical components of DE from side-by-side area measurements taken in underground operations of seven non-metal mining facilities. The Pearson correlation coefficient of the natural log-transformed (Ln)REC measurements with the Ln(CO) measurements was 0.4. The correlation of REC with the other gaseous, organic carbon (OC), and particulate measurements ranged from 0.3 to 0.8. Factor analyses indicated that the gaseous components, including CO, together with REC, loaded most strongly on a presumed 'Diesel exhaust' factor, while the OC and particulate agents loaded predominantly on other factors. In addition, the relationship between Ln(REC) and Ln(CO) was approximately linear over a wide range of REC concentrations. The fact that CO correlated with REC, loaded on the same factor, and increased linearly in log-log space supported the use of CO in estimating historical exposure levels to DE.

  10. Factors controlling the permeability distribution in fault vein zones surrounding granitic intrusions (Ore Mountains/Germany)

    NASA Astrophysics Data System (ADS)

    Achtziger-Zupančič, P.; Loew, S.; Hiller, A.

    2017-03-01

    An outstanding legacy data set has been compiled from underground excavations mostly prospected and mined by the former Soviet (German) Stock Company Wismut describing the hydrology of faulted basement rocks in the Ore Mountains (SE Germany). It consists of more than 5000 detailed descriptions of groundwater inflows to about 660 km of tunnels and 57 km of drillings measured during or shortly after excavation. Inflow measurements (recorded between 1E-8 and 4E-2 m3/s) have been converted to fracture transmissivities using a simplified analytical solution. Discarding site specific effects, the median log transmissivity decreases from 1E-7 to 1E-10 m2/s within the studied depth interval of 0-2000 meters below ground surface (mbgs), and the spacing of conductive fracture increases from 0.1 to 2500 m. This general trend is overprinted at three mining sites by a clear reversal of fracture transmissivity which correlates with contact metamorphic aureoles around Variscan granite intrusions (327-295 Ma). We hypothesize that this transmissivity increase is caused by processes accompanying granite intrusion and contact metamorphism. The thickness of these hydraulically active aureoles is greater in lower-grade metamorphic schist than in higher-grade metamorphic gneisses. Rock mass equivalent continuum conductivities have been estimated by arithmetic averaging of fracture and matrix transmissivities over 100 m intervals and have been converted to permeabilities. The median equivalent continuum permeability decreases with depth according to log(k) = - 1.7 * log(z) - 17.3 (k in m2 and increasing depth z in kilometer being positive). Matrix conductivity controls the bulk conductivity below about 1000 mbgs and is less sensitive to the occurrence of contact metamorphic aureoles.

  11. Digital sleep logs reveal potential impacts of modern temporal structure on class performance in different chronotypes.

    PubMed

    Smarr, Benjamin Lee

    2015-02-01

    Stability of sleep and circadian rhythms are important for healthy learning and memory. While experimental manipulations of lifestyle and learning outcomes present major obstacles, the ongoing increase in data sources allows retrospective data mining of people's sleep timing variation. Here I use digital sleep-log data generated by 1109 students in a biology lab course at the University of Washington to test the hypothesis that higher variance in time asleep and later sleep-onset times negatively correlate with class performance, used here as a real-world proxy for learning and memory. I find that sleep duration variance and mean sleep-onset times both significantly correlate with class performance. These correlations are powerful on weeknights but undetectable on Friday and Saturday nights ("free nights"). Finally, although these data come with no demographic information beyond sex, the constructed demographic groups of "larks" and "owls" within the sexes reveal a significant decrease in performance of owls relative to larks in male students, whereas the correlation of performance with sleep-onset time for all male students was only a near-significant trend. This provides a proof of concept that deeper demographic mining of digital logs in the future may identify subgroups for which certain sleep phenotypes have greater predictive value for performance outcomes. The data analyzed are consistent with known patterns, including sleep-timing delays from weeknights to free nights and sleep-timing delays in men relative to women. These findings support the hypothesis that modern schedule impositions on sleep and circadian timing have consequences for real-world learning and memory. This study also highlights the low-cost, large-scale benefits of personal, daily, digital records as an augmentation of sleep and circadian studies. © 2015 The Author(s).

  12. Using Web-Based and Paper-Based Questionnaires for Collecting Data on Fertility Issues Among Female Childhood Cancer Survivors: Differences in Response Characteristics

    PubMed Central

    Overbeek, Annelies; van der Pal, Helena J; Versluys, A. Birgitta; Bresters, Dorine; van Leeuwen, Flora E; Lambalk, Cornelis B; Kaspers, Gertjan J.L; van Dulmen-den Broeder, Eline

    2011-01-01

    Background Web-based questionnaires have become increasingly popular in health research. However, reported response rates vary and response bias may be introduced. Objective The aim of this study was to evaluate whether sending a mixed invitation (paper-based together with Web-based questionnaire) rather than a Web-only invitation (Web-based questionnaire only) results in higher response and participation rates for female childhood cancer survivors filling out a questionnaire on fertility issues. In addition, differences in type of response and characteristics of the responders and nonresponders were investigated. Moreover, factors influencing preferences for either the Web- or paper-based version of the questionnaire were examined. Methods This study is part of a nationwide study on reproductive function, ovarian reserve, and risk of premature menopause in female childhood cancer survivors. The Web-based version of the questionnaire was available for participants through the Internet by means of a personalized user name and password. Participants were randomly selected to receive either a mixed invitation (paper-based questionnaire together with log-in details for Web-based questionnaire, n = 137) or a Web-only invitation (log-in details only, n = 140). Furthermore, the latter group could request a paper-based version of the questionnaire by filling out a form. Results Overall response rates were comparable in both randomization groups (83% mixed invitation group vs 89% in Web-only invitation group, P = .20). In addition, participation rates appeared not to differ (66% or 90/137, mixed invitation group vs 59% or 83/140, Web-only invitation group, P =.27). However, in the mixed invitation group, significantly more respondents filled out the paper-based questionnaire compared with the Web-only invitation group (83% or 75/90 and 65% or 54/83, respectively, P = .01). The 44 women who filled out the Web-based version of the questionnaire had a higher educational level than the 129 women who filled out the paper-based version (P = .01). Furthermore, the probability of filling out the Web-based questionnaire appeared to be greater for women who were allocated to the Web-only invitation group (OR = 2.85, 95% CI 1.31 - 6.21), were older (OR = 1.08, 95% CI 1.02 - 1.15), had a higher educational level (OR high vs low = 0.06, 95% CI 0.01 - 0.52), or were students (OR employed vs student = 3.25, 95% CI 1.00 - 10.56). Conclusions Although overall response as well as participation rates to both types of invitations were similar, adding a paper version of a questionnaire to a Web-only invitation resulted in more respondents filling out the paper-based version. In addition, women who were older, had a higher level of education, or were students, were more likely to have filled out the Web-based version of the questionnaire. Given the many advantages of Web-based over paper-based questionnaires, researchers should strongly consider using Web-based questionnaires, although possible response bias when using these types of questionnaires should be taken into account. Trial Registration Nederlands Trial Register NTR2922; http://www.trialregister.nl/trialreg/admin/rctview.asp?TC=2922 (Archived by WebCite at http://www.webcitation.org/5zRRdMrDv) PMID:21955527

  13. Using web-based and paper-based questionnaires for collecting data on fertility issues among female childhood cancer survivors: differences in response characteristics.

    PubMed

    van den Berg, Marleen H; Overbeek, Annelies; van der Pal, Helena J; Versluys, A Birgitta; Bresters, Dorine; van Leeuwen, Flora E; Lambalk, Cornelis B; Kaspers, Gertjan J L; van Dulmen-den Broeder, Eline

    2011-09-29

    Web-based questionnaires have become increasingly popular in health research. However, reported response rates vary and response bias may be introduced. The aim of this study was to evaluate whether sending a mixed invitation (paper-based together with Web-based questionnaire) rather than a Web-only invitation (Web-based questionnaire only) results in higher response and participation rates for female childhood cancer survivors filling out a questionnaire on fertility issues. In addition, differences in type of response and characteristics of the responders and nonresponders were investigated. Moreover, factors influencing preferences for either the Web- or paper-based version of the questionnaire were examined. This study is part of a nationwide study on reproductive function, ovarian reserve, and risk of premature menopause in female childhood cancer survivors. The Web-based version of the questionnaire was available for participants through the Internet by means of a personalized user name and password. Participants were randomly selected to receive either a mixed invitation (paper-based questionnaire together with log-in details for Web-based questionnaire, n = 137) or a Web-only invitation (log-in details only, n = 140). Furthermore, the latter group could request a paper-based version of the questionnaire by filling out a form. Overall response rates were comparable in both randomization groups (83% mixed invitation group vs 89% in Web-only invitation group, P = .20). In addition, participation rates appeared not to differ (66% or 90/137, mixed invitation group vs 59% or 83/140, Web-only invitation group, P =.27). However, in the mixed invitation group, significantly more respondents filled out the paper-based questionnaire compared with the Web-only invitation group (83% or 75/90 and 65% or 54/83, respectively, P = .01). The 44 women who filled out the Web-based version of the questionnaire had a higher educational level than the 129 women who filled out the paper-based version (P = .01). Furthermore, the probability of filling out the Web-based questionnaire appeared to be greater for women who were allocated to the Web-only invitation group (OR = 2.85, 95% CI 1.31-6.21), were older (OR = 1.08, 95% CI 1.02-1.15), had a higher educational level (OR high vs low = 0.06, 95% CI 0.01-0.52), or were students (OR employed vs student = 3.25, 95% CI 1.00-10.56). Although overall response as well as participation rates to both types of invitations were similar, adding a paper version of a questionnaire to a Web-only invitation resulted in more respondents filling out the paper-based version. In addition, women who were older, had a higher level of education, or were students, were more likely to have filled out the Web-based version of the questionnaire. Given the many advantages of Web-based over paper-based questionnaires, researchers should strongly consider using Web-based questionnaires, although possible response bias when using these types of questionnaires should be taken into account. Nederlands Trial Register NTR2922; http://www.trialregister.nl/trialreg/admin/rctview.asp?TC=2922 (Archived by WebCite at http://www.webcitation.org/5zRRdMrDv).

  14. 76 FR 18831 - Updating Regulations Issued Under the Fair Labor Standards Act

    Federal Register 2010, 2011, 2012, 2013, 2014

    2011-04-05

    ... time zone, or log onto the WHD's Web site for a nationwide listing of Wage and Hour District and Area... credits. The Department is also not proceeding with the proposed rule that service managers, service writers, service advisors, and service salesman are exempted from the overtime provision. We have also...

  15. Increasing Parent Engagement in Student Learning Using an Intelligent Tutoring System

    ERIC Educational Resources Information Center

    Broderick, Zachary; O'Connor, Christine; Mulcahy, Courtney; Heffernan, Neil; Heffernan, Christina

    2011-01-01

    This study demonstrates the ability of an Intelligent Tutoring System (ITS) to increase parental engagement in student learning. A parent notification feature was developed for the web-based ASSISTment ITS that allows parents to log into their own accounts and access detailed data about their students' performance. Parents from a local middle…

  16. Using Learning Styles and Viewing Styles in Streaming Video

    ERIC Educational Resources Information Center

    de Boer, Jelle; Kommers, Piet A. M.; de Brock, Bert

    2011-01-01

    Improving the effectiveness of learning when students observe video lectures becomes urgent with the rising advent of (web-based) video materials. Vital questions are how students differ in their learning preferences and what patterns in viewing video can be detected in log files. Our experiments inventory students' viewing patterns while watching…

  17. 7 CFR 301.92-5 - Issuance and cancellation of certificates.

    Code of Federal Regulations, 2012 CFR

    2012-01-01

    ... obtained from the Animal and Plant Health Inspection Service, Plant Protection and Quarantine, Invasive Species and Pest Management, 4700 River Road Unit 160, Riverdale, MD 20737, or the APHIS Web site at http... or 10 Firewood, logs, lumber of species listed in 301.92-2(d) and marked with an asterisk are not...

  18. 7 CFR 301.92-5 - Issuance and cancellation of certificates.

    Code of Federal Regulations, 2014 CFR

    2014-01-01

    ... obtained from the Animal and Plant Health Inspection Service, Plant Protection and Quarantine, Invasive Species and Pest Management, 4700 River Road Unit 160, Riverdale, MD 20737, or the APHIS Web site at http... or 10 Firewood, logs, lumber of species listed in 301.92-2(d) and marked with an asterisk are not...

  19. 7 CFR 301.92-5 - Issuance and cancellation of certificates.

    Code of Federal Regulations, 2013 CFR

    2013-01-01

    ... obtained from the Animal and Plant Health Inspection Service, Plant Protection and Quarantine, Invasive Species and Pest Management, 4700 River Road Unit 160, Riverdale, MD 20737, or the APHIS Web site at http... or 10 Firewood, logs, lumber of species listed in 301.92-2(d) and marked with an asterisk are not...

  20. 7 CFR 301.92-5 - Issuance and cancellation of certificates.

    Code of Federal Regulations, 2011 CFR

    2011-01-01

    ... obtained from the Animal and Plant Health Inspection Service, Plant Protection and Quarantine, Invasive Species and Pest Management, 4700 River Road Unit 160, Riverdale, MD 20737, or the APHIS Web site at http... or 10 Firewood, logs, lumber of species listed in 301.92-2(d) and marked with an asterisk are not...

Top