Visual Based Retrieval Systems and Web Mining--Introduction.
ERIC Educational Resources Information Center
Iyengar, S. S.
2001-01-01
Briefly discusses Web mining and image retrieval techniques, and then presents a summary of articles in this special issue. Articles focus on Web content mining, artificial neural networks as tools for image retrieval, content-based image retrieval systems, and personalizing the Web browsing experience using media agents. (AEF)
HC StratoMineR: A Web-Based Tool for the Rapid Analysis of High-Content Datasets.
Omta, Wienand A; van Heesbeen, Roy G; Pagliero, Romina J; van der Velden, Lieke M; Lelieveld, Daphne; Nellen, Mehdi; Kramer, Maik; Yeong, Marley; Saeidi, Amir M; Medema, Rene H; Spruit, Marco; Brinkkemper, Sjaak; Klumperman, Judith; Egan, David A
2016-10-01
High-content screening (HCS) can generate large multidimensional datasets and when aligned with the appropriate data mining tools, it can yield valuable insights into the mechanism of action of bioactive molecules. However, easy-to-use data mining tools are not widely available, with the result that these datasets are frequently underutilized. Here, we present HC StratoMineR, a web-based tool for high-content data analysis. It is a decision-supportive platform that guides even non-expert users through a high-content data analysis workflow. HC StratoMineR is built by using My Structured Query Language for storage and querying, PHP: Hypertext Preprocessor as the main programming language, and jQuery for additional user interface functionality. R is used for statistical calculations, logic and data visualizations. Furthermore, C++ and graphical processor unit power is diffusely embedded in R by using the rcpp and rpud libraries for operations that are computationally highly intensive. We show that we can use HC StratoMineR for the analysis of multivariate data from a high-content siRNA knock-down screen and a small-molecule screen. It can be used to rapidly filter out undesirable data; to select relevant data; and to perform quality control, data reduction, data exploration, morphological hit picking, and data clustering. Our results demonstrate that HC StratoMineR can be used to functionally categorize HCS hits and, thus, provide valuable information for hit prioritization.
Informal Learning through Expertise Mining in the Social Web
ERIC Educational Resources Information Center
Valencia-Garcia, Rafael; Garcia-Sanchez, Francisco; Casado-Lumbreras, Cristina; Castellanos-Nieves, Dagoberto; Fernandez-Breis, Jesualdo Tomas
2012-01-01
The advent of Web 2.0, also called the Social Web, has changed the way people interact with the Web. Assisted by the technologies associated with this new trend, users now play a much more active role as content providers. This Web paradigm shift has also changed how companies operate and interact with their employees, partners and customers. The…
ERIC Educational Resources Information Center
Sathick, Javubar; Venkat, Jaya
2015-01-01
Mining social web data is a challenging task and finding user interest for personalized and non-personalized recommendation systems is another important task. Knowledge sharing among web users has become crucial in determining usage of web data and personalizing content in various social websites as per the user's wish. This paper aims to design a…
Types of Online Hierarchical Repository Structures
ERIC Educational Resources Information Center
Hershkovitz, Arnon; Azran, Ronit; Hardof-Jaffe, Sharon; Nachmias, Rafi
2011-01-01
This study presents an empirical investigation of online hierarchical repositories of items presented to university students in Web-supported course websites, using Web mining methods. To this end, data from 1747 courses were collected, and the use of online repositories of content items in these courses was examined. At a later stage, courses…
RCrawler: An R package for parallel web crawling and scraping
NASA Astrophysics Data System (ADS)
Khalil, Salim; Fakir, Mohamed
RCrawler is a contributed R package for domain-based web crawling and content scraping. As the first implementation of a parallel web crawler in the R environment, RCrawler can crawl, parse, store pages, extract contents, and produce data that can be directly employed for web content mining applications. However, it is also flexible, and could be adapted to other applications. The main features of RCrawler are multi-threaded crawling, content extraction, and duplicate content detection. In addition, it includes functionalities such as URL and content-type filtering, depth level controlling, and a robot.txt parser. Our crawler has a highly optimized system, and can download a large number of pages per second while being robust against certain crashes and spider traps. In this paper, we describe the design and functionality of RCrawler, and report on our experience of implementing it in an R environment, including different optimizations that handle the limitations of R. Finally, we discuss our experimental results.
Using Open Web APIs in Teaching Web Mining
ERIC Educational Resources Information Center
Chen, Hsinchun; Li, Xin; Chau, M.; Ho, Yi-Jen; Tseng, Chunju
2009-01-01
With the advent of the World Wide Web, many business applications that utilize data mining and text mining techniques to extract useful business information on the Web have evolved from Web searching to Web mining. It is important for students to acquire knowledge and hands-on experience in Web mining during their education in information systems…
Multiple-Feature Extracting Modules Based Leak Mining System Design
Cho, Ying-Chiang; Pan, Jen-Yi
2013-01-01
Over the years, human dependence on the Internet has increased dramatically. A large amount of information is placed on the Internet and retrieved from it daily, which makes web security in terms of online information a major concern. In recent years, the most problematic issues in web security have been e-mail address leakage and SQL injection attacks. There are many possible causes of information leakage, such as inadequate precautions during the programming process, which lead to the leakage of e-mail addresses entered online or insufficient protection of database information, a loophole that enables malicious users to steal online content. In this paper, we implement a crawler mining system that is equipped with SQL injection vulnerability detection, by means of an algorithm developed for the web crawler. In addition, we analyze portal sites of the governments of various countries or regions in order to investigate the information leaking status of each site. Subsequently, we analyze the database structure and content of each site, using the data collected. Thus, we make use of practical verification in order to focus on information security and privacy through black-box testing. PMID:24453892
Multiple-feature extracting modules based leak mining system design.
Cho, Ying-Chiang; Pan, Jen-Yi
2013-01-01
Over the years, human dependence on the Internet has increased dramatically. A large amount of information is placed on the Internet and retrieved from it daily, which makes web security in terms of online information a major concern. In recent years, the most problematic issues in web security have been e-mail address leakage and SQL injection attacks. There are many possible causes of information leakage, such as inadequate precautions during the programming process, which lead to the leakage of e-mail addresses entered online or insufficient protection of database information, a loophole that enables malicious users to steal online content. In this paper, we implement a crawler mining system that is equipped with SQL injection vulnerability detection, by means of an algorithm developed for the web crawler. In addition, we analyze portal sites of the governments of various countries or regions in order to investigate the information leaking status of each site. Subsequently, we analyze the database structure and content of each site, using the data collected. Thus, we make use of practical verification in order to focus on information security and privacy through black-box testing.
Web mining in soft computing framework: relevance, state of the art and future directions.
Pal, S K; Talwar, V; Mitra, P
2002-01-01
The paper summarizes the different characteristics of Web data, the basic components of Web mining and its different types, and the current state of the art. The reason for considering Web mining, a separate field from data mining, is explained. The limitations of some of the existing Web mining methods and tools are enunciated, and the significance of soft computing (comprising fuzzy logic (FL), artificial neural networks (ANNs), genetic algorithms (GAs), and rough sets (RSs) are highlighted. A survey of the existing literature on "soft Web mining" is provided along with the commercially available systems. The prospective areas of Web mining where the application of soft computing needs immediate attention are outlined with justification. Scope for future research in developing "soft Web mining" systems is explained. An extensive bibliography is also provided.
Applying Web Usage Mining for Personalizing Hyperlinks in Web-Based Adaptive Educational Systems
ERIC Educational Resources Information Center
Romero, Cristobal; Ventura, Sebastian; Zafra, Amelia; de Bra, Paul
2009-01-01
Nowadays, the application of Web mining techniques in e-learning and Web-based adaptive educational systems is increasing exponentially. In this paper, we propose an advanced architecture for a personalization system to facilitate Web mining. A specific Web mining tool is developed and a recommender engine is integrated into the AHA! system in…
Improving entrepreneurial opportunity recognition through web content analytics
NASA Astrophysics Data System (ADS)
Bakar, Muhamad Shahbani Abu; Azmi, Azwiyati
2017-10-01
The ability to recognize and develop an opportunity into a venture defines an entrepreneur. Research in opportunity recognition has been robust and focuses more on explaining the processes involved in opportunity recognition. Factors such as prior knowledge, cognitive and creative capabilities are shown to affect opportunity recognition in entrepreneurs. Prior knowledge in areas such as customer problems, ways to serve the market, and technology has been shows in various studies to be a factor that facilitates entrepreneurs to identify and recognize opportunities. Findings from research also shows that experienced entrepreneurs search and scan for information to discover opportunities. Searching and scanning for information has also been shown to help novice entrepreneurs who lack prior knowledge to narrow this gap and enable them to better identify and recognize opportunities. There is less focus in research on finding empirically proven techniques and methods to develop and enhance opportunity recognition in student entrepreneurs. This is important as the country pushes for more graduate entrepreneurs that can drive the economy. This paper aims to discuss Opportunity Recognition Support System (ORSS), an information support system to help especially student entrepreneurs in identifying and recognizing business opportunities. The ORSS aims to provide the necessary knowledge to student entrepreneurs to be able to better identify and recognize opportunities. Applying design research, theories in opportunity recognition are applied to identify the requirements for the support system and the requirements in turn dictate the design of the support system. The paper proposes the use of web content mining and analytics as two core components and techniques for the support system. Web content mining can mine the vast knowledge repositories available on the internet and analytics can provide entrepreneurs with further insights into the information needed to recognize opportunities in a given market or industry.
Study on Personalized Recommendation Model of Internet Advertisement
NASA Astrophysics Data System (ADS)
Zhou, Ning; Chen, Yongyue; Zhang, Huiping
With the rapid development of E-Commerce, the audiences put forward higher requirements on personalized Internet advertisement than before. The main function of Personalized Advertising System is to provide the most suitable advertisements for anonymous users on Web sites. The paper offers a personalized Internet advertisement recommendation model. By mining the audiences' historical and current behavior, and the advertisers' and publisher's web site content, etc, the system can recommend appropriate advertisements to corresponding audiences.
TOPSAN: a dynamic web database for structural genomics.
Ellrott, Kyle; Zmasek, Christian M; Weekes, Dana; Sri Krishna, S; Bakolitsa, Constantina; Godzik, Adam; Wooley, John
2011-01-01
The Open Protein Structure Annotation Network (TOPSAN) is a web-based collaboration platform for exploring and annotating structures determined by structural genomics efforts. Characterization of those structures presents a challenge since the majority of the proteins themselves have not yet been characterized. Responding to this challenge, the TOPSAN platform facilitates collaborative annotation and investigation via a user-friendly web-based interface pre-populated with automatically generated information. Semantic web technologies expand and enrich TOPSAN's content through links to larger sets of related databases, and thus, enable data integration from disparate sources and data mining via conventional query languages. TOPSAN can be found at http://www.topsan.org.
Web Mining for Web Image Retrieval.
ERIC Educational Resources Information Center
Chen, Zheng; Wenyin, Liu; Zhang, Feng; Li, Mingjing; Zhang, Hongjiang
2001-01-01
Presents a prototype system for image retrieval from the Internet using Web mining. Discusses the architecture of the Web image retrieval prototype; document space modeling; user log mining; and image retrieval experiments to evaluate the proposed system. (AEF)
Exploring Characterizations of Learning Object Repositories Using Data Mining Techniques
NASA Astrophysics Data System (ADS)
Segura, Alejandra; Vidal, Christian; Menendez, Victor; Zapata, Alfredo; Prieto, Manuel
Learning object repositories provide a platform for the sharing of Web-based educational resources. As these repositories evolve independently, it is difficult for users to have a clear picture of the kind of contents they give access to. Metadata can be used to automatically extract a characterization of these resources by using machine learning techniques. This paper presents an exploratory study carried out in the contents of four public repositories that uses clustering and association rule mining algorithms to extract characterizations of repository contents. The results of the analysis include potential relationships between different attributes of learning objects that may be useful to gain an understanding of the kind of resources available and eventually develop search mechanisms that consider repository descriptions as a criteria in federated search.
Study on online community user motif using web usage mining
NASA Astrophysics Data System (ADS)
Alphy, Meera; Sharma, Ajay
2016-04-01
The Web usage mining is the application of data mining, which is used to extract useful information from the online community. The World Wide Web contains at least 4.73 billion pages according to Indexed Web and it contains at least 228.52 million pages according Dutch Indexed web on 6th august 2015, Thursday. It’s difficult to get needed data from these billions of web pages in World Wide Web. Here is the importance of web usage mining. Personalizing the search engine helps the web user to identify the most used data in an easy way. It reduces the time consumption; automatic site search and automatic restore the useful sites. This study represents the old techniques to latest techniques used in pattern discovery and analysis in web usage mining from 1996 to 2015. Analyzing user motif helps in the improvement of business, e-commerce, personalisation and improvement of websites.
The design and implementation of web mining in web sites security
NASA Astrophysics Data System (ADS)
Li, Jian; Zhang, Guo-Yin; Gu, Guo-Chang; Li, Jian-Li
2003-06-01
The backdoor or information leak of Web servers can be detected by using Web Mining techniques on some abnormal Web log and Web application log data. The security of Web servers can be enhanced and the damage of illegal access can be avoided. Firstly, the system for discovering the patterns of information leakages in CGI scripts from Web log data was proposed. Secondly, those patterns for system administrators to modify their codes and enhance their Web site security were provided. The following aspects were described: one is to combine web application log with web log to extract more information, so web data mining could be used to mine web log for discovering the information that firewall and Information Detection System cannot find. Another approach is to propose an operation module of web site to enhance Web site security. In cluster server session, Density-Based Clustering technique is used to reduce resource cost and obtain better efficiency.
Topic Models for Link Prediction in Document Networks
ERIC Educational Resources Information Center
Kataria, Saurabh
2012-01-01
Recent explosive growth of interconnected document collections such as citation networks, network of web pages, content generated by crowd-sourcing in collaborative environments, etc., has posed several challenging problems for data mining and machine learning community. One central problem in the domain of document networks is that of "link…
A Framework for Web Usage Mining in Electronic Government
NASA Astrophysics Data System (ADS)
Zhou, Ping; Le, Zhongjian
Web usage mining has been a major component of management strategy to enhance organizational analysis and decision. The literature on Web usage mining that deals with strategies and technologies for effectively employing Web usage mining is quite vast. In recent years, E-government has received much attention from researchers and practitioners. Huge amounts of user access data are produced in Electronic government Web site everyday. The role of these data in the success of government management cannot be overstated because they affect government analysis, prediction, strategies, tactical, operational planning and control. Web usage miming in E-government has an important role to play in setting government objectives, discovering citizen behavior, and determining future courses of actions. Web usage mining in E-government has not received adequate attention from researchers or practitioners. We developed a framework to promote a better understanding of the importance of Web usage mining in E-government. Using the current literature, we developed the framework presented herein, in hopes that it would stimulate more interest in this important area.
ERIC Educational Resources Information Center
Chen, Hsinchun
2003-01-01
Discusses information retrieval techniques used on the World Wide Web. Topics include machine learning in information extraction; relevance feedback; information filtering and recommendation; text classification and text clustering; Web mining, based on data mining techniques; hyperlink structure; and Web size. (LRW)
PDBj Mine: design and implementation of relational database interface for Protein Data Bank Japan
Kinjo, Akira R.; Yamashita, Reiko; Nakamura, Haruki
2010-01-01
This article is a tutorial for PDBj Mine, a new database and its interface for Protein Data Bank Japan (PDBj). In PDBj Mine, data are loaded from files in the PDBMLplus format (an extension of PDBML, PDB's canonical XML format, enriched with annotations), which are then served for the user of PDBj via the worldwide web (WWW). We describe the basic design of the relational database (RDB) and web interfaces of PDBj Mine. The contents of PDBMLplus files are first broken into XPath entities, and these paths and data are indexed in the way that reflects the hierarchical structure of the XML files. The data for each XPath type are saved into the corresponding relational table that is named as the XPath itself. The generation of table definitions from the PDBMLplus XML schema is fully automated. For efficient search, frequently queried terms are compiled into a brief summary table. Casual users can perform simple keyword search, and 'Advanced Search' which can specify various conditions on the entries. More experienced users can query the database using SQL statements which can be constructed in a uniform manner. Thus, PDBj Mine achieves a combination of the flexibility of XML documents and the robustness of the RDB. Database URL: http://www.pdbj.org/ PMID:20798081
PDBj Mine: design and implementation of relational database interface for Protein Data Bank Japan.
Kinjo, Akira R; Yamashita, Reiko; Nakamura, Haruki
2010-08-25
This article is a tutorial for PDBj Mine, a new database and its interface for Protein Data Bank Japan (PDBj). In PDBj Mine, data are loaded from files in the PDBMLplus format (an extension of PDBML, PDB's canonical XML format, enriched with annotations), which are then served for the user of PDBj via the worldwide web (WWW). We describe the basic design of the relational database (RDB) and web interfaces of PDBj Mine. The contents of PDBMLplus files are first broken into XPath entities, and these paths and data are indexed in the way that reflects the hierarchical structure of the XML files. The data for each XPath type are saved into the corresponding relational table that is named as the XPath itself. The generation of table definitions from the PDBMLplus XML schema is fully automated. For efficient search, frequently queried terms are compiled into a brief summary table. Casual users can perform simple keyword search, and 'Advanced Search' which can specify various conditions on the entries. More experienced users can query the database using SQL statements which can be constructed in a uniform manner. Thus, PDBj Mine achieves a combination of the flexibility of XML documents and the robustness of the RDB. Database URL: http://www.pdbj.org/
Multi-Filter String Matching and Human-Centric Entity Matching for Information Extraction
ERIC Educational Resources Information Center
Sun, Chong
2012-01-01
More and more information is being generated in text documents, such as Web pages, emails and blogs. To effectively manage this unstructured information, one broadly used approach includes locating relevant content in documents, extracting structured information and integrating the extracted information for querying, mining or further analysis. In…
NASA Astrophysics Data System (ADS)
Madiraju, Praveen; Zhang, Yanqing
2002-03-01
When a user logs in to a website, behind the scenes the user leaves his/her impressions, usage patterns and also access patterns in the web servers log file. A web usage mining agent can analyze these web logs to help web developers to improve the organization and presentation of their websites. They can help system administrators in improving the system performance. Web logs provide invaluable help in creating adaptive web sites and also in analyzing the network traffic analysis. This paper presents the design and implementation of a Web usage mining agent for digging in to the web log files.
The Comprehensive Microbial Resource.
Peterson, J D; Umayam, L A; Dickinson, T; Hickey, E K; White, O
2001-01-01
One challenge presented by large-scale genome sequencing efforts is effective display of uniform information to the scientific community. The Comprehensive Microbial Resource (CMR) contains robust annotation of all complete microbial genomes and allows for a wide variety of data retrievals. The bacterial information has been placed on the Web at http://www.tigr.org/CMR for retrieval using standard web browsing technology. Retrievals can be based on protein properties such as molecular weight or hydrophobicity, GC-content, functional role assignments and taxonomy. The CMR also has special web-based tools to allow data mining using pre-run homology searches, whole genome dot-plots, batch downloading and traversal across genomes using a variety of datatypes.
Effect of Temporal Relationships in Associative Rule Mining for Web Log Data
Mohd Khairudin, Nazli; Mustapha, Aida
2014-01-01
The advent of web-based applications and services has created such diverse and voluminous web log data stored in web servers, proxy servers, client machines, or organizational databases. This paper attempts to investigate the effect of temporal attribute in relational rule mining for web log data. We incorporated the characteristics of time in the rule mining process and analysed the effect of various temporal parameters. The rules generated from temporal relational rule mining are then compared against the rules generated from the classical rule mining approach such as the Apriori and FP-Growth algorithms. The results showed that by incorporating the temporal attribute via time, the number of rules generated is subsequently smaller but is comparable in terms of quality. PMID:24587757
ERIC Educational Resources Information Center
Qin, Jian; Jurisica, Igor; Liddy, Elizabeth D.; Jansen, Bernard J; Spink, Amanda; Priss, Uta; Norton, Melanie J.
2000-01-01
These six articles discuss knowledge discovery in databases (KDD). Topics include data mining; knowledge management systems; applications of knowledge discovery; text and Web mining; text mining and information retrieval; user search patterns through Web log analysis; concept analysis; data collection; and data structure inconsistency. (LRW)
Diamond Eye: a distributed architecture for image data mining
NASA Astrophysics Data System (ADS)
Burl, Michael C.; Fowlkes, Charless; Roden, Joe; Stechert, Andre; Mukhtar, Saleem
1999-02-01
Diamond Eye is a distributed software architecture, which enables users (scientists) to analyze large image collections by interacting with one or more custom data mining servers via a Java applet interface. Each server is coupled with an object-oriented database and a computational engine, such as a network of high-performance workstations. The database provides persistent storage and supports querying of the 'mined' information. The computational engine provides parallel execution of expensive image processing, object recognition, and query-by-content operations. Key benefits of the Diamond Eye architecture are: (1) the design promotes trial evaluation of advanced data mining and machine learning techniques by potential new users (all that is required is to point a web browser to the appropriate URL), (2) software infrastructure that is common across a range of science mining applications is factored out and reused, and (3) the system facilitates closer collaborations between algorithm developers and domain experts.
Stewart, A.R.; Saiki, M.K.; Kuwabara, J.S.; Alpers, Charles N.; Marvin-DiPasquale, M.; Krabbenhoft, D.P.
2008-01-01
Physical and biogeochemical characteristics of the aquatic environment that affect growth dynamics of phytoplankton and the zooplankton communities that depend on them may also affect uptake of methylmercury (MeHg) into the pelagic food web of oligotrophic reservoirs. We evaluated changes in the quality and quantity of suspended particulate material, zooplankton taxonomy, and MeHg concentrations coincident with seasonal changes in water storage of a mining-impacted reservoir in northern California, USA. MeHg concentrations in bulk zooplankton increased from 4 ng??g-1 at low water to 77 ?? 6.1 ng??g-1 at high water and were positively correlated with cladoceran biomass (r = 0.66) and negatively correlated with rotifer biomass (r = -0.65). Stable isotope analysis revealed overall higher MeHg concentrations in the pelagic-based food web relative to the benthic-based food web. Statistically similar patterns of trophic enrichment of MeHg (slopes) for the pelagic and benthic food webs and slightly higher MeHg concentrations in zooplankton than in benthic invertebrates suggest that the difference in MeHg bioaccumulation among trophic pathways is set at the base of the food webs. These results suggest an important role for plankton dynamics in driving the MeHg content of zooplankton and ultimately MeHg bioaccumulation in top predators in pelagic-based food webs. ?? 2008 NRC.
Keynote Talk: Mining the Web 2.0 for Improved Image Search
NASA Astrophysics Data System (ADS)
Baeza-Yates, Ricardo
There are several semantic sources that can be found in the Web that are either explicit, e.g. Wikipedia, or implicit, e.g. derived from Web usage data. Most of them are related to user generated content (UGC) or what is called today the Web 2.0. In this talk we show how to use these sources of evidence in Flickr, such as tags, visual annotations or clicks, which represent the the wisdom of crowds behind UGC, to improve image search. These results are the work of the multimedia retrieval team at Yahoo! Research Barcelona and they are already being used in Yahoo! image search. This work is part of a larger effort to produce a virtuous data feedback circuit based on the right combination many different technologies to leverage the Web itself.
Evaluation of the Kloswall longwall mining system
NASA Astrophysics Data System (ADS)
Guay, P. J.
1982-04-01
A new longwal mining system specifically designed to extract a very deep web (48 inches or deeper) from a longwall panel was studied. Productivity and cost analysis comparing the new mining system with a conventional longwall operation taking a 30 inch wide web is presented. It is shown that the new system will increase annual production and return on investment in most cases. Conceptual drawings and specifications for a high capacity three drum shearer and a unique shield type of roof support specifically designed for very wide web operation are reported. The advantages and problems associated with wide web mining in general and as they relate specifically to the equipment selected for the new mining system are discussed.
The Comprehensive Microbial Resource
Peterson, Jeremy D.; Umayam, Lowell A.; Dickinson, Tanja; Hickey, Erin K.; White, Owen
2001-01-01
One challenge presented by large-scale genome sequencing efforts is effective display of uniform information to the scientific community. The Comprehensive Microbial Resource (CMR) contains robust annotation of all complete microbial genomes and allows for a wide variety of data retrievals. The bacterial information has been placed on the Web at http://www.tigr.org/CMR for retrieval using standard web browsing technology. Retrievals can be based on protein properties such as molecular weight or hydrophobicity, GC-content, functional role assignments and taxonomy. The CMR also has special web-based tools to allow data mining using pre-run homology searches, whole genome dot-plots, batch downloading and traversal across genomes using a variety of datatypes. PMID:11125067
Web Mining: Machine Learning for Web Applications.
ERIC Educational Resources Information Center
Chen, Hsinchun; Chau, Michael
2004-01-01
Presents an overview of machine learning research and reviews methods used for evaluating machine learning systems. Ways that machine-learning algorithms were used in traditional information retrieval systems in the "pre-Web" era are described, and the field of Web mining and how machine learning has been used in different Web mining…
NASA Astrophysics Data System (ADS)
Bini, Claudio; Maleci, Laura; Buffa, Gabriella; Wahsha, Mohammad; Fontana, Silvia
2013-04-01
Response of dandelion (Taraxacum officinale Web) to heavy metals from mine sites: micromorphology of leaves and roots. Maleci L.1 , Bini C.2, Buffa G. 2, Fontana S2., Wahsha M.3 1 - Dept of Biology, University of Florence, Italy. 2 - Dept of Environmental Sciences, Informatics and Statistics. Ca'Foscari University, Venice - Italy. 3 - Marine Science Centre - University of Jordan, Aqaba section, Jordan. Heavy metal accumulation is known to produce significant physiological and biochemical responses in vascular plants. Yet, metabolic and physiological responses of plants to heavy metal concentration can be viewed as potentially adaptive changes of the plants during stress. From this point of view, plants growing on abandoned mine sites are of particular interest, since they are genetically tolerant to high metal concentrations, and can be utilized in soil restoration. Among wild plants, the common dandelion (Taraxacum officinale Web) has received attention as bioindicator plant, and has been also suggested in remediation projects. Wild specimens of Taraxacum officinale Web, with their soil clod, were gathered from three sites with different contamination levels by heavy metals (Cd, Cr, Cu, Fe, Pb, Zn) in the abandoned Imperina Valley mine (Northeast Italy). A control plant was also gathered from a not contaminated site nearby. Plants were cultivated in pots for one year at HBF, and appeared macroscopically not affected by toxic signals (reduced growth, leaf necrosis) possibly induced by soil HM concentration. Leaves and roots taken at the same growing season were observed by LM and TEM. Light microscopy observations carried out on the leaf lamina show a clear difference in the cellular organization of not-contaminated and contaminated samples. The unpolluted samples present a well organized palisade tissue and spongy photosynthetic parenchyma. Samples from contaminated sites, instead, present a palisade parenchyma less organized, and a reduction of leaf thickness proportional to HM concentration. Indeed, at high HM contents, leaf parenchyma is constituted of few roundish cells with large intercellular spaces, while palisade structure is lacking at all. Comparing the leaf morphology with their metal content, it appears that the poor structural organisation, and the reduced foliar thickness of the contaminated plants, are strictly related to soil contamination. Similar observations have been recorded on cortex parenchyma of the roots, which presents a reduced thickness in comparison to the control, proportional to HM content in the soil. Moreover, all the samples examined do not present hairs on the root epidermis, but mycorrhizae, which are well developed in the control, and nearly lacking in the contaminated samples. Preliminary ultrastructure observations of the parenchyma cells of contaminated samples show mitochondrial structure alteration, with lacking or reduced cristae of the internal membrane at increasing metal content, in comparison to the not-contaminated sample. Instead, chloroplast organization does not present significant differences, particularly in number and compartmentalization of thylacoids. Although macromorphology does not present evidence of phytotoxicity, the recorded observations of the micromorphological characteristics of leaves and roots, show a suffering state strictly related to HM content. However, T. officinale, besides the recorded abnormalities, proved to be able to grow on moderately contaminated soils, and therefore may be utilized to colonize polluted sites.
Public health, GIS, and the internet.
Croner, Charles M
2003-01-01
Internet access and use of georeferenced public health information for GIS application will be an important and exciting development for the nation's Department of Health and Human Services and other health agencies in this new millennium. Technological progress toward public health geospatial data integration, analysis, and visualization of space-time events using the Web portends eventual robust use of GIS by public health and other sectors of the economy. Increasing Web resources from distributed spatial data portals and global geospatial libraries, and a growing suite of Web integration tools, will provide new opportunities to advance disease surveillance, control, and prevention, and insure public access and community empowerment in public health decision making. Emerging supercomputing, data mining, compression, and transmission technologies will play increasingly critical roles in national emergency, catastrophic planning and response, and risk management. Web-enabled public health GIS will be guided by Federal Geographic Data Committee spatial metadata, OpenGIS Web interoperability, and GML/XML geospatial Web content standards. Public health will become a responsive and integral part of the National Spatial Data Infrastructure.
Web-video-mining-supported workflow modeling for laparoscopic surgeries.
Liu, Rui; Zhang, Xiaoli; Zhang, Hao
2016-11-01
As quality assurance is of strong concern in advanced surgeries, intelligent surgical systems are expected to have knowledge such as the knowledge of the surgical workflow model (SWM) to support their intuitive cooperation with surgeons. For generating a robust and reliable SWM, a large amount of training data is required. However, training data collected by physically recording surgery operations is often limited and data collection is time-consuming and labor-intensive, severely influencing knowledge scalability of the surgical systems. The objective of this research is to solve the knowledge scalability problem in surgical workflow modeling with a low cost and labor efficient way. A novel web-video-mining-supported surgical workflow modeling (webSWM) method is developed. A novel video quality analysis method based on topic analysis and sentiment analysis techniques is developed to select high-quality videos from abundant and noisy web videos. A statistical learning method is then used to build the workflow model based on the selected videos. To test the effectiveness of the webSWM method, 250 web videos were mined to generate a surgical workflow for the robotic cholecystectomy surgery. The generated workflow was evaluated by 4 web-retrieved videos and 4 operation-room-recorded videos, respectively. The evaluation results (video selection consistency n-index ≥0.60; surgical workflow matching degree ≥0.84) proved the effectiveness of the webSWM method in generating robust and reliable SWM knowledge by mining web videos. With the webSWM method, abundant web videos were selected and a reliable SWM was modeled in a short time with low labor cost. Satisfied performances in mining web videos and learning surgery-related knowledge show that the webSWM method is promising in scaling knowledge for intelligent surgical systems. Copyright © 2016 Elsevier B.V. All rights reserved.
OntoGene web services for biomedical text mining.
Rinaldi, Fabio; Clematide, Simon; Marques, Hernani; Ellendorff, Tilia; Romacker, Martin; Rodriguez-Esteban, Raul
2014-01-01
Text mining services are rapidly becoming a crucial component of various knowledge management pipelines, for example in the process of database curation, or for exploration and enrichment of biomedical data within the pharmaceutical industry. Traditional architectures, based on monolithic applications, do not offer sufficient flexibility for a wide range of use case scenarios, and therefore open architectures, as provided by web services, are attracting increased interest. We present an approach towards providing advanced text mining capabilities through web services, using a recently proposed standard for textual data interchange (BioC). The web services leverage a state-of-the-art platform for text mining (OntoGene) which has been tested in several community-organized evaluation challenges,with top ranked results in several of them.
OntoGene web services for biomedical text mining
2014-01-01
Text mining services are rapidly becoming a crucial component of various knowledge management pipelines, for example in the process of database curation, or for exploration and enrichment of biomedical data within the pharmaceutical industry. Traditional architectures, based on monolithic applications, do not offer sufficient flexibility for a wide range of use case scenarios, and therefore open architectures, as provided by web services, are attracting increased interest. We present an approach towards providing advanced text mining capabilities through web services, using a recently proposed standard for textual data interchange (BioC). The web services leverage a state-of-the-art platform for text mining (OntoGene) which has been tested in several community-organized evaluation challenges, with top ranked results in several of them. PMID:25472638
Automatic generation of Web mining environments
NASA Astrophysics Data System (ADS)
Cibelli, Maurizio; Costagliola, Gennaro
1999-02-01
The main problem related to the retrieval of information from the world wide web is the enormous number of unstructured documents and resources, i.e., the difficulty of locating and tracking appropriate sources. This paper presents a web mining environment (WME), which is capable of finding, extracting and structuring information related to a particular domain from web documents, using general purpose indices. The WME architecture includes a web engine filter (WEF), to sort and reduce the answer set returned by a web engine, a data source pre-processor (DSP), which processes html layout cues in order to collect and qualify page segments, and a heuristic-based information extraction system (HIES), to finally retrieve the required data. Furthermore, we present a web mining environment generator, WMEG, that allows naive users to generate a WME specific to a given domain by providing a set of specifications.
Towards the Geospatial Web: Media Platforms for Managing Geotagged Knowledge Repositories
NASA Astrophysics Data System (ADS)
Scharl, Arno
International media have recognized the visual appeal of geo-browsers such as NASA World Wind and Google Earth, for example, when Web and television coverage on Hurricane Katrina used interactive geospatial projections to illustrate its path and the scale of destruction in August 2005. Yet these early applications only hint at the true potential of geospatial technology to build and maintain virtual communities and to revolutionize the production, distribution and consumption of media products. This chapter investigates this potential by reviewing the literature and discussing the integration of geospatial and semantic reference systems, with an emphasis on extracting geospatial context from unstructured text. A content analysis of news coverage based on a suite of text mining tools (webLyzard) sheds light on the popularity and adoption of geospatial platforms.
Mining a Web Citation Database for Author Co-Citation Analysis.
ERIC Educational Resources Information Center
He, Yulan; Hui, Siu Cheung
2002-01-01
Proposes a mining process to automate author co-citation analysis based on the Web Citation Database, a data warehouse for storing citation indices of Web publications. Describes the use of agglomerative hierarchical clustering for author clustering and multidimensional scaling for displaying author cluster maps, and explains PubSearch, a…
Tourassi, Georgia; Yoon, Hong-Jun; Xu, Songhua; ...
2015-11-27
Background: The World Wide Web has emerged as a powerful data source for epidemiological studies related to infectious disease surveillance. However, its potential for cancer-related epidemiological discoveries is largely unexplored. Methods: Using advanced web crawling and tailored information extraction procedures we automatically collected and analyzed the text content of 79,394 online obituary articles published between 1998 and 2014. The collected data included 51,911 cancer (27,330 breast; 9,470 lung; 6,496 pancreatic; 6,342 ovarian; 2,273 colon) and 27,483 non-cancer cases. With the derived information, we replicated a case-control study design to investigate the association between parity and cancer risk. Age-adjusted odds ratiosmore » (ORs) with 95% confidence intervals (CIs) were calculated for each cancer type and compared to those reported in large-scale epidemiological studies. Results: Parity was found to be associated with a significantly reduced risk of breast cancer (OR=0.78, 95% CI = 0.75 to 0.82), pancreatic cancer (OR=0.78, 95% CI = 0.72 to 0.83), colon cancer (OR=0.67, 95% CI = 0.60 to 0.74), and ovarian cancer (OR=0.58, 95% CI = 0.54 to 0.62). Marginal association was found for lung cancer prevalence (OR=0.87, 95% CI = 0.81 to 0.92). The linear trend between multi-parity and reduced cancer risk was dramatically more pronounced for breast and ovarian cancer than the other cancers included in the analysis. Conclusion: This large web-mining study on parity and cancer risk produced findings very similar to those reported with traditional observational studies. It may be used as a promising strategy to generate study hypotheses for guiding and prioritizing future epidemiological studies.« less
DOE Office of Scientific and Technical Information (OSTI.GOV)
Tourassi, Georgia; Yoon, Hong-Jun; Xu, Songhua
Background: The World Wide Web has emerged as a powerful data source for epidemiological studies related to infectious disease surveillance. However, its potential for cancer-related epidemiological discoveries is largely unexplored. Methods: Using advanced web crawling and tailored information extraction procedures we automatically collected and analyzed the text content of 79,394 online obituary articles published between 1998 and 2014. The collected data included 51,911 cancer (27,330 breast; 9,470 lung; 6,496 pancreatic; 6,342 ovarian; 2,273 colon) and 27,483 non-cancer cases. With the derived information, we replicated a case-control study design to investigate the association between parity and cancer risk. Age-adjusted odds ratiosmore » (ORs) with 95% confidence intervals (CIs) were calculated for each cancer type and compared to those reported in large-scale epidemiological studies. Results: Parity was found to be associated with a significantly reduced risk of breast cancer (OR=0.78, 95% CI = 0.75 to 0.82), pancreatic cancer (OR=0.78, 95% CI = 0.72 to 0.83), colon cancer (OR=0.67, 95% CI = 0.60 to 0.74), and ovarian cancer (OR=0.58, 95% CI = 0.54 to 0.62). Marginal association was found for lung cancer prevalence (OR=0.87, 95% CI = 0.81 to 0.92). The linear trend between multi-parity and reduced cancer risk was dramatically more pronounced for breast and ovarian cancer than the other cancers included in the analysis. Conclusion: This large web-mining study on parity and cancer risk produced findings very similar to those reported with traditional observational studies. It may be used as a promising strategy to generate study hypotheses for guiding and prioritizing future epidemiological studies.« less
A comprehensive review on privacy preserving data mining.
Aldeen, Yousra Abdul Alsahib S; Salleh, Mazleena; Razzaque, Mohammad Abdur
2015-01-01
Preservation of privacy in data mining has emerged as an absolute prerequisite for exchanging confidential information in terms of data analysis, validation, and publishing. Ever-escalating internet phishing posed severe threat on widespread propagation of sensitive information over the web. Conversely, the dubious feelings and contentions mediated unwillingness of various information providers towards the reliability protection of data from disclosure often results utter rejection in data sharing or incorrect information sharing. This article provides a panoramic overview on new perspective and systematic interpretation of a list published literatures via their meticulous organization in subcategories. The fundamental notions of the existing privacy preserving data mining methods, their merits, and shortcomings are presented. The current privacy preserving data mining techniques are classified based on distortion, association rule, hide association rule, taxonomy, clustering, associative classification, outsourced data mining, distributed, and k-anonymity, where their notable advantages and disadvantages are emphasized. This careful scrutiny reveals the past development, present research challenges, future trends, the gaps and weaknesses. Further significant enhancements for more robust privacy protection and preservation are affirmed to be mandatory.
Data Mining for Web-Based Support Systems: A Case Study in e-Custom Systems
NASA Astrophysics Data System (ADS)
Razmerita, Liana; Kirchner, Kathrin
This chapter provides an example of a Web-based support system (WSS) used to streamline trade procedures, prevent potential security threats, and reduce tax-related fraud in cross-border trade. The architecture is based on a service-oriented architecture that includes smart seals and Web services. We discuss the implications and suggest further enhancements to demonstrate how such systems can move toward a Web-based decision support system with the support of data mining methods. We provide a concrete example of how data mining can help to analyze the vast amount of data collected while monitoring the container movements along its supply chain.
Semantic web for integrated network analysis in biomedicine.
Chen, Huajun; Ding, Li; Wu, Zhaohui; Yu, Tong; Dhanapalan, Lavanya; Chen, Jake Y
2009-03-01
The Semantic Web technology enables integration of heterogeneous data on the World Wide Web by making the semantics of data explicit through formal ontologies. In this article, we survey the feasibility and state of the art of utilizing the Semantic Web technology to represent, integrate and analyze the knowledge in various biomedical networks. We introduce a new conceptual framework, semantic graph mining, to enable researchers to integrate graph mining with ontology reasoning in network data analysis. Through four case studies, we demonstrate how semantic graph mining can be applied to the analysis of disease-causal genes, Gene Ontology category cross-talks, drug efficacy analysis and herb-drug interactions analysis.
Towards an Enhanced Aspect-based Contradiction Detection Approach for Online Review Content
NASA Astrophysics Data System (ADS)
Nuradilah Azman, Siti; Ishak, Iskandar; Sharef, Nurfadhlina Mohd; Sidi, Fatimah
2017-09-01
User generated content as such online reviews plays an important role in customer’s purchase decisions. Many works have focused on identifying satisfaction of the reviewer in social media through the study of sentiment analysis (SA) and opinion mining. The large amount of potential application and the increasing number of opinions expresses on the web results in researchers interest on sentiment analysis and opinion mining. However, due to the reviewer’s idiosyncrasy, reviewer may have different preferences and point of view for a particular subject which in this case hotel reviews. There is still limited research that focuses on this contradiction detection in the perspective of tourism online review especially in numerical contradiction. Therefore, the aim of this paper to investigate the type of contradiction in online review which mainly focusing on hotel online review, to provide useful material on process or methods for identifying contradiction which mainly on the review itself and to determine opportunities for relevant future research for online review contradiction detection. We also proposed a model to detect numerical contradiction in user generated content for tourism industry.
NASA Astrophysics Data System (ADS)
Wibonele, Kasanda J.; Zhang, Yanqing
2002-03-01
A web data mining system using granular computing and ASP programming is proposed. This is a web based application, which allows web users to submit survey data for many different companies. This survey is a collection of questions that will help these companies develop and improve their business and customer service with their clients by analyzing survey data. This web application allows users to submit data anywhere. All the survey data is collected into a database for further analysis. An administrator of this web application can login to the system and view all the data submitted. This web application resides on a web server, and the database resides on the MS SQL server.
An Extraction Method of an Informative DOM Node from a Web Page by Using Layout Information
NASA Astrophysics Data System (ADS)
Tsuruta, Masanobu; Masuyama, Shigeru
We propose an informative DOM node extraction method from a Web page for preprocessing of Web content mining. Our proposed method LM uses layout data of DOM nodes generated by a generic Web browser, and the learning set consists of hundreds of Web pages and the annotations of informative DOM nodes of those Web pages. Our method does not require large scale crawling of the whole Web site to which the target Web page belongs. We design LM so that it uses the information of the learning set more efficiently in comparison to the existing method that uses the same learning set. By experiments, we evaluate the methods obtained by combining one that consists of the method for extracting the informative DOM node both the proposed method and the existing methods, and the existing noise elimination methods: Heur removes advertisements and link-lists by some heuristics and CE removes the DOM nodes existing in the Web pages in the same Web site to which the target Web page belongs. Experimental results show that 1) LM outperforms other methods for extracting the informative DOM node, 2) the combination method (LM, {CE(10), Heur}) based on LM (precision: 0.755, recall: 0.826, F-measure: 0.746) outperforms other combination methods.
The Effectiveness of Web-Based Learning Environment: A Case Study of Public Universities in Kenya
ERIC Educational Resources Information Center
Kirui, Paul A.; Mutai, Sheila J.
2010-01-01
Web mining is emerging in many aspects of e-learning, aiming at improving online learning and teaching processes and making them more transparent and effective. Researchers using Web mining tools and techniques are challenged to learn more about the online students' reshaping online courses and educational websites, and create tools for…
Earth Science Mining Web Services
NASA Astrophysics Data System (ADS)
Pham, L. B.; Lynnes, C. S.; Hegde, M.; Graves, S.; Ramachandran, R.; Maskey, M.; Keiser, K.
2008-12-01
To allow scientists further capabilities in the area of data mining and web services, the Goddard Earth Sciences Data and Information Services Center (GES DISC) and researchers at the University of Alabama in Huntsville (UAH) have developed a system to mine data at the source without the need of network transfers. The system has been constructed by linking together several pre-existing technologies: the Simple Scalable Script-based Science Processor for Measurements (S4PM), a processing engine at the GES DISC; the Algorithm Development and Mining (ADaM) system, a data mining toolkit from UAH that can be configured in a variety of ways to create customized mining processes; ActiveBPEL, a workflow execution engine based on BPEL (Business Process Execution Language); XBaya, a graphical workflow composer; and the EOS Clearinghouse (ECHO). XBaya is used to construct an analysis workflow at UAH using ADaM components, which are also installed remotely at the GES DISC, wrapped as Web Services. The S4PM processing engine searches ECHO for data using space-time criteria, staging them to cache, allowing the ActiveBPEL engine to remotely orchestrates the processing workflow within S4PM. As mining is completed, the output is placed in an FTP holding area for the end user. The goals are to give users control over the data they want to process, while mining data at the data source using the server's resources rather than transferring the full volume over the internet. These diverse technologies have been infused into a functioning, distributed system with only minor changes to the underlying technologies. The key to this infusion is the loosely coupled, Web- Services based architecture: All of the participating components are accessible (one way or another) through (Simple Object Access Protocol) SOAP-based Web Services.
Earth Science Mining Web Services
NASA Technical Reports Server (NTRS)
Pham, Long; Lynnes, Christopher; Hegde, Mahabaleshwa; Graves, Sara; Ramachandran, Rahul; Maskey, Manil; Keiser, Ken
2008-01-01
To allow scientists further capabilities in the area of data mining and web services, the Goddard Earth Sciences Data and Information Services Center (GES DISC) and researchers at the University of Alabama in Huntsville (UAH) have developed a system to mine data at the source without the need of network transfers. The system has been constructed by linking together several pre-existing technologies: the Simple Scalable Script-based Science Processor for Measurements (S4PM), a processing engine at he GES DISC; the Algorithm Development and Mining (ADaM) system, a data mining toolkit from UAH that can be configured in a variety of ways to create customized mining processes; ActiveBPEL, a workflow execution engine based on BPEL (Business Process Execution Language); XBaya, a graphical workflow composer; and the EOS Clearinghouse (ECHO). XBaya is used to construct an analysis workflow at UAH using ADam components, which are also installed remotely at the GES DISC, wrapped as Web Services. The S4PM processing engine searches ECHO for data using space-time criteria, staging them to cache, allowing the ActiveBPEL engine to remotely orchestras the processing workflow within S4PM. As mining is completed, the output is placed in an FTP holding area for the end user. The goals are to give users control over the data they want to process, while mining data at the data source using the server's resources rather than transferring the full volume over the internet. These diverse technologies have been infused into a functioning, distributed system with only minor changes to the underlying technologies. The key to the infusion is the loosely coupled, Web-Services based architecture: All of the participating components are accessible (one way or another) through (Simple Object Access Protocol) SOAP-based Web Services.
Content-Aware DataGuide with Incremental Index Update using Frequently Used Paths
NASA Astrophysics Data System (ADS)
Sharma, A. K.; Duhan, Neelam; Khattar, Priyanka
2010-11-01
Size of the WWW is increasing day by day. Due to the absence of structured data on the Web, it becomes very difficult for information retrieval tools to fully utilize the Web information. As a solution to this problem, XML pages come into play, which provide structural information to the users to some extent. Without efficient indexes, query processing can be quite inefficient due to an exhaustive traversal on XML data. In this paper an improved content-centric approach of Content-Aware DataGuide, which is an indexing technique for XML databases, is being proposed that uses frequently used paths from historical query logs to improve query performance. The index can be updated incrementally according to the changes in query workload and thus, the overhead of reconstruction can be minimized. Frequently used paths are extracted using any Sequential Pattern mining algorithm on subsequent queries in the query workload. After this, the data structures are incrementally updated. This indexing technique proves to be efficient as partial matching queries can be executed efficiently and users can now get the more relevant documents in results.
Mining and integration of pathway diagrams from imaging data.
Kozhenkov, Sergey; Baitaluk, Michael
2012-03-01
Pathway diagrams from PubMed and World Wide Web (WWW) contain valuable highly curated information difficult to reach without tools specifically designed and customized for the biological semantics and high-content density of the images. There is currently no search engine or tool that can analyze pathway images, extract their pathway components (molecules, genes, proteins, organelles, cells, organs, etc.) and indicate their relationships. Here, we describe a resource of pathway diagrams retrieved from article and web-page images through optical character recognition, in conjunction with data mining and data integration methods. The recognized pathways are integrated into the BiologicalNetworks research environment linking them to a wealth of data available in the BiologicalNetworks' knowledgebase, which integrates data from >100 public data sources and the biomedical literature. Multiple search and analytical tools are available that allow the recognized cellular pathways, molecular networks and cell/tissue/organ diagrams to be studied in the context of integrated knowledge, experimental data and the literature. BiologicalNetworks software and the pathway repository are freely available at www.biologicalnetworks.org. Supplementary data are available at Bioinformatics online.
Lightweight monitoring and control system for coal mine safety using REST style.
Cheng, Bo; Cheng, Xin; Chen, Junliang
2015-01-01
The complex environment of a coal mine requires the underground environment, devices and miners to be constantly monitored to ensure safe coal production. However, existing coal mines do not meet these coverage requirements because blind spots occur when using a wired network. In this paper, we develop a Web-based, lightweight remote monitoring and control platform using a wireless sensor network (WSN) with the REST style to collect temperature, humidity and methane concentration data in a coal mine using sensor nodes. This platform also collects information on personnel positions inside the mine. We implement a RESTful application programming interface (API) that provides access to underground sensors and instruments through the Web such that underground coal mine physical devices can be easily interfaced to remote monitoring and control applications. We also implement three different scenarios for Web-based, lightweight remote monitoring and control of coal mine safety and measure and analyze the system performance. Finally, we present the conclusions from this study and discuss future work. Copyright © 2014 ISA. Published by Elsevier Ltd. All rights reserved.
Service-based analysis of biological pathways
Zheng, George; Bouguettaya, Athman
2009-01-01
Background Computer-based pathway discovery is concerned with two important objectives: pathway identification and analysis. Conventional mining and modeling approaches aimed at pathway discovery are often effective at achieving either objective, but not both. Such limitations can be effectively tackled leveraging a Web service-based modeling and mining approach. Results Inspired by molecular recognitions and drug discovery processes, we developed a Web service mining tool, named PathExplorer, to discover potentially interesting biological pathways linking service models of biological processes. The tool uses an innovative approach to identify useful pathways based on graph-based hints and service-based simulation verifying user's hypotheses. Conclusion Web service modeling of biological processes allows the easy access and invocation of these processes on the Web. Web service mining techniques described in this paper enable the discovery of biological pathways linking these process service models. Algorithms presented in this paper for automatically highlighting interesting subgraph within an identified pathway network enable the user to formulate hypothesis, which can be tested out using our simulation algorithm that are also described in this paper. PMID:19796403
Exploiting Recurring Structure in a Semantic Network
NASA Technical Reports Server (NTRS)
Wolfe, Shawn R.; Keller, Richard M.
2004-01-01
With the growing popularity of the Semantic Web, an increasing amount of information is becoming available in machine interpretable, semantically structured networks. Within these semantic networks are recurring structures that could be mined by existing or novel knowledge discovery methods. The mining of these semantic structures represents an interesting area that focuses on mining both for and from the Semantic Web, with surprising applicability to problems confronting the developers of Semantic Web applications. In this paper, we present representative examples of recurring structures and show how these structures could be used to increase the utility of a semantic repository deployed at NASA.
Data Mining Web Services for Science Data Repositories
NASA Astrophysics Data System (ADS)
Graves, S.; Ramachandran, R.; Keiser, K.; Maskey, M.; Lynnes, C.; Pham, L.
2006-12-01
The maturation of web services standards and technologies sets the stage for a distributed "Service-Oriented Architecture" (SOA) for NASA's next generation science data processing. This architecture will allow members of the scientific community to create and combine persistent distributed data processing services and make them available to other users over the Internet. NASA has initiated a project to create a suite of specialized data mining web services designed specifically for science data. The project leverages the Algorithm Development and Mining (ADaM) toolkit as its basis. The ADaM toolkit is a robust, mature and freely available science data mining toolkit that is being used by several research organizations and educational institutions worldwide. These mining services will give the scientific community a powerful and versatile data mining capability that can be used to create higher order products such as thematic maps from current and future NASA satellite data records with methods that are not currently available. The package of mining and related services are being developed using Web Services standards so that community-based measurement processing systems can access and interoperate with them. These standards-based services allow users different options for utilizing them, from direct remote invocation by a client application to deployment of a Business Process Execution Language (BPEL) solutions package where a complex data mining workflow is exposed to others as a single service. The ability to deploy and operate these services at a data archive allows the data mining algorithms to be run where the data are stored, a more efficient scenario than moving large amounts of data over the network. This will be demonstrated in a scenario in which a user uses a remote Web-Service-enabled clustering algorithm to create cloud masks from satellite imagery at the Goddard Earth Sciences Data and Information Services Center (GES DISC).
A user-oriented web crawler for selectively acquiring online content in e-health research.
Xu, Songhua; Yoon, Hong-Jun; Tourassi, Georgia
2014-01-01
Life stories of diseased and healthy individuals are abundantly available on the Internet. Collecting and mining such online content can offer many valuable insights into patients' physical and emotional states throughout the pre-diagnosis, diagnosis, treatment and post-treatment stages of the disease compared with those of healthy subjects. However, such content is widely dispersed across the web. Using traditional query-based search engines to manually collect relevant materials is rather labor intensive and often incomplete due to resource constraints in terms of human query composition and result parsing efforts. The alternative option, blindly crawling the whole web, has proven inefficient and unaffordable for e-health researchers. We propose a user-oriented web crawler that adaptively acquires user-desired content on the Internet to meet the specific online data source acquisition needs of e-health researchers. Experimental results on two cancer-related case studies show that the new crawler can substantially accelerate the acquisition of highly relevant online content compared with the existing state-of-the-art adaptive web crawling technology. For the breast cancer case study using the full training set, the new method achieves a cumulative precision between 74.7 and 79.4% after 5 h of execution till the end of the 20-h long crawling session as compared with the cumulative precision between 32.8 and 37.0% using the peer method for the same time period. For the lung cancer case study using the full training set, the new method achieves a cumulative precision between 56.7 and 61.2% after 5 h of execution till the end of the 20-h long crawling session as compared with the cumulative precision between 29.3 and 32.4% using the peer method. Using the reduced training set in the breast cancer case study, the cumulative precision of our method is between 44.6 and 54.9%, whereas the cumulative precision of the peer method is between 24.3 and 26.3%; for the lung cancer case study using the reduced training set, the cumulative precisions of our method and the peer method are, respectively, between 35.7 and 46.7% versus between 24.1 and 29.6%. These numbers clearly show a consistently superior accuracy of our method in discovering and acquiring user-desired online content for e-health research. The implementation of our user-oriented web crawler is freely available to non-commercial users via the following Web site: http://bsec.ornl.gov/AdaptiveCrawler.shtml. The Web site provides a step-by-step guide on how to execute the web crawler implementation. In addition, the Web site provides the two study datasets including manually labeled ground truth, initial seeds and the crawling results reported in this article.
Design and implementation of website information disclosure assessment system.
Cho, Ying-Chiang; Pan, Jen-Yi
2015-01-01
Internet application technologies, such as cloud computing and cloud storage, have increasingly changed people's lives. Websites contain vast amounts of personal privacy information. In order to protect this information, network security technologies, such as database protection and data encryption, attract many researchers. The most serious problems concerning web vulnerability are e-mail address and network database leakages. These leakages have many causes. For example, malicious users can steal database contents, taking advantage of mistakes made by programmers and administrators. In order to mitigate this type of abuse, a website information disclosure assessment system is proposed in this study. This system utilizes a series of technologies, such as web crawler algorithms, SQL injection attack detection, and web vulnerability mining, to assess a website's information disclosure. Thirty websites, randomly sampled from the top 50 world colleges, were used to collect leakage information. This testing showed the importance of increasing the security and privacy of website information for academic websites.
NASA Astrophysics Data System (ADS)
Zhang, Xiaowen; Chen, Bingfeng
2017-08-01
Based on the frequent sub-tree mining algorithm, this paper proposes a construction scheme of web page comment information extraction system based on frequent subtree mining, referred to as FSM system. The entire system architecture and the various modules to do a brief introduction, and then the core of the system to do a detailed description, and finally give the system prototype.
Kernel Methods for Mining Instance Data in Ontologies
NASA Astrophysics Data System (ADS)
Bloehdorn, Stephan; Sure, York
The amount of ontologies and meta data available on the Web is constantly growing. The successful application of machine learning techniques for learning of ontologies from textual data, i.e. mining for the Semantic Web, contributes to this trend. However, no principal approaches exist so far for mining from the Semantic Web. We investigate how machine learning algorithms can be made amenable for directly taking advantage of the rich knowledge expressed in ontologies and associated instance data. Kernel methods have been successfully employed in various learning tasks and provide a clean framework for interfacing between non-vectorial data and machine learning algorithms. In this spirit, we express the problem of mining instances in ontologies as the problem of defining valid corresponding kernels. We present a principled framework for designing such kernels by means of decomposing the kernel computation into specialized kernels for selected characteristics of an ontology which can be flexibly assembled and tuned. Initial experiments on real world Semantic Web data enjoy promising results and show the usefulness of our approach.
Abandoned Uranium Mines (AUM) Site Screening Map Service, 2016, US EPA Region 9
As described in detail in the Five-Year Report, US EPA completed on-the-ground screening of 521 abandoned uranium mine areas. US EPA and the Navajo EPA are using the Comprehensive Database and Atlas to determine which mines should be cleaned up first. US EPA continues to research and identify Potentially Responsible Parties (PRPs) under Superfund to contribute to the costs of cleanup efforts.This US EPA Region 9 web service contains the following map layers:Abandoned Uranium Mines, Priority Mines, Tronox Mines, Navajo Environmental Response Trust Mines, Mines with Enforcement Actions, Superfund AUM Regions, Navajo Nation Administrative Boundaries and Chapter Houses.Mine points have a maximum scale of 1:220,000, while Mine polygons have a minimum scale of 1:220,000. Chapter houses have a minimum scale of 1:200,000. BLM Land Status has a minimum scale of 1:150,000.Full FGDC metadata records for each layer can be found by clicking the layer name at the web service endpoint and viewing the layer description. Data used to create this web service are available for download at https://edg.epa.gov/metadata/catalog/data/data.page.Security Classification: Public. Access Constraints: None. Use Constraints: None. Please check sources, scale, accuracy, currentness and other available information. Please confirm that you are using the most recent copy of both data and metadata. Acknowledgement of the EPA would be appreciated.
Supporting the education evidence portal via text mining
Ananiadou, Sophia; Thompson, Paul; Thomas, James; Mu, Tingting; Oliver, Sandy; Rickinson, Mark; Sasaki, Yutaka; Weissenbacher, Davy; McNaught, John
2010-01-01
The UK Education Evidence Portal (eep) provides a single, searchable, point of access to the contents of the websites of 33 organizations relating to education, with the aim of revolutionizing work practices for the education community. Use of the portal alleviates the need to spend time searching multiple resources to find relevant information. However, the combined content of the websites of interest is still very large (over 500 000 documents and growing). This means that searches using the portal can produce very large numbers of hits. As users often have limited time, they would benefit from enhanced methods of performing searches and viewing results, allowing them to drill down to information of interest more efficiently, without having to sift through potentially long lists of irrelevant documents. The Joint Information Systems Committee (JISC)-funded ASSIST project has produced a prototype web interface to demonstrate the applicability of integrating a number of text-mining tools and methods into the eep, to facilitate an enhanced searching, browsing and document-viewing experience. New features include automatic classification of documents according to a taxonomy, automatic clustering of search results according to similar document content, and automatic identification and highlighting of key terms within documents. PMID:20643679
NASA Astrophysics Data System (ADS)
Huang, Yin; Chen, Jianhua; Xiong, Shaojun
2009-07-01
Mobile-Learning (M-learning) makes many learners get the advantages of both traditional learning and E-learning. Currently, Web-based Mobile-Learning Systems have created many new ways and defined new relationships between educators and learners. Association rule mining is one of the most important fields in data mining and knowledge discovery in databases. Rules explosion is a serious problem which causes great concerns, as conventional mining algorithms often produce too many rules for decision makers to digest. Since Web-based Mobile-Learning System collects vast amounts of student profile data, data mining and knowledge discovery techniques can be applied to find interesting relationships between attributes of learners, assessments, the solution strategies adopted by learners and so on. Therefore ,this paper focus on a new data-mining algorithm, combined with the advantages of genetic algorithm and simulated annealing algorithm , called ARGSA(Association rules based on an improved Genetic Simulated Annealing Algorithm), to mine the association rules. This paper first takes advantage of the Parallel Genetic Algorithm and Simulated Algorithm designed specifically for discovering association rules. Moreover, the analysis and experiment are also made to show the proposed method is superior to the Apriori algorithm in this Mobile-Learning system.
Graph Mining Meets the Semantic Web
DOE Office of Scientific and Technical Information (OSTI.GOV)
Lee, Sangkeun; Sukumar, Sreenivas R; Lim, Seung-Hwan
The Resource Description Framework (RDF) and SPARQL Protocol and RDF Query Language (SPARQL) were introduced about a decade ago to enable flexible schema-free data interchange on the Semantic Web. Today, data scientists use the framework as a scalable graph representation for integrating, querying, exploring and analyzing data sets hosted at different sources. With increasing adoption, the need for graph mining capabilities for the Semantic Web has emerged. We address that need through implementation of three popular iterative Graph Mining algorithms (Triangle count, Connected component analysis, and PageRank). We implement these algorithms as SPARQL queries, wrapped within Python scripts. We evaluatemore » the performance of our implementation on 6 real world data sets and show graph mining algorithms (that have a linear-algebra formulation) can indeed be unleashed on data represented as RDF graphs using the SPARQL query interface.« less
Mining Longitudinal Web Queries: Trends and Patterns.
ERIC Educational Resources Information Center
Wang, Peiling; Berry, Michael W.; Yang, Yiheng
2003-01-01
Analyzed user queries submitted to an academic Web site during a four-year period, using a relational database, to examine users' query behavior, to identify problems they encounter, and to develop techniques for optimizing query analysis and mining. Linguistic analyses focus on query structures, lexicon, and word associations using statistical…
An Expertise Recommender using Web Mining
NASA Technical Reports Server (NTRS)
Joshi, Anupam; Chandrasekaran, Purnima; ShuYang, Michelle; Ramakrishnan, Ramya
2001-01-01
This report explored techniques to mine web pages of scientists to extract information regarding their expertise, build expertise chains and referral webs, and semi automatically combine this information with directory information services to create a recommender system that permits query by expertise. The approach included experimenting with existing techniques that have been reported in research literature in recent past , and adapted them as needed. In addition, software tools were developed to capture and use this information.
Visualization of usability and functionality of a professional website through web-mining.
Jones, Josette F; Mahoui, Malika; Gopa, Venkata Devi Pragna
2007-10-11
Functional interface design requires understanding of the information system structure and the user. Web logs record user interactions with the interface, and thus provide some insight into user search behavior and efficiency of the search process. The present study uses a data-mining approach with techniques such as association rules, clustering and classification, to visualize the usability and functionality of a digital library through in depth analyses of web logs.
Application of data mining in science and technology management information system based on WebGIS
NASA Astrophysics Data System (ADS)
Wu, Xiaofang; Xu, Zhiyong; Bao, Shitai; Chen, Feixiang
2009-10-01
With the rapid development of science and technology and the quick increase of information, a great deal of data is accumulated in the management department of science and technology. Usually, many knowledge and rules are contained and concealed in the data. Therefore, how to excavate and use the knowledge fully is very important in the management of science and technology. It will help to examine and approve the project of science and technology more scientifically and make the achievement transformed as the realistic productive forces easier. Therefore, the data mine technology will be researched and applied to the science and technology management information system to find and excavate the knowledge in the paper. According to analyzing the disadvantages of traditional science and technology management information system, the database technology, data mining and web geographic information systems (WebGIS) technology will be introduced to develop and construct the science and technology management information system based on WebGIS. The key problems are researched in detail such as data mining and statistical analysis. What's more, the prototype system is developed and validated based on the project data of National Natural Science Foundation Committee. The spatial data mining is done from the axis of time, space and other factors. Then the variety of knowledge and rules will be excavated by using data mining technology, which helps to provide an effective support for decisionmaking.
Binary Coded Web Access Pattern Tree in Education Domain
ERIC Educational Resources Information Center
Gomathi, C.; Moorthi, M.; Duraiswamy, K.
2008-01-01
Web Access Pattern (WAP), which is the sequence of accesses pursued by users frequently, is a kind of interesting and useful knowledge in practice. Sequential Pattern mining is the process of applying data mining techniques to a sequential database for the purposes of discovering the correlation relationships that exist among an ordered list of…
A user-oriented web crawler for selectively acquiring online content in e-health research
Xu, Songhua; Yoon, Hong-Jun; Tourassi, Georgia
2014-01-01
Motivation: Life stories of diseased and healthy individuals are abundantly available on the Internet. Collecting and mining such online content can offer many valuable insights into patients’ physical and emotional states throughout the pre-diagnosis, diagnosis, treatment and post-treatment stages of the disease compared with those of healthy subjects. However, such content is widely dispersed across the web. Using traditional query-based search engines to manually collect relevant materials is rather labor intensive and often incomplete due to resource constraints in terms of human query composition and result parsing efforts. The alternative option, blindly crawling the whole web, has proven inefficient and unaffordable for e-health researchers. Results: We propose a user-oriented web crawler that adaptively acquires user-desired content on the Internet to meet the specific online data source acquisition needs of e-health researchers. Experimental results on two cancer-related case studies show that the new crawler can substantially accelerate the acquisition of highly relevant online content compared with the existing state-of-the-art adaptive web crawling technology. For the breast cancer case study using the full training set, the new method achieves a cumulative precision between 74.7 and 79.4% after 5 h of execution till the end of the 20-h long crawling session as compared with the cumulative precision between 32.8 and 37.0% using the peer method for the same time period. For the lung cancer case study using the full training set, the new method achieves a cumulative precision between 56.7 and 61.2% after 5 h of execution till the end of the 20-h long crawling session as compared with the cumulative precision between 29.3 and 32.4% using the peer method. Using the reduced training set in the breast cancer case study, the cumulative precision of our method is between 44.6 and 54.9%, whereas the cumulative precision of the peer method is between 24.3 and 26.3%; for the lung cancer case study using the reduced training set, the cumulative precisions of our method and the peer method are, respectively, between 35.7 and 46.7% versus between 24.1 and 29.6%. These numbers clearly show a consistently superior accuracy of our method in discovering and acquiring user-desired online content for e-health research. Availability and implementation: The implementation of our user-oriented web crawler is freely available to non-commercial users via the following Web site: http://bsec.ornl.gov/AdaptiveCrawler.shtml. The Web site provides a step-by-step guide on how to execute the web crawler implementation. In addition, the Web site provides the two study datasets including manually labeled ground truth, initial seeds and the crawling results reported in this article. Contact: xus1@ornl.gov Supplementary information: Supplementary data are available at Bioinformatics online. PMID:24078710
Methane Content Estimation in DuongHuy Coal Mine
NASA Astrophysics Data System (ADS)
Nguyen, Van Thinh; Mijał, Waldemar; Dang, Vu Chi; Nguyen, Thi Tuyet Mai
2018-03-01
Methane hazard has always been considered for underground coal mining as it can lead to methane explosion. In Quang Ninh province, several coal mines such as Mạo Khe coal mine, Khe Cham coal mine, especially Duong Huy mine that have high methane content. Experimental data to examine contents of methane bearing coal seams at different depths are not similar in Duong coal mine. In order to ensure safety, this report has been undertaken to determine a pattern of changing methane contents of coal seams at different exploitation depths in Duong Huy underground coal mine.
Text and Structural Data Mining of Influenza Mentions in Web and Social Media
DOE Office of Scientific and Technical Information (OSTI.GOV)
Corley, Courtney D.; Cook, Diane; Mikler, Armin R.
Text and structural data mining of Web and social media (WSM) provides a novel disease surveillance resource and can identify online communities for targeted public health communications (PHC) to assure wide dissemination of pertinent information. WSM that mention influenza are harvested over a 24-week period, 5-October-2008 to 21-March-2009. Link analysis reveals communities for targeted PHC. Text mining is shown to identify trends in flu posts that correlate to real-world influenza-like-illness patient report data. We also bring to bear a graph-based data mining technique to detect anomalies among flu blogs connected by publisher type, links, and user-tags.
NASA Astrophysics Data System (ADS)
Abdillah, T.; Dai, R.; Setiawan, E.
2018-02-01
This study aims to develop the application of Web Services technology with RestFul Protocol to optimize the information presentation on mining potential. This study used User Interface Design approach for the information accuracy and relevance as well as the Web Service for the reliability in presenting the information. The results show that: the information accuracy and relevance regarding mining potential can be seen from the achievement of User Interface implementation in the application that is based on the following rules: The consideration of the appropriate colours and objects, the easiness of using the navigation, and users’ interaction with the applications that employs symbols and languages understood by the users; the information accuracy and relevance related to mining potential can be observed by the information presented by using charts and Tool Tip Text to help the users understand the provided chart/figure; the reliability of the information presentation is evident by the results of Web Services testing in Figure 4.5.6. This study finds out that User Interface Design and Web Services approaches (for the access of different Platform apps) are able to optimize the presentation. The results of this study can be used as a reference for software developers and Provincial Government of Gorontalo.
Document Exploration and Automatic Knowledge Extraction for Unstructured Biomedical Text
NASA Astrophysics Data System (ADS)
Chu, S.; Totaro, G.; Doshi, N.; Thapar, S.; Mattmann, C. A.; Ramirez, P.
2015-12-01
We describe our work on building a web-browser based document reader with built-in exploration tool and automatic concept extraction of medical entities for biomedical text. Vast amounts of biomedical information are offered in unstructured text form through scientific publications and R&D reports. Utilizing text mining can help us to mine information and extract relevant knowledge from a plethora of biomedical text. The ability to employ such technologies to aid researchers in coping with information overload is greatly desirable. In recent years, there has been an increased interest in automatic biomedical concept extraction [1, 2] and intelligent PDF reader tools with the ability to search on content and find related articles [3]. Such reader tools are typically desktop applications and are limited to specific platforms. Our goal is to provide researchers with a simple tool to aid them in finding, reading, and exploring documents. Thus, we propose a web-based document explorer, which we called Shangri-Docs, which combines a document reader with automatic concept extraction and highlighting of relevant terms. Shangri-Docsalso provides the ability to evaluate a wide variety of document formats (e.g. PDF, Words, PPT, text, etc.) and to exploit the linked nature of the Web and personal content by performing searches on content from public sites (e.g. Wikipedia, PubMed) and private cataloged databases simultaneously. Shangri-Docsutilizes Apache cTAKES (clinical Text Analysis and Knowledge Extraction System) [4] and Unified Medical Language System (UMLS) to automatically identify and highlight terms and concepts, such as specific symptoms, diseases, drugs, and anatomical sites, mentioned in the text. cTAKES was originally designed specially to extract information from clinical medical records. Our investigation leads us to extend the automatic knowledge extraction process of cTAKES for biomedical research domain by improving the ontology guided information extraction process. We will describe our experience and implementation of our system and share lessons learned from our development. We will also discuss ways in which this could be adapted to other science fields. [1] Funk et al., 2014. [2] Kang et al., 2014. [3] Utopia Documents, http://utopiadocs.com [4] Apache cTAKES, http://ctakes.apache.org
Design and Implementation of Website Information Disclosure Assessment System
Cho, Ying-Chiang; Pan, Jen-Yi
2015-01-01
Internet application technologies, such as cloud computing and cloud storage, have increasingly changed people’s lives. Websites contain vast amounts of personal privacy information. In order to protect this information, network security technologies, such as database protection and data encryption, attract many researchers. The most serious problems concerning web vulnerability are e-mail address and network database leakages. These leakages have many causes. For example, malicious users can steal database contents, taking advantage of mistakes made by programmers and administrators. In order to mitigate this type of abuse, a website information disclosure assessment system is proposed in this study. This system utilizes a series of technologies, such as web crawler algorithms, SQL injection attack detection, and web vulnerability mining, to assess a website’s information disclosure. Thirty websites, randomly sampled from the top 50 world colleges, were used to collect leakage information. This testing showed the importance of increasing the security and privacy of website information for academic websites. PMID:25768434
Motivation Mining: Prospecting the Web.
ERIC Educational Resources Information Center
Small, Ruth V.; Arnone, Marilyn P.
1999-01-01
Describes WebMAC instruments, which differ from other Web-evaluation instruments because they have a theoretical base, are user-centered, are designed for students in grades 7 through 12, and assess the motivational quality of Web sites. Examples are given of uses of WebMAC Middle and WebMAC Senior in activities to promote evaluation and…
Wang, Weiqi; Wang, Yanbo Justin; Bañares-Alcántara, René; Coenen, Frans; Cui, Zhanfeng
2009-12-01
In this paper, data mining is used to analyze the data on the differentiation of mammalian Mesenchymal Stem Cells (MSCs), aiming at discovering known and hidden rules governing MSC differentiation, following the establishment of a web-based public database containing experimental data on the MSC proliferation and differentiation. To this effect, a web-based public interactive database comprising the key parameters which influence the fate and destiny of mammalian MSCs has been constructed and analyzed using Classification Association Rule Mining (CARM) as a data-mining technique. The results show that the proposed approach is technically feasible and performs well with respect to the accuracy of (classification) prediction. Key rules mined from the constructed MSC database are consistent with experimental observations, indicating the validity of the method developed and the first step in the application of data mining to the study of MSCs.
NASA Astrophysics Data System (ADS)
Ma, Kevin; Wang, Ximing; Lerner, Alex; Shiroishi, Mark; Amezcua, Lilyana; Liu, Brent
2015-03-01
In the past, we have developed and displayed a multiple sclerosis eFolder system for patient data storage, image viewing, and automatic lesion quantification results stored in DICOM-SR format. The web-based system aims to be integrated in DICOM-compliant clinical and research environments to aid clinicians in patient treatments and disease tracking. This year, we have further developed the eFolder system to handle big data analysis and data mining in today's medical imaging field. The database has been updated to allow data mining and data look-up from DICOM-SR lesion analysis contents. Longitudinal studies are tracked, and any changes in lesion volumes and brain parenchyma volumes are calculated and shown on the webbased user interface as graphical representations. Longitudinal lesion characteristic changes are compared with patients' disease history, including treatments, symptom progressions, and any other changes in the disease profile. The image viewer is updated such that imaging studies can be viewed side-by-side to allow visual comparisons. We aim to use the web-based medical imaging informatics eFolder system to demonstrate big data analysis in medical imaging, and use the analysis results to predict MS disease trends and patterns in Hispanic and Caucasian populations in our pilot study. The discovery of disease patterns among the two ethnicities is a big data analysis result that will help lead to personalized patient care and treatment planning.
Mining Student Data Captured from a Web-Based Tutoring Tool: Initial Exploration and Results
ERIC Educational Resources Information Center
Merceron, Agathe; Yacef, Kalina
2004-01-01
In this article we describe the initial investigations that we have conducted on student data collected from a web-based tutoring tool. We have used some data mining techniques such as association rule and symbolic data analysis, as well as traditional SQL queries to gain further insight on the students' learning and deduce information to improve…
Beyond accuracy: creating interoperable and scalable text-mining web services.
Wei, Chih-Hsuan; Leaman, Robert; Lu, Zhiyong
2016-06-15
The biomedical literature is a knowledge-rich resource and an important foundation for future research. With over 24 million articles in PubMed and an increasing growth rate, research in automated text processing is becoming increasingly important. We report here our recently developed web-based text mining services for biomedical concept recognition and normalization. Unlike most text-mining software tools, our web services integrate several state-of-the-art entity tagging systems (DNorm, GNormPlus, SR4GN, tmChem and tmVar) and offer a batch-processing mode able to process arbitrary text input (e.g. scholarly publications, patents and medical records) in multiple formats (e.g. BioC). We support multiple standards to make our service interoperable and allow simpler integration with other text-processing pipelines. To maximize scalability, we have preprocessed all PubMed articles, and use a computer cluster for processing large requests of arbitrary text. Our text-mining web service is freely available at http://www.ncbi.nlm.nih.gov/CBBresearch/Lu/Demo/tmTools/#curl : Zhiyong.Lu@nih.gov. Published by Oxford University Press 2016. This work is written by US Government employees and is in the public domain in the US.
Stratification-Based Outlier Detection over the Deep Web.
Xian, Xuefeng; Zhao, Pengpeng; Sheng, Victor S; Fang, Ligang; Gu, Caidong; Yang, Yuanfeng; Cui, Zhiming
2016-01-01
For many applications, finding rare instances or outliers can be more interesting than finding common patterns. Existing work in outlier detection never considers the context of deep web. In this paper, we argue that, for many scenarios, it is more meaningful to detect outliers over deep web. In the context of deep web, users must submit queries through a query interface to retrieve corresponding data. Therefore, traditional data mining methods cannot be directly applied. The primary contribution of this paper is to develop a new data mining method for outlier detection over deep web. In our approach, the query space of a deep web data source is stratified based on a pilot sample. Neighborhood sampling and uncertainty sampling are developed in this paper with the goal of improving recall and precision based on stratification. Finally, a careful performance evaluation of our algorithm confirms that our approach can effectively detect outliers in deep web.
Stratification-Based Outlier Detection over the Deep Web
Xian, Xuefeng; Zhao, Pengpeng; Sheng, Victor S.; Fang, Ligang; Gu, Caidong; Yang, Yuanfeng; Cui, Zhiming
2016-01-01
For many applications, finding rare instances or outliers can be more interesting than finding common patterns. Existing work in outlier detection never considers the context of deep web. In this paper, we argue that, for many scenarios, it is more meaningful to detect outliers over deep web. In the context of deep web, users must submit queries through a query interface to retrieve corresponding data. Therefore, traditional data mining methods cannot be directly applied. The primary contribution of this paper is to develop a new data mining method for outlier detection over deep web. In our approach, the query space of a deep web data source is stratified based on a pilot sample. Neighborhood sampling and uncertainty sampling are developed in this paper with the goal of improving recall and precision based on stratification. Finally, a careful performance evaluation of our algorithm confirms that our approach can effectively detect outliers in deep web. PMID:27313603
Yoon, Hong-Jun; Xu, Songhua; Han, Xuesong
2016-01-01
Background The World Wide Web has emerged as a powerful data source for epidemiological studies related to infectious disease surveillance. However, its potential for cancer-related epidemiological discoveries is largely unexplored. Methods Using advanced web crawling and tailored information extraction procedures, the authors automatically collected and analyzed the text content of 79 394 online obituary articles published between 1998 and 2014. The collected data included 51 911 cancer (27 330 breast; 9470 lung; 6496 pancreatic; 6342 ovarian; 2273 colon) and 27 483 non-cancer cases. With the derived information, the authors replicated a case-control study design to investigate the association between parity (i.e., childbearing) and cancer risk. Age-adjusted odds ratios (ORs) with 95% confidence intervals (CIs) were calculated for each cancer type and compared to those reported in large-scale epidemiological studies. Results Parity was found to be associated with a significantly reduced risk of breast cancer (OR = 0.78, 95% CI, 0.75-0.82), pancreatic cancer (OR = 0.78, 95% CI, 0.72-0.83), colon cancer (OR = 0.67, 95% CI, 0.60-0.74), and ovarian cancer (OR = 0.58, 95% CI, 0.54-0.62). Marginal association was found for lung cancer risk (OR = 0.87, 95% CI, 0.81-0.92). The linear trend between increased parity and reduced cancer risk was dramatically more pronounced for breast and ovarian cancer than the other cancers included in the analysis. Conclusion This large web-mining study on parity and cancer risk produced findings very similar to those reported with traditional observational studies. It may be used as a promising strategy to generate study hypotheses for guiding and prioritizing future epidemiological studies. PMID:26615183
Deploying and sharing U-Compare workflows as web services.
Kontonatsios, Georgios; Korkontzelos, Ioannis; Kolluru, Balakrishna; Thompson, Paul; Ananiadou, Sophia
2013-02-18
U-Compare is a text mining platform that allows the construction, evaluation and comparison of text mining workflows. U-Compare contains a large library of components that are tuned to the biomedical domain. Users can rapidly develop biomedical text mining workflows by mixing and matching U-Compare's components. Workflows developed using U-Compare can be exported and sent to other users who, in turn, can import and re-use them. However, the resulting workflows are standalone applications, i.e., software tools that run and are accessible only via a local machine, and that can only be run with the U-Compare platform. We address the above issues by extending U-Compare to convert standalone workflows into web services automatically, via a two-click process. The resulting web services can be registered on a central server and made publicly available. Alternatively, users can make web services available on their own servers, after installing the web application framework, which is part of the extension to U-Compare. We have performed a user-oriented evaluation of the proposed extension, by asking users who have tested the enhanced functionality of U-Compare to complete questionnaires that assess its functionality, reliability, usability, efficiency and maintainability. The results obtained reveal that the new functionality is well received by users. The web services produced by U-Compare are built on top of open standards, i.e., REST and SOAP protocols, and therefore, they are decoupled from the underlying platform. Exported workflows can be integrated with any application that supports these open standards. We demonstrate how the newly extended U-Compare enhances the cross-platform interoperability of workflows, by seamlessly importing a number of text mining workflow web services exported from U-Compare into Taverna, i.e., a generic scientific workflow construction platform.
Deploying and sharing U-Compare workflows as web services
2013-01-01
Background U-Compare is a text mining platform that allows the construction, evaluation and comparison of text mining workflows. U-Compare contains a large library of components that are tuned to the biomedical domain. Users can rapidly develop biomedical text mining workflows by mixing and matching U-Compare’s components. Workflows developed using U-Compare can be exported and sent to other users who, in turn, can import and re-use them. However, the resulting workflows are standalone applications, i.e., software tools that run and are accessible only via a local machine, and that can only be run with the U-Compare platform. Results We address the above issues by extending U-Compare to convert standalone workflows into web services automatically, via a two-click process. The resulting web services can be registered on a central server and made publicly available. Alternatively, users can make web services available on their own servers, after installing the web application framework, which is part of the extension to U-Compare. We have performed a user-oriented evaluation of the proposed extension, by asking users who have tested the enhanced functionality of U-Compare to complete questionnaires that assess its functionality, reliability, usability, efficiency and maintainability. The results obtained reveal that the new functionality is well received by users. Conclusions The web services produced by U-Compare are built on top of open standards, i.e., REST and SOAP protocols, and therefore, they are decoupled from the underlying platform. Exported workflows can be integrated with any application that supports these open standards. We demonstrate how the newly extended U-Compare enhances the cross-platform interoperability of workflows, by seamlessly importing a number of text mining workflow web services exported from U-Compare into Taverna, i.e., a generic scientific workflow construction platform. PMID:23419017
PREDOSE: A Semantic Web Platform for Drug Abuse Epidemiology using Social Media
Cameron, Delroy; Smith, Gary A.; Daniulaityte, Raminta; Sheth, Amit P.; Dave, Drashti; Chen, Lu; Anand, Gaurish; Carlson, Robert; Watkins, Kera Z.; Falck, Russel
2013-01-01
Objectives The role of social media in biomedical knowledge mining, including clinical, medical and healthcare informatics, prescription drug abuse epidemiology and drug pharmacology, has become increasingly significant in recent years. Social media offers opportunities for people to share opinions and experiences freely in online communities, which may contribute information beyond the knowledge of domain professionals. This paper describes the development of a novel Semantic Web platform called PREDOSE (PREscription Drug abuse Online Surveillance and Epidemiology), which is designed to facilitate the epidemiologic study of prescription (and related) drug abuse practices using social media. PREDOSE uses web forum posts and domain knowledge, modeled in a manually created Drug Abuse Ontology (DAO) (pronounced dow), to facilitate the extraction of semantic information from User Generated Content (UGC). A combination of lexical, pattern-based and semantics-based techniques is used together with the domain knowledge to extract fine-grained semantic information from UGC. In a previous study, PREDOSE was used to obtain the datasets from which new knowledge in drug abuse research was derived. Here, we report on various platform enhancements, including an updated DAO, new components for relationship and triple extraction, and tools for content analysis, trend detection and emerging patterns exploration, which enhance the capabilities of the PREDOSE platform. Given these enhancements, PREDOSE is now more equipped to impact drug abuse research by alleviating traditional labor-intensive content analysis tasks. Methods Using custom web crawlers that scrape UGC from publicly available web forums, PREDOSE first automates the collection of web-based social media content for subsequent semantic annotation. The annotation scheme is modeled in the DAO, and includes domain specific knowledge such as prescription (and related) drugs, methods of preparation, side effects, routes of administration, etc. The DAO is also used to help recognize three types of data, namely: 1) entities, 2) relationships and 3) triples. PREDOSE then uses a combination of lexical and semantic-based techniques to extract entities and relationships from the scraped content, and a top-down approach for triple extraction that uses patterns expressed in the DAO. In addition, PREDOSE uses publicly available lexicons to identify initial sentiment expressions in text, and then a probabilistic optimization algorithm (from related research) to extract the final sentiment expressions. Together, these techniques enable the capture of fine-grained semantic information from UGC, and querying, search, trend analysis and overall content analysis of social media related to prescription drug abuse. Moreover, extracted data are also made available to domain experts for the creation of training and test sets for use in evaluation and refinements in information extraction techniques. Results A recent evaluation of the information extraction techniques applied in the PREDOSE platform indicates 85% precision and 72% recall in entity identification, on a manually created gold standard dataset. In another study, PREDOSE achieved 36% precision in relationship identification and 33% precision in triple extraction, through manual evaluation by domain experts. Given the complexity of the relationship and triple extraction tasks and the abstruse nature of social media texts, we interpret these as favorable initial results. Extracted semantic information is currently in use in an online discovery support system, by prescription drug abuse researchers at the Center for Interventions, Treatment and Addictions Research (CITAR) at Wright State University. Conclusion A comprehensive platform for entity, relationship, triple and sentiment extraction from such abstruse texts has never been developed for drug abuse research. PREDOSE has already demonstrated the importance of mining social media by providing data from which new findings in drug abuse research were uncovered. Given the recent platform enhancements, including the refined DAO, components for relationship and triple extraction, and tools for content, trend and emerging pattern analysis, it is expected that PREDOSE will play a significant role in advancing drug abuse epidemiology in future. PMID:23892295
PREDOSE: a semantic web platform for drug abuse epidemiology using social media.
Cameron, Delroy; Smith, Gary A; Daniulaityte, Raminta; Sheth, Amit P; Dave, Drashti; Chen, Lu; Anand, Gaurish; Carlson, Robert; Watkins, Kera Z; Falck, Russel
2013-12-01
The role of social media in biomedical knowledge mining, including clinical, medical and healthcare informatics, prescription drug abuse epidemiology and drug pharmacology, has become increasingly significant in recent years. Social media offers opportunities for people to share opinions and experiences freely in online communities, which may contribute information beyond the knowledge of domain professionals. This paper describes the development of a novel semantic web platform called PREDOSE (PREscription Drug abuse Online Surveillance and Epidemiology), which is designed to facilitate the epidemiologic study of prescription (and related) drug abuse practices using social media. PREDOSE uses web forum posts and domain knowledge, modeled in a manually created Drug Abuse Ontology (DAO--pronounced dow), to facilitate the extraction of semantic information from User Generated Content (UGC), through combination of lexical, pattern-based and semantics-based techniques. In a previous study, PREDOSE was used to obtain the datasets from which new knowledge in drug abuse research was derived. Here, we report on various platform enhancements, including an updated DAO, new components for relationship and triple extraction, and tools for content analysis, trend detection and emerging patterns exploration, which enhance the capabilities of the PREDOSE platform. Given these enhancements, PREDOSE is now more equipped to impact drug abuse research by alleviating traditional labor-intensive content analysis tasks. Using custom web crawlers that scrape UGC from publicly available web forums, PREDOSE first automates the collection of web-based social media content for subsequent semantic annotation. The annotation scheme is modeled in the DAO, and includes domain specific knowledge such as prescription (and related) drugs, methods of preparation, side effects, and routes of administration. The DAO is also used to help recognize three types of data, namely: (1) entities, (2) relationships and (3) triples. PREDOSE then uses a combination of lexical and semantic-based techniques to extract entities and relationships from the scraped content, and a top-down approach for triple extraction that uses patterns expressed in the DAO. In addition, PREDOSE uses publicly available lexicons to identify initial sentiment expressions in text, and then a probabilistic optimization algorithm (from related research) to extract the final sentiment expressions. Together, these techniques enable the capture of fine-grained semantic information, which facilitate search, trend analysis and overall content analysis using social media on prescription drug abuse. Moreover, extracted data are also made available to domain experts for the creation of training and test sets for use in evaluation and refinements in information extraction techniques. A recent evaluation of the information extraction techniques applied in the PREDOSE platform indicates 85% precision and 72% recall in entity identification, on a manually created gold standard dataset. In another study, PREDOSE achieved 36% precision in relationship identification and 33% precision in triple extraction, through manual evaluation by domain experts. Given the complexity of the relationship and triple extraction tasks and the abstruse nature of social media texts, we interpret these as favorable initial results. Extracted semantic information is currently in use in an online discovery support system, by prescription drug abuse researchers at the Center for Interventions, Treatment and Addictions Research (CITAR) at Wright State University. A comprehensive platform for entity, relationship, triple and sentiment extraction from such abstruse texts has never been developed for drug abuse research. PREDOSE has already demonstrated the importance of mining social media by providing data from which new findings in drug abuse research were uncovered. Given the recent platform enhancements, including the refined DAO, components for relationship and triple extraction, and tools for content, trend and emerging pattern analysis, it is expected that PREDOSE will play a significant role in advancing drug abuse epidemiology in future. Copyright © 2013 Elsevier Inc. All rights reserved.
Masseroli, Marco; Stella, Andrea; Meani, Natalia; Alcalay, Myriam; Pinciroli, Francesco
2004-12-12
High-throughput technologies create the necessity to mine large amounts of gene annotations from diverse databanks, and to integrate the resulting data. Most databanks can be interrogated only via Web, for a single gene at a time, and query results are generally available only in the HTML format. Although some databanks provide batch retrieval of data via FTP, this requires expertise and resources for locally reimplementing the databank. We developed MyWEST, a tool aimed at researchers without extensive informatics skills or resources, which exploits user-defined templates to easily mine selected annotations from different Web-interfaced databanks, and aggregates and structures results in an automatically updated database. Using microarray results from a model system of retinoic acid-induced differentiation, MyWEST effectively gathered relevant annotations from various biomolecular databanks, highlighted significant biological characteristics and supported a global approach to the understanding of complex cellular mechanisms. MyWEST is freely available for non-profit use at http://www.medinfopoli.polimi.it/MyWEST/
Federal Register 2010, 2011, 2012, 2013, 2014
2011-09-29
...--County on January 28, 2011. The public notice is available on Charleston District's public Web site at... eight open mining pits over a twelve-year period, with pit depths ranging from 110 to 840 feet deep. The... of January 28, 2011, and are available on Charleston District's public Web site at http://www.sac...
Web usage mining at an academic health sciences library: an exploratory study.
Bracke, Paul J
2004-10-01
This paper explores the potential of multinomial logistic regression analysis to perform Web usage mining for an academic health sciences library Website. Usage of database-driven resource gateway pages was logged for a six-month period, including information about users' network addresses, referring uniform resource locators (URLs), and types of resource accessed. It was found that referring URL did vary significantly by two factors: whether a user was on-campus and what type of resource was accessed. Although the data available for analysis are limited by the nature of the Web and concerns for privacy, this method demonstrates the potential for gaining insight into Web usage that supplements Web log analysis. It can be used to improve the design of static and dynamic Websites today and could be used in the design of more advanced Web systems in the future.
Effective Filtering of Query Results on Updated User Behavioral Profiles in Web Mining
Sadesh, S.; Suganthe, R. C.
2015-01-01
Web with tremendous volume of information retrieves result for user related queries. With the rapid growth of web page recommendation, results retrieved based on data mining techniques did not offer higher performance filtering rate because relationships between user profile and queries were not analyzed in an extensive manner. At the same time, existing user profile based prediction in web data mining is not exhaustive in producing personalized result rate. To improve the query result rate on dynamics of user behavior over time, Hamilton Filtered Regime Switching User Query Probability (HFRS-UQP) framework is proposed. HFRS-UQP framework is split into two processes, where filtering and switching are carried out. The data mining based filtering in our research work uses the Hamilton Filtering framework to filter user result based on personalized information on automatic updated profiles through search engine. Maximized result is fetched, that is, filtered out with respect to user behavior profiles. The switching performs accurate filtering updated profiles using regime switching. The updating in profile change (i.e., switches) regime in HFRS-UQP framework identifies the second- and higher-order association of query result on the updated profiles. Experiment is conducted on factors such as personalized information search retrieval rate, filtering efficiency, and precision ratio. PMID:26221626
DOE Office of Scientific and Technical Information (OSTI.GOV)
Raymond, David W.; Gaither, Katherine N.; Polsky, Yarom
Sandia National Laboratories (Sandia) has a long history in developing compact, mobile, very high-speed drilling systems and this technology could be applied to increasing the rate at which boreholes are drilled during a mine accident response. The present study reviews current technical approaches, primarily based on technology developed under other programs, analyzes mine rescue specific requirements to develop a conceptual mine rescue drilling approach, and finally, proposes development of a phased mine rescue drilling system (MRDS) that accomplishes (1) development of rapid drilling MRDS equipment; (2) structuring improved web communication through the Mine Safety & Health Administration (MSHA) web site;more » (3) development of an improved protocol for employment of existing drilling technology in emergencies; (4) deployment of advanced technologies to complement mine rescue drilling operations during emergency events; and (5) preliminary discussion of potential future technology development of specialized MRDS equipment. This phased approach allows for rapid fielding of a basic system for improved rescue drilling, with the ability to improve the system over time at a reasonable cost.« less
Hogsden, Kristy L; Harding, Jon S
2012-03-01
We compared food web structure in 20 streams with either anthropogenic or natural sources of acidity and metals or circumneutral water chemistry in New Zealand. Community and diet analysis indicated that mining streams receiving anthropogenic inputs of acidic and metal-rich drainage had much simpler food webs (fewer species, shorter food chains, less links) than those in naturally acidic, naturally high metal, and circumneutral streams. Food webs of naturally high metal streams were structurally similar to those in mining streams, lacking fish predators and having few species. Whereas, webs in naturally acidic streams differed very little from those in circumneutral streams due to strong similarities in community composition and diets of secondary and top consumers. The combined negative effects of acidity and metals on stream food webs are clear. However, elevated metal concentrations, regardless of source, appear to play a more important role than acidity in driving food web structure. Copyright © 2011 Elsevier Ltd. All rights reserved.
Kamel Boulos, Maged N; Sanfilippo, Antonio P; Corley, Courtney D; Wheeler, Steve
2010-10-01
This paper explores Technosocial Predictive Analytics (TPA) and related methods for Web "data mining" where users' posts and queries are garnered from Social Web ("Web 2.0") tools such as blogs, micro-blogging and social networking sites to form coherent representations of real-time health events. The paper includes a brief introduction to commonly used Social Web tools such as mashups and aggregators, and maps their exponential growth as an open architecture of participation for the masses and an emerging way to gain insight about people's collective health status of whole populations. Several health related tool examples are described and demonstrated as practical means through which health professionals might create clear location specific pictures of epidemiological data such as flu outbreaks. Copyright 2010 Elsevier Ireland Ltd. All rights reserved.
Basner, Jodi E; Theisz, Katrina I; Jensen, Unni S; Jones, C David; Ponomarev, Ilya; Sulima, Pawel; Jo, Karen; Eljanne, Mariam; Espey, Michael G; Franca-Koh, Jonathan; Hanlon, Sean E; Kuhn, Nastaran Z; Nagahara, Larry A; Schnell, Joshua D; Moore, Nicole M
2013-12-01
Development of effective quantitative indicators and methodologies to assess the outcomes of cross-disciplinary collaborative initiatives has the potential to improve scientific program management and scientific output. This article highlights an example of a prospective evaluation that has been developed to monitor and improve progress of the National Cancer Institute Physical Sciences-Oncology Centers (PS-OC) program. Study data, including collaboration information, was captured through progress reports and compiled using the web-based analytic database: Interdisciplinary Team Reporting, Analysis, and Query Resource. Analysis of collaborations was further supported by data from the Thomson Reuters Web of Science database, MEDLINE database, and a web-based survey. Integration of novel and standard data sources was augmented by the development of automated methods to mine investigator pre-award publications, assign investigator disciplines, and distinguish cross-disciplinary publication content. The results highlight increases in cross-disciplinary authorship collaborations from pre- to post-award years among the primary investigators and confirm that a majority of cross-disciplinary collaborations have resulted in publications with cross-disciplinary content that rank in the top third of their field. With these evaluation data, PS-OC Program officials have provided ongoing feedback to participating investigators to improve center productivity and thereby facilitate a more successful initiative. Future analysis will continue to expand these methods and metrics to adapt to new advances in research evaluation and changes in the program.
Rule-based statistical data mining agents for an e-commerce application
NASA Astrophysics Data System (ADS)
Qin, Yi; Zhang, Yan-Qing; King, K. N.; Sunderraman, Rajshekhar
2003-03-01
Intelligent data mining techniques have useful e-Business applications. Because an e-Commerce application is related to multiple domains such as statistical analysis, market competition, price comparison, profit improvement and personal preferences, this paper presents a hybrid knowledge-based e-Commerce system fusing intelligent techniques, statistical data mining, and personal information to enhance QoS (Quality of Service) of e-Commerce. A Web-based e-Commerce application software system, eDVD Web Shopping Center, is successfully implemented uisng Java servlets and an Oracle81 database server. Simulation results have shown that the hybrid intelligent e-Commerce system is able to make smart decisions for different customers.
NASA Astrophysics Data System (ADS)
Damm, Bodo; Klose, Martin
2014-05-01
This contribution presents an initiative to develop a national landslide database for the Federal Republic of Germany. It highlights structure and contents of the landslide database and outlines its major data sources and the strategy of information retrieval. Furthermore, the contribution exemplifies the database potentials in applied landslide impact research, including statistics of landslide damage, repair, and mitigation. The landslide database offers due to systematic regional data compilation a differentiated data pool of more than 5,000 data sets and over 13,000 single data files. It dates back to 1137 AD and covers landslide sites throughout Germany. In seven main data blocks, the landslide database stores besides information on landslide types, dimensions, and processes, additional data on soil and bedrock properties, geomorphometry, and climatic or other major triggering events. A peculiarity of this landslide database is its storage of data sets on land use effects, damage impacts, hazard mitigation, and landslide costs. Compilation of landslide data is based on a two-tier strategy of data collection. The first step of information retrieval includes systematic web content mining and exploration of online archives of emergency agencies, fire and police departments, and news organizations. Using web and RSS feeds and soon also a focused web crawler, this enables effective nationwide data collection for recent landslides. On the basis of this information, in-depth data mining is performed to deepen and diversify the data pool in key landslide areas. This enables to gather detailed landslide information from, amongst others, agency records, geotechnical reports, climate statistics, maps, and satellite imagery. Landslide data is extracted from these information sources using a mix of methods, including statistical techniques, imagery analysis, and qualitative text interpretation. The landslide database is currently migrated to a spatial database system running on PostgreSQL/PostGIS. This provides advanced functionality for spatial data analysis and forms the basis for future data provision and visualization using a WebGIS application. Analysis of landslide database contents shows that in most parts of Germany landslides primarily affect transportation infrastructures. Although with distinct lower frequency, recent landslides are also recorded to cause serious damage to hydraulic facilities and waterways, supply and disposal infrastructures, sites of cultural heritage, as well as forest, agricultural, and mining areas. The main types of landslide damage are failure of cut and fill slopes, destruction of retaining walls, street lights, and forest stocks, burial of roads, backyards, and garden areas, as well as crack formation in foundations, sewer lines, and building walls. Landslide repair and mitigation at transportation infrastructures is dominated by simple solutions such as catch barriers or rock fall drapery. These solutions are often undersized and fail under stress. The use of costly slope stabilization or protection systems is proven to reduce these risks effectively over longer maintenance cycles. The right balancing of landslide mitigation is thus a crucial problem in managing landslide risks. Development and analysis of such landslide databases helps to support decision-makers in finding efficient solutions to minimize landslide risks for human beings, infrastructures, and financial assets.
Research on the optimization strategy of web search engine based on data mining
NASA Astrophysics Data System (ADS)
Chen, Ronghua
2018-04-01
With the wide application of search engines, web site information has become an important way for people to obtain information. People have found that they are growing in an increasingly explosive manner. Web site information is verydifficult to find the information they need, and now the search engine can not meet the need, so there is an urgent need for the network to provide website personalized information service, data mining technology for this new challenge is to find a breakthrough. In order to improve people's accuracy of finding information from websites, a website search engine optimization strategy based on data mining is proposed, and verified by website search engine optimization experiment. The results show that the proposed strategy improves the accuracy of the people to find information, and reduces the time for people to find information. It has an important practical value.
The Voice of Chinese Health Consumers: A Text Mining Approach to Web-Based Physician Reviews
Zhang, Kunpeng
2016-01-01
Background Many Web-based health care platforms allow patients to evaluate physicians by posting open-end textual reviews based on their experiences. These reviews are helpful resources for other patients to choose high-quality doctors, especially in countries like China where no doctor referral systems exist. Analyzing such a large amount of user-generated content to understand the voice of health consumers has attracted much attention from health care providers and health care researchers. Objective The aim of this paper is to automatically extract hidden topics from Web-based physician reviews using text-mining techniques to examine what Chinese patients have said about their doctors and whether these topics differ across various specialties. This knowledge will help health care consumers, providers, and researchers better understand this information. Methods We conducted two-fold analyses on the data collected from the “Good Doctor Online” platform, the largest online health community in China. First, we explored all reviews from 2006-2014 using descriptive statistics. Second, we applied the well-known topic extraction algorithm Latent Dirichlet Allocation to more than 500,000 textual reviews from over 75,000 Chinese doctors across four major specialty areas to understand what Chinese health consumers said online about their doctor visits. Results On the “Good Doctor Online” platform, 112,873 out of 314,624 doctors had been reviewed at least once by April 11, 2014. Among the 772,979 textual reviews, we chose to focus on four major specialty areas that received the most reviews: Internal Medicine, Surgery, Obstetrics/Gynecology and Pediatrics, and Chinese Traditional Medicine. Among the doctors who received reviews from those four medical specialties, two-thirds of them received more than two reviews and in a few extreme cases, some doctors received more than 500 reviews. Across the four major areas, the most popular topics reviewers found were the experience of finding doctors, doctors’ technical skills and bedside manner, general appreciation from patients, and description of various symptoms. Conclusions To the best of our knowledge, our work is the first study using an automated text-mining approach to analyze a large amount of unstructured textual data of Web-based physician reviews in China. Based on our analysis, we found that Chinese reviewers mainly concentrate on a few popular topics. This is consistent with the goal of Chinese online health platforms and demonstrates the health care focus in China’s health care system. Our text-mining approach reveals a new research area on how to use big data to help health care providers, health care administrators, and policy makers hear patient voices, target patient concerns, and improve the quality of care in this age of patient-centered care. Also, on the health care consumer side, our text mining technique helps patients make more informed decisions about which specialists to see without reading thousands of reviews, which is simply not feasible. In addition, our comparison analysis of Web-based physician reviews in China and the United States also indicates some cultural differences. PMID:27165558
The Voice of Chinese Health Consumers: A Text Mining Approach to Web-Based Physician Reviews.
Hao, Haijing; Zhang, Kunpeng
2016-05-10
Many Web-based health care platforms allow patients to evaluate physicians by posting open-end textual reviews based on their experiences. These reviews are helpful resources for other patients to choose high-quality doctors, especially in countries like China where no doctor referral systems exist. Analyzing such a large amount of user-generated content to understand the voice of health consumers has attracted much attention from health care providers and health care researchers. The aim of this paper is to automatically extract hidden topics from Web-based physician reviews using text-mining techniques to examine what Chinese patients have said about their doctors and whether these topics differ across various specialties. This knowledge will help health care consumers, providers, and researchers better understand this information. We conducted two-fold analyses on the data collected from the "Good Doctor Online" platform, the largest online health community in China. First, we explored all reviews from 2006-2014 using descriptive statistics. Second, we applied the well-known topic extraction algorithm Latent Dirichlet Allocation to more than 500,000 textual reviews from over 75,000 Chinese doctors across four major specialty areas to understand what Chinese health consumers said online about their doctor visits. On the "Good Doctor Online" platform, 112,873 out of 314,624 doctors had been reviewed at least once by April 11, 2014. Among the 772,979 textual reviews, we chose to focus on four major specialty areas that received the most reviews: Internal Medicine, Surgery, Obstetrics/Gynecology and Pediatrics, and Chinese Traditional Medicine. Among the doctors who received reviews from those four medical specialties, two-thirds of them received more than two reviews and in a few extreme cases, some doctors received more than 500 reviews. Across the four major areas, the most popular topics reviewers found were the experience of finding doctors, doctors' technical skills and bedside manner, general appreciation from patients, and description of various symptoms. To the best of our knowledge, our work is the first study using an automated text-mining approach to analyze a large amount of unstructured textual data of Web-based physician reviews in China. Based on our analysis, we found that Chinese reviewers mainly concentrate on a few popular topics. This is consistent with the goal of Chinese online health platforms and demonstrates the health care focus in China's health care system. Our text-mining approach reveals a new research area on how to use big data to help health care providers, health care administrators, and policy makers hear patient voices, target patient concerns, and improve the quality of care in this age of patient-centered care. Also, on the health care consumer side, our text mining technique helps patients make more informed decisions about which specialists to see without reading thousands of reviews, which is simply not feasible. In addition, our comparison analysis of Web-based physician reviews in China and the United States also indicates some cultural differences.
Geochemical baseline studies of soil in Finland
NASA Astrophysics Data System (ADS)
Pihlaja, Jouni
2017-04-01
The soil element concentrations regionally vary a lot in Finland. Mostly this is caused by the different bedrock types, which are reflected in the soil qualities. Geological Survey of Finland (GTK) is carrying out geochemical baseline studies in Finland. In the previous phase, the research is focusing on urban areas and mine environments. The information can, for example, be used to determine the need for soil remediation, to assess environmental impacts or to measure the natural state of soil in industrial areas or mine districts. The field work is done by taking soil samples, typically at depth between 0-10 cm. Sampling sites are chosen to represent the most vulnerable areas when thinking of human impacts by possible toxic soil element contents: playgrounds, day-care centers, schools, parks and residential areas. In the mine districts the samples are taken from the areas locating outside the airborne dust effected areas. Element contents of the soil samples are then analyzed with ICP-AES and ICP-MS, Hg with CV-AAS. The results of the geochemical baseline studies are published in the Finnish national geochemical baseline database (TAPIR). The geochemical baseline map service is free for all users via internet browser. Through this map service it is possible to calculate regional soil baseline values using geochemical data stored in the map service database. Baseline data for 17 elements in total is provided in the map service and it can be viewed on the GTK's web pages (http://gtkdata.gtk.fi/Tapir/indexEN.html).
Web page provides narrative of What's New?, Site Description, Site Risk, Cleanup Progress, Community Involvement, Next Steps, Site Documents, FAQ, Contacts and LInks for the Upper Animas Mining District site in San Juan County, Colorado.
Data Mining of Extremely Large Ad-Hoc Data Sets to Produce Reverse Web-Link Graphs
2017-03-01
in most of the MR cases. From these studies , we also learned that computing -optimized instances should be chosen for serialized/compressed input data...maximum 200 words) Data mining can be a valuable tool, particularly in the acquisition of military intelligence. As the second study within a larger Naval...open web crawler data set Common Crawl. Similar to previous studies , this research employs MapReduce (MR) for sorting and categorizing output value
MiMiR – an integrated platform for microarray data sharing, mining and analysis
Tomlinson, Chris; Thimma, Manjula; Alexandrakis, Stelios; Castillo, Tito; Dennis, Jayne L; Brooks, Anthony; Bradley, Thomas; Turnbull, Carly; Blaveri, Ekaterini; Barton, Geraint; Chiba, Norie; Maratou, Klio; Soutter, Pat; Aitman, Tim; Game, Laurence
2008-01-01
Background Despite considerable efforts within the microarray community for standardising data format, content and description, microarray technologies present major challenges in managing, sharing, analysing and re-using the large amount of data generated locally or internationally. Additionally, it is recognised that inconsistent and low quality experimental annotation in public data repositories significantly compromises the re-use of microarray data for meta-analysis. MiMiR, the Microarray data Mining Resource was designed to tackle some of these limitations and challenges. Here we present new software components and enhancements to the original infrastructure that increase accessibility, utility and opportunities for large scale mining of experimental and clinical data. Results A user friendly Online Annotation Tool allows researchers to submit detailed experimental information via the web at the time of data generation rather than at the time of publication. This ensures the easy access and high accuracy of meta-data collected. Experiments are programmatically built in the MiMiR database from the submitted information and details are systematically curated and further annotated by a team of trained annotators using a new Curation and Annotation Tool. Clinical information can be annotated and coded with a clinical Data Mapping Tool within an appropriate ethical framework. Users can visualise experimental annotation, assess data quality, download and share data via a web-based experiment browser called MiMiR Online. All requests to access data in MiMiR are routed through a sophisticated middleware security layer thereby allowing secure data access and sharing amongst MiMiR registered users prior to publication. Data in MiMiR can be mined and analysed using the integrated EMAAS open source analysis web portal or via export of data and meta-data into Rosetta Resolver data analysis package. Conclusion The new MiMiR suite of software enables systematic and effective capture of extensive experimental and clinical information with the highest MIAME score, and secure data sharing prior to publication. MiMiR currently contains more than 150 experiments corresponding to over 3000 hybridisations and supports the Microarray Centre's large microarray user community and two international consortia. The MiMiR flexible and scalable hardware and software architecture enables secure warehousing of thousands of datasets, including clinical studies, from microarray and potentially other -omics technologies. PMID:18801157
MiMiR--an integrated platform for microarray data sharing, mining and analysis.
Tomlinson, Chris; Thimma, Manjula; Alexandrakis, Stelios; Castillo, Tito; Dennis, Jayne L; Brooks, Anthony; Bradley, Thomas; Turnbull, Carly; Blaveri, Ekaterini; Barton, Geraint; Chiba, Norie; Maratou, Klio; Soutter, Pat; Aitman, Tim; Game, Laurence
2008-09-18
Despite considerable efforts within the microarray community for standardising data format, content and description, microarray technologies present major challenges in managing, sharing, analysing and re-using the large amount of data generated locally or internationally. Additionally, it is recognised that inconsistent and low quality experimental annotation in public data repositories significantly compromises the re-use of microarray data for meta-analysis. MiMiR, the Microarray data Mining Resource was designed to tackle some of these limitations and challenges. Here we present new software components and enhancements to the original infrastructure that increase accessibility, utility and opportunities for large scale mining of experimental and clinical data. A user friendly Online Annotation Tool allows researchers to submit detailed experimental information via the web at the time of data generation rather than at the time of publication. This ensures the easy access and high accuracy of meta-data collected. Experiments are programmatically built in the MiMiR database from the submitted information and details are systematically curated and further annotated by a team of trained annotators using a new Curation and Annotation Tool. Clinical information can be annotated and coded with a clinical Data Mapping Tool within an appropriate ethical framework. Users can visualise experimental annotation, assess data quality, download and share data via a web-based experiment browser called MiMiR Online. All requests to access data in MiMiR are routed through a sophisticated middleware security layer thereby allowing secure data access and sharing amongst MiMiR registered users prior to publication. Data in MiMiR can be mined and analysed using the integrated EMAAS open source analysis web portal or via export of data and meta-data into Rosetta Resolver data analysis package. The new MiMiR suite of software enables systematic and effective capture of extensive experimental and clinical information with the highest MIAME score, and secure data sharing prior to publication. MiMiR currently contains more than 150 experiments corresponding to over 3000 hybridisations and supports the Microarray Centre's large microarray user community and two international consortia. The MiMiR flexible and scalable hardware and software architecture enables secure warehousing of thousands of datasets, including clinical studies, from microarray and potentially other -omics technologies.
Mining Social Media and Web Searches For Disease Detection
Yang, Y. Tony; Horneffer, Michael; DiLisio, Nicole
2013-01-01
Web-based social media is increasingly being used across different settings in the health care industry. The increased frequency in the use of the Internet via computer or mobile devices provides an opportunity for social media to be the medium through which people can be provided with valuable health information quickly and directly. While traditional methods of detection relied predominately on hierarchical or bureaucratic lines of communication, these often failed to yield timely and accurate epidemiological intelligence. New web-based platforms promise increased opportunities for a more timely and accurate spreading of information and analysis. This article aims to provide an overview and discussion of the availability of timely and accurate information. It is especially useful for the rapid identification of an outbreak of an infectious disease that is necessary to promptly and effectively develop public health responses. These web-based platforms include search queries, data mining of web and social media, process and analysis of blogs containing epidemic key words, text mining, and geographical information system data analyses. These new sources of analysis and information are intended to complement traditional sources of epidemic intelligence. Despite the attractiveness of these new approaches, further study is needed to determine the accuracy of blogger statements, as increases in public participation may not necessarily mean the information provided is more accurate. PMID:25170475
Mining social media and web searches for disease detection.
Yang, Y Tony; Horneffer, Michael; DiLisio, Nicole
2013-04-28
Web-based social media is increasingly being used across different settings in the health care industry. The increased frequency in the use of the Internet via computer or mobile devices provides an opportunity for social media to be the medium through which people can be provided with valuable health information quickly and directly. While traditional methods of detection relied predominately on hierarchical or bureaucratic lines of communication, these often failed to yield timely and accurate epidemiological intelligence. New web-based platforms promise increased opportunities for a more timely and accurate spreading of information and analysis. This article aims to provide an overview and discussion of the availability of timely and accurate information. It is especially useful for the rapid identification of an outbreak of an infectious disease that is necessary to promptly and effectively develop public health responses. These web-based platforms include search queries, data mining of web and social media, process and analysis of blogs containing epidemic key words, text mining, and geographical information system data analyses. These new sources of analysis and information are intended to complement traditional sources of epidemic intelligence. Despite the attractiveness of these new approaches, further study is needed to determine the accuracy of blogger statements, as increases in public participation may not necessarily mean the information provided is more accurate.
GoWeb: a semantic search engine for the life science web.
Dietze, Heiko; Schroeder, Michael
2009-10-01
Current search engines are keyword-based. Semantic technologies promise a next generation of semantic search engines, which will be able to answer questions. Current approaches either apply natural language processing to unstructured text or they assume the existence of structured statements over which they can reason. Here, we introduce a third approach, GoWeb, which combines classical keyword-based Web search with text-mining and ontologies to navigate large results sets and facilitate question answering. We evaluate GoWeb on three benchmarks of questions on genes and functions, on symptoms and diseases, and on proteins and diseases. The first benchmark is based on the BioCreAtivE 1 Task 2 and links 457 gene names with 1352 functions. GoWeb finds 58% of the functional GeneOntology annotations. The second benchmark is based on 26 case reports and links symptoms with diseases. GoWeb achieves 77% success rate improving an existing approach by nearly 20%. The third benchmark is based on 28 questions in the TREC genomics challenge and links proteins to diseases. GoWeb achieves a success rate of 79%. GoWeb's combination of classical Web search with text-mining and ontologies is a first step towards answering questions in the biomedical domain. GoWeb is online at: http://www.gopubmed.org/goweb.
Beyond Google: The Invisible Web in the Academic Library
ERIC Educational Resources Information Center
Devine, Jane; Egger-Sider, Francine
2004-01-01
This article analyzes the concept of the Invisible Web and its implication for academic librarianship. It offers a guide to tools that can be used to mine the Invisible Web and discusses the benefits of using the Invisible Web to promote interest in library services. In addition, the article includes an expanded definition, a literature review,…
Intelligent Information Retrieval and Web Mining Architecture Using SOA
ERIC Educational Resources Information Center
El-Bathy, Naser Ibrahim
2010-01-01
The study of this dissertation provides a solution to a very specific problem instance in the area of data mining, data warehousing, and service-oriented architecture in publishing and newspaper industries. The research question focuses on the integration of data mining and data warehousing. The research problem focuses on the development of…
Introducing Text Analytics as a Graduate Business School Course
ERIC Educational Resources Information Center
Edgington, Theresa M.
2011-01-01
Text analytics refers to the process of analyzing unstructured data from documented sources, including open-ended surveys, blogs, and other types of web dialog. Text analytics has enveloped the concept of text mining, an analysis approach influenced heavily from data mining. While text mining has been covered extensively in various computer…
Rating prediction using textual reviews
NASA Astrophysics Data System (ADS)
NithyaKalyani, A.; Ushasukhanya, S.; Nagamalleswari, TYJ; Girija, S.
2018-04-01
Information today is present in the form of opinions. Two & a half quintillion bytes are exchanged today in Internet everyday and a large amount consists of people’s speculation and reflection over an issue. It is the need of the hour to be able to mine this information that is presented to us. Sentimental analysis refers to mining of this raw information to make sense. The discipline of opinion mining has seen a lot of encouragement in the past few years augmented by involvement of social media like Instagram, Facebook, and twitter. The hidden message in this web of information is useful in several fields such as marketing, political polls, product review, forecast market movement, Identifying detractor and promoter. In this endeavor, we introduced sentiment rating system for a particular text or paragraph to determine the opinions polarity. Firstly we resolve the searching problem, tokenization, classification, and reliable content identification. Secondly we extract probability for given text or paragraph for both positive & negative sentiment value using naive bayes classifier. At last we use sentiment dictionary (SD), sentiment degree dictionary (SDD) and negation dictionary (ND) for more accuracy. Later we blend all above mentioned factor into given formula to find the rating for the review.
Zhu, Yong-heng; Li, Ke-zhong; Zhang, Heng; Han, Fei; Zhou, Ju-hua; Gao, Ting-ting
2015-02-01
A survey was carried out to investigate soil nematode communities in the plant associations of gramineae (Arthraxon lanceolatus, AL; Imperata cylindrica, IC) and leguminosae (Glycine soja, GS) in reclaimed land of copper-mine-tailings and in the plant associations of gramineae (Digitaria chrysoblephara, DC-CK) of peripheral control in Fenghuang Mountain, Tongling City. A total of 1277 nematodes were extracted and sorted into 51 genera. The average individual density of the nematodes was 590 individuals · 100 g(-1) dry soil. In order to analyze the distribution character- istics of soil nematode communities in reclaimed land of copper-mine-tailings, Shannon community diversity index and soil food web structure indices were applied in the research. The results showed that the total number of nematode genus and the Shannon community diversity index of soil nematode in the three plant associations of AL, IC and GS were less than that in the plant associations of DC-CK. Compared with the ecological indices of soil nematode communities among the different plant associations in reclaimed land of copper-mine-tailings and peripheral natural habitat, we found that the structure of soil food web in the plant associations of GS was more mature, with bacterial decomposition being dominant in the soil organic matter decomposition, and that the soil ecosystem in the plant associations of GS was not stable with low interference. This indicated that the soil food web in the plant associations of leguminosae had a greater development potential to improve the ecological stability of the reclaimed land of copper-mine-tailings. On the other hand, the structure of soil food web in the plant associations of AL and IC were relatively stable in a structured state with fungal decomposition being dominant in the decomposition of soil organic matter. This indicated that the soil food web in the plant associations of gramineae was at a poor development level.
Mining Tasks from the Web Anchor Text Graph: MSR Notebook Paper for the TREC 2015 Tasks Track
2015-11-20
Mining Tasks from the Web Anchor Text Graph: MSR Notebook Paper for the TREC 2015 Tasks Track Paul N. Bennett Microsoft Research Redmond, USA pauben...anchor text graph has proven useful in the general realm of query reformulation [2], we sought to quantify the value of extracting key phrases from...anchor text in the broader setting of the task understanding track. Given a query, our approach considers a simple method for identifying a relevant
Web Video Event Recognition by Semantic Analysis From Ubiquitous Documents.
Yu, Litao; Yang, Yang; Huang, Zi; Wang, Peng; Song, Jingkuan; Shen, Heng Tao
2016-12-01
In recent years, the task of event recognition from videos has attracted increasing interest in multimedia area. While most of the existing research was mainly focused on exploring visual cues to handle relatively small-granular events, it is difficult to directly analyze video content without any prior knowledge. Therefore, synthesizing both the visual and semantic analysis is a natural way for video event understanding. In this paper, we study the problem of Web video event recognition, where Web videos often describe large-granular events and carry limited textual information. Key challenges include how to accurately represent event semantics from incomplete textual information and how to effectively explore the correlation between visual and textual cues for video event understanding. We propose a novel framework to perform complex event recognition from Web videos. In order to compensate the insufficient expressive power of visual cues, we construct an event knowledge base by deeply mining semantic information from ubiquitous Web documents. This event knowledge base is capable of describing each event with comprehensive semantics. By utilizing this base, the textual cues for a video can be significantly enriched. Furthermore, we introduce a two-view adaptive regression model, which explores the intrinsic correlation between the visual and textual cues of the videos to learn reliable classifiers. Extensive experiments on two real-world video data sets show the effectiveness of our proposed framework and prove that the event knowledge base indeed helps improve the performance of Web video event recognition.
CMD: a Cotton Microsatellite Database resource for Gossypium genomics
Blenda, Anna; Scheffler, Jodi; Scheffler, Brian; Palmer, Michael; Lacape, Jean-Marc; Yu, John Z; Jesudurai, Christopher; Jung, Sook; Muthukumar, Sriram; Yellambalase, Preetham; Ficklin, Stephen; Staton, Margaret; Eshelman, Robert; Ulloa, Mauricio; Saha, Sukumar; Burr, Ben; Liu, Shaolin; Zhang, Tianzhen; Fang, Deqiu; Pepper, Alan; Kumpatla, Siva; Jacobs, John; Tomkins, Jeff; Cantrell, Roy; Main, Dorrie
2006-01-01
Background The Cotton Microsatellite Database (CMD) is a curated and integrated web-based relational database providing centralized access to publicly available cotton microsatellites, an invaluable resource for basic and applied research in cotton breeding. Description At present CMD contains publication, sequence, primer, mapping and homology data for nine major cotton microsatellite projects, collectively representing 5,484 microsatellites. In addition, CMD displays data for three of the microsatellite projects that have been screened against a panel of core germplasm. The standardized panel consists of 12 diverse genotypes including genetic standards, mapping parents, BAC donors, subgenome representatives, unique breeding lines, exotic introgression sources, and contemporary Upland cottons with significant acreage. A suite of online microsatellite data mining tools are accessible at CMD. These include an SSR server which identifies microsatellites, primers, open reading frames, and GC-content of uploaded sequences; BLAST and FASTA servers providing sequence similarity searches against the existing cotton SSR sequences and primers, a CAP3 server to assemble EST sequences into longer transcripts prior to mining for SSRs, and CMap, a viewer for comparing cotton SSR maps. Conclusion The collection of publicly available cotton SSR markers in a centralized, readily accessible and curated web-enabled database provides a more efficient utilization of microsatellite resources and will help accelerate basic and applied research in molecular breeding and genetic mapping in Gossypium spp. PMID:16737546
Tourassi, Georgia; Yoon, Hong-Jun; Xu, Songhua; Han, Xuesong
2016-05-01
The World Wide Web has emerged as a powerful data source for epidemiological studies related to infectious disease surveillance. However, its potential for cancer-related epidemiological discoveries is largely unexplored. Using advanced web crawling and tailored information extraction procedures, the authors automatically collected and analyzed the text content of 79 394 online obituary articles published between 1998 and 2014. The collected data included 51 911 cancer (27 330 breast; 9470 lung; 6496 pancreatic; 6342 ovarian; 2273 colon) and 27 483 non-cancer cases. With the derived information, the authors replicated a case-control study design to investigate the association between parity (i.e., childbearing) and cancer risk. Age-adjusted odds ratios (ORs) with 95% confidence intervals (CIs) were calculated for each cancer type and compared to those reported in large-scale epidemiological studies. Parity was found to be associated with a significantly reduced risk of breast cancer (OR = 0.78, 95% CI, 0.75-0.82), pancreatic cancer (OR = 0.78, 95% CI, 0.72-0.83), colon cancer (OR = 0.67, 95% CI, 0.60-0.74), and ovarian cancer (OR = 0.58, 95% CI, 0.54-0.62). Marginal association was found for lung cancer risk (OR = 0.87, 95% CI, 0.81-0.92). The linear trend between increased parity and reduced cancer risk was dramatically more pronounced for breast and ovarian cancer than the other cancers included in the analysis. This large web-mining study on parity and cancer risk produced findings very similar to those reported with traditional observational studies. It may be used as a promising strategy to generate study hypotheses for guiding and prioritizing future epidemiological studies. © The Author 2015. Published by Oxford University Press on behalf of the American Medical Informatics Association. All rights reserved. For Permissions, please email: journals.permissions@oup.com.
BAGEL4: a user-friendly web server to thoroughly mine RiPPs and bacteriocins.
van Heel, Auke J; de Jong, Anne; Song, Chunxu; Viel, Jakob H; Kok, Jan; Kuipers, Oscar P
2018-05-21
Interest in secondary metabolites such as RiPPs (ribosomally synthesized and posttranslationally modified peptides) is increasing worldwide. To facilitate the research in this field we have updated our mining web server. BAGEL4 is faster than its predecessor and is now fully independent from ORF-calling. Gene clusters of interest are discovered using the core-peptide database and/or through HMM motifs that are present in associated context genes. The databases used for mining have been updated and extended with literature references and links to UniProt and NCBI. Additionally, we have included automated promoter and terminator prediction and the option to upload RNA expression data, which can be displayed along with the identified clusters. Further improvements include the annotation of the context genes, which is now based on a fast blast against the prokaryote part of the UniRef90 database, and the improved web-BLAST feature that dynamically loads structural data such as internal cross-linking from UniProt. Overall BAGEL4 provides the user with more information through a user-friendly web-interface which simplifies data evaluation. BAGEL4 is freely accessible at http://bagel4.molgenrug.nl.
Asymmetric threat data mining and knowledge discovery
NASA Astrophysics Data System (ADS)
Gilmore, John F.; Pagels, Michael A.; Palk, Justin
2001-03-01
Asymmetric threats differ from the conventional force-on- force military encounters that the Defense Department has historically been trained to engage. Terrorism by its nature is now an operational activity that is neither easily detected or countered as its very existence depends on small covert attacks exploiting the element of surprise. But terrorism does have defined forms, motivations, tactics and organizational structure. Exploiting a terrorism taxonomy provides the opportunity to discover and assess knowledge of terrorist operations. This paper describes the Asymmetric Threat Terrorist Assessment, Countering, and Knowledge (ATTACK) system. ATTACK has been developed to (a) data mine open source intelligence (OSINT) information from web-based newspaper sources, video news web casts, and actual terrorist web sites, (b) evaluate this information against a terrorism taxonomy, (c) exploit country/region specific social, economic, political, and religious knowledge, and (d) discover and predict potential terrorist activities and association links. Details of the asymmetric threat structure and the ATTACK system architecture are presented with results of an actual terrorist data mining and knowledge discovery test case shown.
Duncan, Dean F; Kum, Hye-Chung; Weigensberg, Elizabeth Caplick; Flair, Kimberly A; Stewart, C Joy
2008-11-01
Proper management and implementation of an effective child welfare agency requires the constant use of information about the experiences and outcomes of children involved in the system, emphasizing the need for comprehensive, timely, and accurate data. In the past 20 years, there have been many advances in technology that can maximize the potential of administrative data to promote better evaluation and management in the field of child welfare. Specifically, this article discusses the use of knowledge discovery and data mining (KDD), which makes it possible to create longitudinal data files from administrative data sources, extract valuable knowledge, and make the information available via a user-friendly public Web site. This article demonstrates a successful project in North Carolina where knowledge discovery and data mining technology was used to develop a comprehensive set of child welfare outcomes available through a public Web site to facilitate information sharing of child welfare data to improve policy and practice.
Chemotext: A Publicly Available Web Server for Mining Drug-Target-Disease Relationships in PubMed.
Capuzzi, Stephen J; Thornton, Thomas E; Liu, Kammy; Baker, Nancy; Lam, Wai In; O'Banion, Colin P; Muratov, Eugene N; Pozefsky, Diane; Tropsha, Alexander
2018-02-26
Elucidation of the mechanistic relationships between drugs, their targets, and diseases is at the core of modern drug discovery research. Thousands of studies relevant to the drug-target-disease (DTD) triangle have been published and annotated in the Medline/PubMed database. Mining this database affords rapid identification of all published studies that confirm connections between vertices of this triangle or enable new inferences of such connections. To this end, we describe the development of Chemotext, a publicly available Web server that mines the entire compendium of published literature in PubMed annotated by Medline Subject Heading (MeSH) terms. The goal of Chemotext is to identify all known DTD relationships and infer missing links between vertices of the DTD triangle. As a proof-of-concept, we show that Chemotext could be instrumental in generating new drug repurposing hypotheses or annotating clinical outcomes pathways for known drugs. The Chemotext Web server is freely available at http://chemotext.mml.unc.edu .
Basner, Jodi E.; Theisz, Katrina I.; Jensen, Unni S.; Jones, C. David; Ponomarev, Ilya; Sulima, Pawel; Jo, Karen; Eljanne, Mariam; Espey, Michael G.; Franca-Koh, Jonathan; Hanlon, Sean E.; Kuhn, Nastaran Z.; Nagahara, Larry A.; Schnell, Joshua D.; Moore, Nicole M.
2013-01-01
Development of effective quantitative indicators and methodologies to assess the outcomes of cross-disciplinary collaborative initiatives has the potential to improve scientific program management and scientific output. This article highlights an example of a prospective evaluation that has been developed to monitor and improve progress of the National Cancer Institute Physical Sciences—Oncology Centers (PS-OC) program. Study data, including collaboration information, was captured through progress reports and compiled using the web-based analytic database: Interdisciplinary Team Reporting, Analysis, and Query Resource. Analysis of collaborations was further supported by data from the Thomson Reuters Web of Science database, MEDLINE database, and a web-based survey. Integration of novel and standard data sources was augmented by the development of automated methods to mine investigator pre-award publications, assign investigator disciplines, and distinguish cross-disciplinary publication content. The results highlight increases in cross-disciplinary authorship collaborations from pre- to post-award years among the primary investigators and confirm that a majority of cross-disciplinary collaborations have resulted in publications with cross-disciplinary content that rank in the top third of their field. With these evaluation data, PS-OC Program officials have provided ongoing feedback to participating investigators to improve center productivity and thereby facilitate a more successful initiative. Future analysis will continue to expand these methods and metrics to adapt to new advances in research evaluation and changes in the program. PMID:24808632
30 CFR 75.403-1 - Incombustible content.
Code of Federal Regulations, 2013 CFR
2013-07-01
... 30 Mineral Resources 1 2013-07-01 2013-07-01 false Incombustible content. 75.403-1 Section 75.403-1 Mineral Resources MINE SAFETY AND HEALTH ADMINISTRATION, DEPARTMENT OF LABOR COAL MINE SAFETY AND HEALTH MANDATORY SAFETY STANDARDS-UNDERGROUND COAL MINES Combustible Materials and Rock Dusting § 75.403-1 Incombustible content. Moisture contained...
30 CFR 75.403-1 - Incombustible content.
Code of Federal Regulations, 2014 CFR
2014-07-01
... 30 Mineral Resources 1 2014-07-01 2014-07-01 false Incombustible content. 75.403-1 Section 75.403-1 Mineral Resources MINE SAFETY AND HEALTH ADMINISTRATION, DEPARTMENT OF LABOR COAL MINE SAFETY AND HEALTH MANDATORY SAFETY STANDARDS-UNDERGROUND COAL MINES Combustible Materials and Rock Dusting § 75.403-1 Incombustible content. Moisture contained...
30 CFR 75.403-1 - Incombustible content.
Code of Federal Regulations, 2012 CFR
2012-07-01
... 30 Mineral Resources 1 2012-07-01 2012-07-01 false Incombustible content. 75.403-1 Section 75.403-1 Mineral Resources MINE SAFETY AND HEALTH ADMINISTRATION, DEPARTMENT OF LABOR COAL MINE SAFETY AND HEALTH MANDATORY SAFETY STANDARDS-UNDERGROUND COAL MINES Combustible Materials and Rock Dusting § 75.403-1 Incombustible content. Moisture contained...
30 CFR 75.403-1 - Incombustible content.
Code of Federal Regulations, 2010 CFR
2010-07-01
... 30 Mineral Resources 1 2010-07-01 2010-07-01 false Incombustible content. 75.403-1 Section 75.403-1 Mineral Resources MINE SAFETY AND HEALTH ADMINISTRATION, DEPARTMENT OF LABOR COAL MINE SAFETY AND HEALTH MANDATORY SAFETY STANDARDS-UNDERGROUND COAL MINES Combustible Materials and Rock Dusting § 75.403-1 Incombustible content. Moisture contained...
30 CFR 75.403-1 - Incombustible content.
Code of Federal Regulations, 2011 CFR
2011-07-01
... 30 Mineral Resources 1 2011-07-01 2011-07-01 false Incombustible content. 75.403-1 Section 75.403-1 Mineral Resources MINE SAFETY AND HEALTH ADMINISTRATION, DEPARTMENT OF LABOR COAL MINE SAFETY AND HEALTH MANDATORY SAFETY STANDARDS-UNDERGROUND COAL MINES Combustible Materials and Rock Dusting § 75.403-1 Incombustible content. Moisture contained...
Data mining for personal navigation
NASA Astrophysics Data System (ADS)
Hariharan, Gurushyam; Franti, Pasi; Mehta, Sandeep
2002-03-01
Relevance is the key in defining what data is to be extracted from the Internet. Traditionally, relevance has been defined mainly by keywords and user profiles. In this paper we discuss a fairly untouched dimension to relevance: location. Any navigational information sought by a user at large on earth is evidently governed by his location. We believe that task oriented data mining of the web amalgamated with location information is the key to providing relevant information for personal navigation. We explore the existential hurdles and propose novel approaches to tackle them. We also present naive, task-oriented data mining based approaches and their implementations in Java, to extract location based information. Ad-hoc pairing of data with coordinates (x, y) is very rare on the web. But if the same co-ordinates are converted to a logical address (state/city/street), a wide spectrum of location-based information base opens up. Hence, given the coordinates (x, y) on the earth, the scheme points to the logical address of the user. Location based information could either be picked up from fixed and known service providers (e.g. Yellow Pages) or from any arbitrary website on the Web. Once the web servers providing information relevant to the logical address are located, task oriented data mining is performed over these sites keeping in mind what information is interesting to the contemporary user. After all this, a simple data stream is provided to the user with information scaled to his convenience. The scheme has been implemented for cities of Finland.
Large-Scale Overlays and Trends: Visually Mining, Panning and Zooming the Observable Universe.
Luciani, Timothy Basil; Cherinka, Brian; Oliphant, Daniel; Myers, Sean; Wood-Vasey, W Michael; Labrinidis, Alexandros; Marai, G Elisabeta
2014-07-01
We introduce a web-based computing infrastructure to assist the visual integration, mining and interactive navigation of large-scale astronomy observations. Following an analysis of the application domain, we design a client-server architecture to fetch distributed image data and to partition local data into a spatial index structure that allows prefix-matching of spatial objects. In conjunction with hardware-accelerated pixel-based overlays and an online cross-registration pipeline, this approach allows the fetching, displaying, panning and zooming of gigabit panoramas of the sky in real time. To further facilitate the integration and mining of spatial and non-spatial data, we introduce interactive trend images-compact visual representations for identifying outlier objects and for studying trends within large collections of spatial objects of a given class. In a demonstration, images from three sky surveys (SDSS, FIRST and simulated LSST results) are cross-registered and integrated as overlays, allowing cross-spectrum analysis of astronomy observations. Trend images are interactively generated from catalog data and used to visually mine astronomy observations of similar type. The front-end of the infrastructure uses the web technologies WebGL and HTML5 to enable cross-platform, web-based functionality. Our approach attains interactive rendering framerates; its power and flexibility enables it to serve the needs of the astronomy community. Evaluation on three case studies, as well as feedback from domain experts emphasize the benefits of this visual approach to the observational astronomy field; and its potential benefits to large scale geospatial visualization in general.
Web Content Management and One EPA Web Factsheet
One EPA Web is a multi-year project to improve EPA’s website to better meet the needs of our Web visitors. Content is developed and managed in the WebCMS which supports One EPA Web goals by standardizing how we create and publish content.
ERIC Educational Resources Information Center
Cagirgan Gulten, Dilek
2013-01-01
This research aims to investigate primary preservice mathematics teachers' views on distance education and web pedagogical content knowledge in terms of the subscales of general web, communicative web, pedagogical web, web pedagogical content and attitude towards web based instruction. The research was conducted with 46 senior students in the…
Mining Hidden Gems Beneath the Surface: A Look At the Invisible Web.
ERIC Educational Resources Information Center
Carlson, Randal D.; Repman, Judi
2002-01-01
Describes resources for researchers called the Invisible Web that are hidden from the usual search engines and other tools and contrasts them with those resources available on the surface Web. Identifies specialized search tools, databases, and strategies that can be used to locate credible in-depth information. (Author/LRW)
Opinion Integration and Summarization
ERIC Educational Resources Information Center
Lu, Yue
2011-01-01
As Web 2.0 applications become increasingly popular, more and more people express their opinions on the Web in various ways in real time. Such wide coverage of topics and abundance of users make the Web an extremely valuable source for mining people's opinions about all kinds of topics. However, since the opinions are usually expressed as…
Mining Formative Evaluation Rules Using Web-Based Learning Portfolios for Web-Based Learning Systems
ERIC Educational Resources Information Center
Chen, Chih-Ming; Hong, Chin-Ming; Chen, Shyuan-Yi; Liu, Chao-Yu
2006-01-01
Learning performance assessment aims to evaluate what knowledge learners have acquired from teaching activities. Objective technical measures of learning performance are difficult to develop, but are extremely important for both teachers and learners. Learning performance assessment using learning portfolios or web server log data is becoming an…
Federal Register 2010, 2011, 2012, 2013, 2014
2013-12-24
... gold mine and associated processing and ancillary facilities. The project would be located on public... media, newspapers and the BLM Web site at: http://www.blm.gov/nv/st/en/fo/battle_mountain_field.html... to construct, operate, reclaim, and close an open pit, heap leach, gold mining operation known as the...
Entomopathogenic nematode food webs in an ancient, mining pollution gradient in Spain.
Campos-Herrera, Raquel; Rodríguez Martín, José Antonio; Escuer, Miguel; García-González, María Teresa; Duncan, Larry W; Gutiérrez, Carmen
2016-12-01
Mining activities pollute the environment with by-products that cause unpredictable impacts in surrounding areas. Cartagena-La Unión mine (Southeastern-Spain) was active for >2500years. Despite its closure in 1991, high concentrations of metals and waste residues remain in this area. A previous study using nematodes suggested that high lead content diminished soil biodiversity. However, the effects of mine pollution on specific ecosystem services remain unknown. Entomopathogenic nematodes (EPN) play a major role in the biocontrol of insect pests. Because EPNs are widespread throughout the world, we speculated that EPNs would be present in the mined areas, but at increased incidence with distance from the pollution focus. We predicted that the natural enemies of nematodes would follow a similar spatial pattern. We used qPCR techniques to measure abundance of five EPN species, five nematophagous fungi species, two bacterial ectoparasites of EPNs and one group of free-living nematodes that compete for the insect-cadaver. The study comprised 193 soil samples taken from mining sites, natural areas and agricultural fields. The highest concentrations of iron and zinc were detected in the mined area as was previously described for lead, cadmium and nickel. Molecular tools detected very low numbers of EPNs in samples found to be negative by insect-baiting, demonstrating the importance of the approach. EPNs were detected at low numbers in 13% of the localities, without relationship to heavy-metal concentrations. Only Acrobeloides-group nematodes were inversely related to the pollution gradient. Factors associated with agricultural areas explained 98.35% of the biotic variability, including EPN association with agricultural areas. Our study suggests that EPNs have adapted to polluted habitats that might support arthropod hosts. By contrast, the relationship between abundance of Acrobeloides-group and heavy-metal levels, revealed these taxa as especially well suited bio-indicators of soil mining pollution. Copyright © 2016 Elsevier B.V. All rights reserved.
Bendell, L I
2011-02-15
Archived samples of blue grouse (Dendragapus obscurus) gizzard contents, inclusive of grit, collected yearly between 1959 and 1970 were analyzed for cadmium, lead, zinc, and copper content. Approximately halfway through the 12-year sampling period, an open-pit copper mine began activities, then ceased operations 2 years later. Thus the archived samples provided a unique opportunity to determine if avian gizzard contents, inclusive of grit, could reveal patterns in the anthropogenic deposition of trace metals associated with mining activities. Gizzard concentrations of cadmium and copper strongly coincided with the onset of opening and the closing of the pit mining activity. Gizzard zinc and lead demonstrated significant among year variation; however, maximum concentrations did not correlate to mining activity. The archived gizzard contents did provide a useful tool for documenting trends in metal depositional patterns related to an anthropogenic activity. Further, blue grouse ingesting grit particles during the time of active mining activity would have been exposed to toxicologically significant levels of cadmium. Gizzard lead concentrations were also of toxicological significance but not related to mining activity. This type of "pulse" toxic metal exposure as a consequence of open-pit mining activity would not necessarily have been revealed through a "snap-shot" of soil, plant or avian tissue trace metal analysis post-mining activity. Copyright © 2010 Elsevier B.V. All rights reserved.
Opinion Mining for Educational Video Lectures.
Kravvaris, Dimitrios; Kermanidis, Katia Lida
2017-01-01
The search for relevant educational videos is a time consuming process for the users. Furthermore, the increasing demand for educational videos intensifies the problem and calls for the users to utilize whichever information is offered by the hosting web pages, and choose the most appropriate one. This research focuses on the classification of user views, based on the comments on educational videos, into positive or negative ones. The aim is to give users a picture of the positive and negative comments that have been recorded, so as to provide a qualitative view of the final selection at their disposal. The present paper's innovation is the automatic identification of the most important words of the verbal content of the video lectures and the filtering of the comments based on them, thus limiting the comments to the ones that have a substantial semantic connection with the video content.
QuadBase2: web server for multiplexed guanine quadruplex mining and visualization
Dhapola, Parashar; Chowdhury, Shantanu
2016-01-01
DNA guanine quadruplexes or G4s are non-canonical DNA secondary structures which affect genomic processes like replication, transcription and recombination. G4s are computationally identified by specific nucleotide motifs which are also called putative G4 (PG4) motifs. Despite the general relevance of these structures, there is currently no tool available that can allow batch queries and genome-wide analysis of these motifs in a user-friendly interface. QuadBase2 (quadbase.igib.res.in) presents a completely reinvented web server version of previously published QuadBase database. QuadBase2 enables users to mine PG4 motifs in up to 178 eukaryotes through the EuQuad module. This module interfaces with Ensembl Compara database, to allow users mine PG4 motifs in the orthologues of genes of interest across eukaryotes. PG4 motifs can be mined across genes and their promoter sequences in 1719 prokaryotes through ProQuad module. This module includes a feature that allows genome-wide mining of PG4 motifs and their visualization as circular histograms. TetraplexFinder, the module for mining PG4 motifs in user-provided sequences is now capable of handling up to 20 MB of data. QuadBase2 is a comprehensive PG4 motif mining tool that further expands the configurations and algorithms for mining PG4 motifs in a user-friendly way. PMID:27185890
Using ant-behavior-based simulation model AntWeb to improve website organization
NASA Astrophysics Data System (ADS)
Li, Weigang; Pinheiro Dib, Marcos V.; Teles, Wesley M.; Morais de Andrade, Vlaudemir; Alves de Melo, Alba C. M.; Cariolano, Judas T.
2002-03-01
Some web usage mining algorithms showed the potential application to find the difference among the organizations expected by visitors to the website. However, there are still no efficient method and criterion for a web administrator to measure the performance of the modification. In this paper, we developed an AntWeb, a model inspired by ants' behavior to simulate the sequence of visiting the website, in order to measure the efficient of the web structure. We implemented a web usage mining algorithm using backtrack to the intranet website of the Politec Informatic Ltd., Brazil. We defined throughput (the number of visitors to reach their target pages per time unit relates to the total number of visitors) as an index to measure the website's performance. We also used the link in a web page to represent the effect of visitors' pheromone trails. For every modification in the website organization, for example, putting a link from the expected location to the target object, the simulation reported the value of throughput as a quick answer about this modification. The experiment showed the stability of our simulation model, and a positive modification to the intranet website of the Politec.
Mining of the social network extraction
NASA Astrophysics Data System (ADS)
Nasution, M. K. M.; Hardi, M.; Syah, R.
2017-01-01
The use of Web as social media is steadily gaining ground in the study of social actor behaviour. However, information in Web can be interpreted in accordance with the ability of the method such as superficial methods for extracting social networks. Each method however has features and drawbacks: it cannot reveal the behaviour of social actors, but it has the hidden information about them. Therefore, this paper aims to reveal such information in the social networks mining. Social behaviour could be expressed through a set of words extracted from the list of snippets.
Analysing Customer Opinions with Text Mining Algorithms
NASA Astrophysics Data System (ADS)
Consoli, Domenico
2009-08-01
Knowing what the customer thinks of a particular product/service helps top management to introduce improvements in processes and products, thus differentiating the company from their competitors and gain competitive advantages. The customers, with their preferences, determine the success or failure of a company. In order to know opinions of the customers we can use technologies available from the web 2.0 (blog, wiki, forums, chat, social networking, social commerce). From these web sites, useful information must be extracted, for strategic purposes, using techniques of sentiment analysis or opinion mining.
Prusti, Marjo; Lehtineva, Susanna; Pohjanoksa-Mäntylä, Marika; Bell, J Simon
2012-01-01
The Internet is a frequently used source of drug information, including among people with mental disorders. Online drug information may be narrow in scope, incomplete, and contain errors of omission. To evaluate the quality of online antidepressant drug information in English and Finnish. Forty Web sites were identified using the search terms antidepressants and masennuslääkkeet in English and Finnish, respectively. Included Web sites (14 English, 8 Finnish) were evaluated for aesthetics, interactivity, content coverage, and content correctness using published criteria. All Web sites were assessed using the Date, Author, References, Type, Sponsor (DARTS) and DISCERN quality assessment tools. English and Finnish Web sites had similar aesthetics, content coverage, and content correctness scores. English Web sites were more interactive than Finnish Web sites (P<.05). Overall, adverse drug reactions were covered on 21 of 22 Web sites; however, drug-alcohol interactions were addressed on only 9 of 22 Web sites, and dose was addressed on only 6 of 22 Web sites. Few (2/22 Web sites) provided incorrect information. The DISCERN score was significantly correlated with content coverage (r=0.670, P<.01), content correctness (r=0.663, P<.01), and the DARTS score (r=0.459, P<.05). No Web site provided information about all aspects of antidepressant treatment. Nevertheless, few Web sites provided incorrect information. Both English and Finnish Web sites were similar in terms of aesthetics, content coverage, and content correctness. Copyright © 2012 Elsevier Inc. All rights reserved.
ERIC Educational Resources Information Center
Arnone, Marilyn P.; Small, Ruth V.
Designed for elementary or middle school teachers and library media specialists, this book provides educators with practical, easy-to-use ways of applying motivation assessment techniques when selecting World Wide Web sites for inclusion in their lessons and offers concrete examples of how to use Web evaluation with young learners. WebMAC…
Astrophysical data mining with GPU. A case study: Genetic classification of globular clusters
NASA Astrophysics Data System (ADS)
Cavuoti, S.; Garofalo, M.; Brescia, M.; Paolillo, M.; Pescape', A.; Longo, G.; Ventre, G.
2014-01-01
We present a multi-purpose genetic algorithm, designed and implemented with GPGPU/CUDA parallel computing technology. The model was derived from our CPU serial implementation, named GAME (Genetic Algorithm Model Experiment). It was successfully tested and validated on the detection of candidate Globular Clusters in deep, wide-field, single band HST images. The GPU version of GAME will be made available to the community by integrating it into the web application DAMEWARE (DAta Mining Web Application REsource, http://dame.dsf.unina.it/beta_info.html), a public data mining service specialized on massive astrophysical data. Since genetic algorithms are inherently parallel, the GPGPU computing paradigm leads to a speedup of a factor of 200× in the training phase with respect to the CPU based version.
Wiegers, Thomas C; Davis, Allan Peter; Mattingly, Carolyn J
2014-01-01
The Critical Assessment of Information Extraction systems in Biology (BioCreAtIvE) challenge evaluation tasks collectively represent a community-wide effort to evaluate a variety of text-mining and information extraction systems applied to the biological domain. The BioCreative IV Workshop included five independent subject areas, including Track 3, which focused on named-entity recognition (NER) for the Comparative Toxicogenomics Database (CTD; http://ctdbase.org). Previously, CTD had organized document ranking and NER-related tasks for the BioCreative Workshop 2012; a key finding of that effort was that interoperability and integration complexity were major impediments to the direct application of the systems to CTD's text-mining pipeline. This underscored a prevailing problem with software integration efforts. Major interoperability-related issues included lack of process modularity, operating system incompatibility, tool configuration complexity and lack of standardization of high-level inter-process communications. One approach to potentially mitigate interoperability and general integration issues is the use of Web services to abstract implementation details; rather than integrating NER tools directly, HTTP-based calls from CTD's asynchronous, batch-oriented text-mining pipeline could be made to remote NER Web services for recognition of specific biological terms using BioC (an emerging family of XML formats) for inter-process communications. To test this concept, participating groups developed Representational State Transfer /BioC-compliant Web services tailored to CTD's NER requirements. Participants were provided with a comprehensive set of training materials. CTD evaluated results obtained from the remote Web service-based URLs against a test data set of 510 manually curated scientific articles. Twelve groups participated in the challenge. Recall, precision, balanced F-scores and response times were calculated. Top balanced F-scores for gene, chemical and disease NER were 61, 74 and 51%, respectively. Response times ranged from fractions-of-a-second to over a minute per article. We present a description of the challenge and summary of results, demonstrating how curation groups can effectively use interoperable NER technologies to simplify text-mining pipeline implementation. Database URL: http://ctdbase.org/ © The Author(s) 2014. Published by Oxford University Press.
Wiegers, Thomas C.; Davis, Allan Peter; Mattingly, Carolyn J.
2014-01-01
The Critical Assessment of Information Extraction systems in Biology (BioCreAtIvE) challenge evaluation tasks collectively represent a community-wide effort to evaluate a variety of text-mining and information extraction systems applied to the biological domain. The BioCreative IV Workshop included five independent subject areas, including Track 3, which focused on named-entity recognition (NER) for the Comparative Toxicogenomics Database (CTD; http://ctdbase.org). Previously, CTD had organized document ranking and NER-related tasks for the BioCreative Workshop 2012; a key finding of that effort was that interoperability and integration complexity were major impediments to the direct application of the systems to CTD's text-mining pipeline. This underscored a prevailing problem with software integration efforts. Major interoperability-related issues included lack of process modularity, operating system incompatibility, tool configuration complexity and lack of standardization of high-level inter-process communications. One approach to potentially mitigate interoperability and general integration issues is the use of Web services to abstract implementation details; rather than integrating NER tools directly, HTTP-based calls from CTD's asynchronous, batch-oriented text-mining pipeline could be made to remote NER Web services for recognition of specific biological terms using BioC (an emerging family of XML formats) for inter-process communications. To test this concept, participating groups developed Representational State Transfer /BioC-compliant Web services tailored to CTD's NER requirements. Participants were provided with a comprehensive set of training materials. CTD evaluated results obtained from the remote Web service-based URLs against a test data set of 510 manually curated scientific articles. Twelve groups participated in the challenge. Recall, precision, balanced F-scores and response times were calculated. Top balanced F-scores for gene, chemical and disease NER were 61, 74 and 51%, respectively. Response times ranged from fractions-of-a-second to over a minute per article. We present a description of the challenge and summary of results, demonstrating how curation groups can effectively use interoperable NER technologies to simplify text-mining pipeline implementation. Database URL: http://ctdbase.org/ PMID:24919658
Operating System Support for Shared Hardware Data Structures
2013-01-31
Carbon [73] uses hardware queues to improve fine-grained multitasking for Recognition, Mining , and Synthesis. Compared to software ap- proaches...web transaction processing, data mining , and multimedia. Early work in database processors [114, 96, 79, 111] reduce the costs of relational database...assignment can be solved statically or dynamically. Static assignment deter- mines offline which data structures are assigned to use HWDS resources and at
Federal Register 2010, 2011, 2012, 2013, 2014
2012-01-27
... comments related to the Hycroft Mine Expansion Draft EIS by any of the following methods: Web site: www.blm..., Nevada 89445, Attn. Kathleen Rehberg. Copies of the Hycroft Mine Expansion Draft EIS are available in the... hours. The FIRS is available 24 hours a day, 7 days a week, to leave a message or question with the...
Sealife: a semantic grid browser for the life sciences applied to the study of infectious diseases.
Schroeder, Michael; Burger, Albert; Kostkova, Patty; Stevens, Robert; Habermann, Bianca; Dieng-Kuntz, Rose
2006-01-01
The objective of Sealife is the conception and realisation of a semantic Grid browser for the life sciences, which will link the existing Web to the currently emerging eScience infrastructure. The SeaLife Browser will allow users to automatically link a host of Web servers and Web/Grid services to the Web content he/she is visiting. This will be accomplished using eScience's growing number of Web/Grid Services and its XML-based standards and ontologies. The browser will identify terms in the pages being browsed through the background knowledge held in ontologies. Through the use of Semantic Hyperlinks, which link identified ontology terms to servers and services, the SeaLife Browser will offer a new dimension of context-based information integration. In this paper, we give an overview over the different components of the browser and their interplay. This SeaLife Browser will be demonstrated within three application scenarios in evidence-based medicine, literature & patent mining, and molecular biology, all relating to the study of infectious diseases. The three applications vertically integrate the molecule/cell, the tissue/organ and the patient/population level by covering the analysis of high-throughput screening data for endocytosis (the molecular entry pathway into the cell), the expression of proteins in the spatial context of tissue and organs, and a high-level library on infectious diseases designed for clinicians and their patients. For more information see http://www.biote.ctu-dresden.de/sealife.
Content and Workflow Management for Library Websites: Case Studies
ERIC Educational Resources Information Center
Yu, Holly, Ed.
2005-01-01
Using database-driven web pages or web content management (WCM) systems to manage increasingly diverse web content and to streamline workflows is a commonly practiced solution recognized in libraries today. However, limited library web content management models and funding constraints prevent many libraries from purchasing commercially available…
Mercury methylation in mine wastes collected from abandoned mercury mines in the USA
Gray, J.E.; Hines, M.E.; Biester, H.; Lasorsa, B.K.; ,
2003-01-01
Speciation and transformation of Hg was studied in mine wastes collected from abandoned Hg mines at McDermitt, Nevada, and Terlingua, Texas, to evaluate formation of methyl-Hg, which is highly toxic. In these mine wastes, we measured total Hg and methyl-Hg contents, identified various Hg compounds using a pyrolysis technique, and determined rates of Hg methylation and methyl-Hg demethylation using isotopic-tracer methods. Mine wastes contain total Hg contents as high as 14000 ??g/g and methyl-Hg concentrations as high as 88 ng/g. Mine wastes were found to contain variable amounts of cinnabar, metacinnabar, Hg salts, Hg0, and Hg0 and Hg2+ sorbed onto matrix particulates. Samples with Hg0 and matrix-sorbed Hg generally contained significant methyl-Hg contents. Similarly, samples containing Hg0 compounds generally produced significant Hg methylation rates, as much as 26%/day. Samples containing mostly cinnabar showed little or no Hg methylation. Mine wastes with high methyl-Hg contents generally showed low methyl-Hg demethylation, suggesting that Hg methylation was dominant. Methyl-Hg demethylation was by both oxidative and microbial pathways. The correspondence of mine wastes containing Hg0 compounds and measured Hg methylation suggests that Hg0 oxidizes to Hg2+, which is subsequently bioavailable for microbial Hg methylation.
Alkemio: association of chemicals with biomedical topics by text and data mining
Gijón-Correas, José A.; Andrade-Navarro, Miguel A.; Fontaine, Jean F.
2014-01-01
The PubMed® database of biomedical citations allows the retrieval of scientific articles studying the function of chemicals in biology and medicine. Mining millions of available citations to search reported associations between chemicals and topics of interest would require substantial human time. We have implemented the Alkemio text mining web tool and SOAP web service to help in this task. The tool uses biomedical articles discussing chemicals (including drugs), predicts their relatedness to the query topic with a naïve Bayesian classifier and ranks all chemicals by P-values computed from random simulations. Benchmarks on seven human pathways showed good retrieval performance (areas under the receiver operating characteristic curves ranged from 73.6 to 94.5%). Comparison with existing tools to retrieve chemicals associated to eight diseases showed the higher precision and recall of Alkemio when considering the top 10 candidate chemicals. Alkemio is a high performing web tool ranking chemicals for any biomedical topics and it is free to non-commercial users. Availability: http://cbdm.mdc-berlin.de/∼medlineranker/cms/alkemio. PMID:24838570
Geovisualization of Local and Regional Migration Using Web-mined Demographics
NASA Astrophysics Data System (ADS)
Schuermann, R. T.; Chow, T. E.
2014-11-01
The intent of this research was to augment and facilitate analyses, which gauges the feasibility of web-mined demographics to study spatio-temporal dynamics of migration. As a case study, we explored the spatio-temporal dynamics of Vietnamese Americans (VA) in Texas through geovisualization of mined demographic microdata from the World Wide Web. Based on string matching across all demographic attributes, including full name, address, date of birth, age and phone number, multiple records of the same entity (i.e. person) over time were resolved and reconciled into a database. Migration trajectories were geovisualized through animated sprites by connecting the different addresses associated with the same person and segmenting the trajectory into small fragments. Intra-metropolitan migration patterns appeared at the local scale within many metropolitan areas. At the scale of metropolitan area, varying degrees of immigration and emigration manifest different types of migration clusters. This paper presents a methodology incorporating GIS methods and cartographic design to produce geovisualization animation, enabling the cognitive identification of migration patterns at multiple scales. Identification of spatio-temporal patterns often stimulates further research to better understand the phenomenon and enhance subsequent modeling.
NASA Astrophysics Data System (ADS)
Pierce, M. E.; Aktas, M. S.; Aydin, G.; Fox, G. C.; Gadgil, H.; Sayar, A.
2005-12-01
We examine the application of Web Service Architectures and Grid-based distributed computing technologies to geophysics and geo-informatics. We are particularly interested in the integration of Geographical Information System (GIS) services with distributed data mining applications. GIS services provide the general purpose framework for building archival data services, real time streaming data services, and map-based visualization services that may be integrated with data mining and other applications through the use of distributed messaging systems and Web Service orchestration tools. Building upon on our previous work in these areas, we present our current research efforts. These include fundamental investigations into increasing XML-based Web service performance, supporting real time data streams, and integrating GIS mapping tools with audio/video collaboration systems for shared display and annotation.
NASA Astrophysics Data System (ADS)
Szlązak, Nikodem; Korzec, Marek
2016-06-01
Methane has a bad influence on safety in underground mines as it is emitted to the air during mining works. Appropriate identification of methane hazard is essential to determining methane hazard prevention methods, ventilation systems and methane drainage systems. Methane hazard is identified while roadways are driven and boreholes are drilled. Coalbed methane content is one of the parameters which is used to assess this threat. This is a requirement according to the Decree of the Minister of Economy dated 28 June 2002 on work safety and hygiene, operation and special firefighting protection in underground mines. For this purpose a new method for determining coalbed methane content in underground coal mines has been developed. This method consists of two stages - collecting samples in a mine and testing the sample in the laboratory. The stage of determining methane content in a coal sample in a laboratory is essential. This article presents the estimation of measurement uncertainty of determining methane content in a coal sample according to this methodology.
Literature Mining Methods for Toxicology and Construction of ...
Webinar Presentation on text-mining methodologies in use at NCCT and how they can be used to assist with the OECD Retinoid project. Presentation to 1st Workshop/Scientific Expert Group meeting on the OECD Retinoid Project - April 26, 2016 –Brussels, Presented remotely via web.
Monitoring food safety violation reports from internet forums.
Kate, Kiran; Negi, Sumit; Kalagnanam, Jayant
2014-01-01
Food-borne illness is a growing public health concern in the world. Government bodies, which regulate and monitor the state of food safety, solicit citizen feedback about food hygiene practices followed by food establishments. They use traditional channels like call center, e-mail for such feedback collection. With the growing popularity of Web 2.0 and social media, citizens often post such feedback on internet forums, message boards etc. The system proposed in this paper applies text mining techniques to identify and mine such food safety complaints posted by citizens on web data sources thereby enabling the government agencies to gather more information about the state of food safety. In this paper, we discuss the architecture of our system and the text mining methods used. We also present results which demonstrate the effectiveness of this system in a real-world deployment.
Huang, Zhenzhen; Duan, Huilong; Li, Haomin
2015-01-01
Large-scale human cancer genomics projects, such as TCGA, generated large genomics data for further study. Exploring and mining these data to obtain meaningful analysis results can help researchers find potential genomics alterations that intervene the development and metastasis of tumors. We developed a web-based gene analysis platform, named TCGA4U, which used statistics methods and models to help translational investigators explore, mine and visualize human cancer genomic characteristic information from the TCGA datasets. Furthermore, through Gene Ontology (GO) annotation and clinical data integration, the genomic data were transformed into biological process, molecular function, cellular component and survival curves to help researchers identify potential driver genes. Clinical researchers without expertise in data analysis will benefit from such a user-friendly genomic analysis platform.
30 CFR 28.31 - Quality control plans; contents.
Code of Federal Regulations, 2012 CFR
2012-07-01
... 28.31 Mineral Resources MINE SAFETY AND HEALTH ADMINISTRATION, DEPARTMENT OF LABOR TESTING... PROTECTION FOR TRAILING CABLES IN COAL MINES Quality Control § 28.31 Quality control plans; contents. (a... Specification MIL-F-15160D is available for examination at the U.S. Department of Labor, Mine Safety and Health...
30 CFR 28.31 - Quality control plans; contents.
Code of Federal Regulations, 2014 CFR
2014-07-01
... 28.31 Mineral Resources MINE SAFETY AND HEALTH ADMINISTRATION, DEPARTMENT OF LABOR TESTING... PROTECTION FOR TRAILING CABLES IN COAL MINES Quality Control § 28.31 Quality control plans; contents. (a... Specification MIL-F-15160D is available for examination at the U.S. Department of Labor, Mine Safety and Health...
30 CFR 28.31 - Quality control plans; contents.
Code of Federal Regulations, 2013 CFR
2013-07-01
... 28.31 Mineral Resources MINE SAFETY AND HEALTH ADMINISTRATION, DEPARTMENT OF LABOR TESTING... PROTECTION FOR TRAILING CABLES IN COAL MINES Quality Control § 28.31 Quality control plans; contents. (a... Specification MIL-F-15160D is available for examination at the U.S. Department of Labor, Mine Safety and Health...
30 CFR 28.31 - Quality control plans; contents.
Code of Federal Regulations, 2011 CFR
2011-07-01
... 28.31 Mineral Resources MINE SAFETY AND HEALTH ADMINISTRATION, DEPARTMENT OF LABOR TESTING... PROTECTION FOR TRAILING CABLES IN COAL MINES Quality Control § 28.31 Quality control plans; contents. (a... Specification MIL-F-15160D is available for examination at the U.S. Department of Labor, Mine Safety and Health...
DOE Office of Scientific and Technical Information (OSTI.GOV)
Kargupta, H.; Stafford, B.; Hamzaoglu, I.
This paper describes an experimental parallel/distributed data mining system PADMA (PArallel Data Mining Agents) that uses software agents for local data accessing and analysis and a web based interface for interactive data visualization. It also presents the results of applying PADMA for detecting patterns in unstructured texts of postmortem reports and laboratory test data for Hepatitis C patients.
Data warehousing as a basis for web-based documentation of data mining and analysis.
Karlsson, J; Eklund, P; Hallgren, C G; Sjödin, J G
1999-01-01
In this paper we present a case study for data warehousing intended to support data mining and analysis. We also describe a prototype for data retrieval. Further we discuss some technical issues related to a particular choice of a patient record environment.
On-Board Mining in the Sensor Web
NASA Astrophysics Data System (ADS)
Tanner, S.; Conover, H.; Graves, S.; Ramachandran, R.; Rushing, J.
2004-12-01
On-board data mining can contribute to many research and engineering applications, including natural hazard detection and prediction, intelligent sensor control, and the generation of customized data products for direct distribution to users. The ability to mine sensor data in real time can also be a critical component of autonomous operations, supporting deep space missions, unmanned aerial and ground-based vehicles (UAVs, UGVs), and a wide range of sensor meshes, webs and grids. On-board processing is expected to play a significant role in the next generation of NASA, Homeland Security, Department of Defense and civilian programs, providing for greater flexibility and versatility in measurements of physical systems. In addition, the use of UAV and UGV systems is increasing in military, emergency response and industrial applications. As research into the autonomy of these vehicles progresses, especially in fleet or web configurations, the applicability of on-board data mining is expected to increase significantly. Data mining in real time on board sensor platforms presents unique challenges. Most notably, the data to be mined is a continuous stream, rather than a fixed store such as a database. This means that the data mining algorithms must be modified to make only a single pass through the data. In addition, the on-board environment requires real time processing with limited computing resources, thus the algorithms must use fixed and relatively small amounts of processing time and memory. The University of Alabama in Huntsville is developing an innovative processing framework for the on-board data and information environment. The Environment for On-Board Processing (EVE) and the Adaptive On-board Data Processing (AODP) projects serve as proofs-of-concept of advanced information systems for remote sensing platforms. The EVE real-time processing infrastructure will upload, schedule and control the execution of processing plans on board remote sensors. These plans provide capabilities for autonomous data mining, classification and feature extraction using both streaming and buffered data sources. A ground-based testbed provides a heterogeneous, embedded hardware and software environment representing both space-based and ground-based sensor platforms, including wireless sensor mesh architectures. The AODP project explores the EVE concepts in the world of sensor-networks, including ad-hoc networks of small sensor platforms.
Designing and Managing Your Digital Library.
ERIC Educational Resources Information Center
Guenther, Kim
2000-01-01
Discusses digital libraries and Web site design issues. Highlights include accessibility issues, including standards, markup languages like HTML and XML, and metadata; building virtual communities; the use of Web portals for customized delivery of information; quality assurance tools, including data mining; and determining user needs, including…
A Dynamic Recommender System for Improved Web Usage Mining and CRM Using Swarm Intelligence.
Alphy, Anna; Prabakaran, S
2015-01-01
In modern days, to enrich e-business, the websites are personalized for each user by understanding their interests and behavior. The main challenges of online usage data are information overload and their dynamic nature. In this paper, to address these issues, a WebBluegillRecom-annealing dynamic recommender system that uses web usage mining techniques in tandem with software agents developed for providing dynamic recommendations to users that can be used for customizing a website is proposed. The proposed WebBluegillRecom-annealing dynamic recommender uses swarm intelligence from the foraging behavior of a bluegill fish. It overcomes the information overload by handling dynamic behaviors of users. Our dynamic recommender system was compared against traditional collaborative filtering systems. The results show that the proposed system has higher precision, coverage, F1 measure, and scalability than the traditional collaborative filtering systems. Moreover, the recommendations given by our system overcome the overspecialization problem by including variety in recommendations.
A Dynamic Recommender System for Improved Web Usage Mining and CRM Using Swarm Intelligence
Alphy, Anna; Prabakaran, S.
2015-01-01
In modern days, to enrich e-business, the websites are personalized for each user by understanding their interests and behavior. The main challenges of online usage data are information overload and their dynamic nature. In this paper, to address these issues, a WebBluegillRecom-annealing dynamic recommender system that uses web usage mining techniques in tandem with software agents developed for providing dynamic recommendations to users that can be used for customizing a website is proposed. The proposed WebBluegillRecom-annealing dynamic recommender uses swarm intelligence from the foraging behavior of a bluegill fish. It overcomes the information overload by handling dynamic behaviors of users. Our dynamic recommender system was compared against traditional collaborative filtering systems. The results show that the proposed system has higher precision, coverage, F1 measure, and scalability than the traditional collaborative filtering systems. Moreover, the recommendations given by our system overcome the overspecialization problem by including variety in recommendations. PMID:26229978
NASA Astrophysics Data System (ADS)
Maleci, Laura; Bini, Claudio; Spiandorello, Massimo; Wahsha, Mohammad
2014-05-01
Heavy metal accumulation produces significant physiological and biochemical responses in vascular plants. Plants growing on abandoned mine sites are of particular interest, since they are genetically tolerant to high metal concentrations. In this work we examined the effect of heavy metals (HM) on the morphology of T. officinale growing on mine soils, with the following objectives: - to determine the fate of HM within the soil-plant system; - to highlight possible damage at anatomical and cytological level; - to assess the resilience capacity of Taraxacum officinale after three years of pot cultivation. Wild specimens of Taraxacum officinale Web, with their soil clod, were gathered from four sites with different contamination levels by heavy metals (Cu, Fe, Pb, Zn) in the abandoned Imperina Valley mine (Northeast Italy). Plants were cultivated in pots at the botanical garden of the University of Florence (HBF), and appeared macroscopically not affected by toxic signals (e.g. reduced growth, leaf necrosis) possibly induced by soil HM concentration. Leaves and roots taken at the same growing season were observed by light microscopy (LM) and transmission electron microscopy (TEM). Light microscopy observations show a clear difference in the cell organization of not-contaminated and contaminated samples. The unpolluted samples present a well organized palisade tissue and spongy photosynthetic parenchyma. Samples from contaminated sites, instead, present a palisade parenchyma less organized, and a reduction of leaf thickness proportional to HM concentration. The poor structural organisations, and the reduced foliar thickness of the contaminated plants, are related to soil contamination. Differences in roots micromorphology concern the cortical parenchyma. Moreover, all the samples examined present mycorrhiza. Ultrastructure observations of the parenchyma cells show mitochondrial structure alteration, with lacking or reduced cristae of the internal membrane at increasing metal content. Instead, chloroplast organization does not present significant differences, particularly in number and compartmentalization of thylakoids. Although macromorphology does not present evidence of phytotoxicity, the recorded observations of the micromorphological characteristics of leaves and roots, show a suffering state of the plants, strictly related to HM content. Leaching reduced partly the HM content of the soil, therefore decreasing their phytotoxic effect. A gradual restoration of leaf organization suggests that somewhat resilience occurred in plants. Moreover, the presence of stress-tolerant mycorrhizal fungi could contribute to reduce metal toxicity. The resilience capacity suggests that Taraxacum could be a useful species in remediation projects. Keywords: Heavy metals • Mine soils • Plant morphology • Taraxacum officinale • Ultrastructure
The Islamic State Battle Plan: Press Release Natural Language Processing
2016-06-01
Processing, text mining , corpus, generalized linear model, cascade, R Shiny, leaflet, data visualization 15. NUMBER OF PAGES 83 16. PRICE CODE...Terrorism and Responses to Terrorism TDM Term Document Matrix TF Term Frequency TF-IDF Term Frequency-Inverse Document Frequency tm text mining (R...package=leaflet. Feinerer I, Hornik K (2015) Text Mining Package “tm,” Version 0.6-2. (Jul 3) https://cran.r-project.org/web/packages/tm/tm.pdf
Numerical linear algebra in data mining
NASA Astrophysics Data System (ADS)
Eldén, Lars
Ideas and algorithms from numerical linear algebra are important in several areas of data mining. We give an overview of linear algebra methods in text mining (information retrieval), pattern recognition (classification of handwritten digits), and PageRank computations for web search engines. The emphasis is on rank reduction as a method of extracting information from a data matrix, low-rank approximation of matrices using the singular value decomposition and clustering, and on eigenvalue methods for network analysis.
A Structural and Content-Based Analysis for Web Filtering.
ERIC Educational Resources Information Center
Lee, P. Y.; Hui, S. C.; Fong, A. C. M.
2003-01-01
Presents an analysis of the distinguishing features of pornographic Web pages so that effective filtering techniques can be developed. Surveys the existing techniques for Web content filtering and describes the implementation of a Web content filtering system that uses an artificial neural network. (Author/LRW)
30 CFR 769.13 - Contents of petitions.
Code of Federal Regulations, 2013 CFR
2013-07-01
... 30 Mineral Resources 3 2013-07-01 2013-07-01 false Contents of petitions. 769.13 Section 769.13 Mineral Resources OFFICE OF SURFACE MINING RECLAMATION AND ENFORCEMENT, DEPARTMENT OF THE INTERIOR AREAS UNSUITABLE FOR MINING PETITION PROCESS FOR DESIGNATION OF FEDERAL LANDS AS UNSUITABLE FOR ALL OR CERTAIN TYPES OF SURFACE COAL MINING OPERATIONS...
Find resources and guidance on writing for the web, keeping your content relevant, using social media, meeting accessibility standards, and how to transform your content into the WebCMS to meet One EPA Web standards.
Koopmans, Bastijn; Smit, August B; Verhage, Matthijs; Loos, Maarten
2017-04-04
Systematic, standardized and in-depth phenotyping and data analyses of rodent behaviour empowers gene-function studies, drug testing and therapy design. However, no data repositories are currently available for standardized quality control, data analysis and mining at the resolution of individual mice. Here, we present AHCODA-DB, a public data repository with standardized quality control and exclusion criteria aimed to enhance robustness of data, enabled with web-based mining tools for the analysis of individually and group-wise collected mouse phenotypic data. AHCODA-DB allows monitoring in vivo effects of compounds collected from conventional behavioural tests and from automated home-cage experiments assessing spontaneous behaviour, anxiety and cognition without human interference. AHCODA-DB includes such data from mutant mice (transgenics, knock-out, knock-in), (recombinant) inbred strains, and compound effects in wildtype mice and disease models. AHCODA-DB provides real time statistical analyses with single mouse resolution and versatile suite of data presentation tools. On March 9th, 2017 AHCODA-DB contained 650 k data points on 2419 parameters from 1563 mice. AHCODA-DB provides users with tools to systematically explore mouse behavioural data, both with positive and negative outcome, published and unpublished, across time and experiments with single mouse resolution. The standardized (automated) experimental settings and the large current dataset (1563 mice) in AHCODA-DB provide a unique framework for the interpretation of behavioural data and drug effects. The use of common ontologies allows data export to other databases such as the Mouse Phenome Database. Unbiased presentation of positive and negative data obtained under the highly standardized screening conditions increase cost efficiency of publicly funded mouse screening projects and help to reach consensus conclusions on drug responses and mouse behavioural phenotypes. The website is publicly accessible through https://public.sylics.com and can be viewed in every recent version of all commonly used browsers.
Utilizing Social Bookmarking Tag Space for Web Content Discovery: A Social Network Analysis Approach
ERIC Educational Resources Information Center
Wei, Wei
2010-01-01
Social bookmarking has gained popularity since the advent of Web 2.0. Keywords known as tags are created to annotate web content, and the resulting tag space composed of the tags, the resources, and the users arises as a new platform for web content discovery. Useful and interesting web resources can be located through searching and browsing based…
Creating and sharing clinical decision support content with Web 2.0: Issues and examples.
Wright, Adam; Bates, David W; Middleton, Blackford; Hongsermeier, Tonya; Kashyap, Vipul; Thomas, Sean M; Sittig, Dean F
2009-04-01
Clinical decision support is a powerful tool for improving healthcare quality and patient safety. However, developing a comprehensive package of decision support interventions is costly and difficult. If used well, Web 2.0 methods may make it easier and less costly to develop decision support. Web 2.0 is characterized by online communities, open sharing, interactivity and collaboration. Although most previous attempts at sharing clinical decision support content have worked outside of the Web 2.0 framework, several initiatives are beginning to use Web 2.0 to share and collaborate on decision support content. We present case studies of three efforts: the Clinfowiki, a world-accessible wiki for developing decision support content; Partners Healthcare eRooms, web-based tools for developing decision support within a single organization; and Epic Systems Corporation's Community Library, a repository for sharing decision support content for customers of a single clinical system vendor. We evaluate the potential of Web 2.0 technologies to enable collaborative development and sharing of clinical decision support systems through the lens of three case studies; analyzing technical, legal and organizational issues for developers, consumers and organizers of clinical decision support content in Web 2.0. We believe the case for Web 2.0 as a tool for collaborating on clinical decision support content appears strong, particularly for collaborative content development within an organization.
Frouz, Jan; Thébault, Elisa; Pižl, Václav; Adl, Sina; Cajthaml, Tomáš; Baldrián, Petr; Háněl, Ladislav; Starý, Josef; Tajovský, Karel; Materna, Jan; Nováková, Alena; de Ruiter, Peter C
2013-01-01
Parameters characterizing the structure of the decomposer food web, biomass of the soil microflora (bacteria and fungi) and soil micro-, meso- and macrofauna were studied at 14 non-reclaimed 1- 41-year-old post-mining sites near the town of Sokolov (Czech Republic). These observations on the decomposer food webs were compared with knowledge of vegetation and soil microstructure development from previous studies. The amount of carbon entering the food web increased with succession age in a similar way as the total amount of C in food web biomass and the number of functional groups in the food web. Connectance did not show any significant changes with succession age, however. In early stages of the succession, the bacterial channel dominated the food web. Later on, in shrub-dominated stands, the fungal channel took over. Even later, in the forest stage, the bacterial channel prevailed again. The best predictor of fungal bacterial ratio is thickness of fermentation layer. We argue that these changes correspond with changes in topsoil microstructure driven by a combination of plant organic matter input and engineering effects of earthworms. In early stages, soil is alkaline, and a discontinuous litter layer on the soil surface promotes bacterial biomass growth, so the bacterial food web channel can dominate. Litter accumulation on the soil surface supports the development of the fungal channel. In older stages, earthworms arrive, mix litter into the mineral soil and form an organo-mineral topsoil, which is beneficial for bacteria and enhances the bacterial food web channel.
Frouz, Jan; Thébault, Elisa; Pižl, Václav; Adl, Sina; Cajthaml, Tomáš; Baldrián, Petr; Háněl, Ladislav; Starý, Josef; Tajovský, Karel; Materna, Jan; Nováková, Alena; de Ruiter, Peter C.
2013-01-01
Parameters characterizing the structure of the decomposer food web, biomass of the soil microflora (bacteria and fungi) and soil micro-, meso- and macrofauna were studied at 14 non-reclaimed 1– 41-year-old post-mining sites near the town of Sokolov (Czech Republic). These observations on the decomposer food webs were compared with knowledge of vegetation and soil microstructure development from previous studies. The amount of carbon entering the food web increased with succession age in a similar way as the total amount of C in food web biomass and the number of functional groups in the food web. Connectance did not show any significant changes with succession age, however. In early stages of the succession, the bacterial channel dominated the food web. Later on, in shrub-dominated stands, the fungal channel took over. Even later, in the forest stage, the bacterial channel prevailed again. The best predictor of fungal bacterial ratio is thickness of fermentation layer. We argue that these changes correspond with changes in topsoil microstructure driven by a combination of plant organic matter input and engineering effects of earthworms. In early stages, soil is alkaline, and a discontinuous litter layer on the soil surface promotes bacterial biomass growth, so the bacterial food web channel can dominate. Litter accumulation on the soil surface supports the development of the fungal channel. In older stages, earthworms arrive, mix litter into the mineral soil and form an organo-mineral topsoil, which is beneficial for bacteria and enhances the bacterial food web channel. PMID:24260281
Does content affect whether users remember that Web pages were hyperlinked?
Jones, Keith S; Ballew, Timothy V; Probst, C Adam
2008-10-01
We determined whether memory for hyperlinks improved when they represented relations between the contents of the Web pages. J. S. Farris (2003) found that memory for hyperlinks improved when they represented relations between the contents of the Web pages. However, Farris's (2003) participants could have used their knowledge of site content to answer questions about relations that were instantiated via the site's content and its hyperlinks. In Experiment 1, users navigated a Web site and then answered questions about relations that were instantiated only via content, only via hyperlinks, and via content and hyperlinks. Unlike Farris (2003), we split the latter into two sets. One asked whether certain content elements were related, and the other asked whether certain Web pages were hyperlinked. Experiment 2 replicated Experiment 1 with one modification: The questions that were asked about relations instantiated via content and hyperlinks were changed so that each question's wrong answer was also related to the question's target. Memory for hyperlinks improved when they represented relations instantiated within the content of the Web pages. This was true when (a) questions about content and hyperlinks were separated (Experiment 1) and (b) each question's wrong answer was also related to the question's target (Experiment 2). The accuracy of users' mental representations of local architecture depended on whether hyperlinks were related to the site's content. Designers who want users to remember hyperlinks should associate those hyperlinks with content that reflects the relation between the contents on the Web pages.
Alkemio: association of chemicals with biomedical topics by text and data mining.
Gijón-Correas, José A; Andrade-Navarro, Miguel A; Fontaine, Jean F
2014-07-01
The PubMed® database of biomedical citations allows the retrieval of scientific articles studying the function of chemicals in biology and medicine. Mining millions of available citations to search reported associations between chemicals and topics of interest would require substantial human time. We have implemented the Alkemio text mining web tool and SOAP web service to help in this task. The tool uses biomedical articles discussing chemicals (including drugs), predicts their relatedness to the query topic with a naïve Bayesian classifier and ranks all chemicals by P-values computed from random simulations. Benchmarks on seven human pathways showed good retrieval performance (areas under the receiver operating characteristic curves ranged from 73.6 to 94.5%). Comparison with existing tools to retrieve chemicals associated to eight diseases showed the higher precision and recall of Alkemio when considering the top 10 candidate chemicals. Alkemio is a high performing web tool ranking chemicals for any biomedical topics and it is free to non-commercial users. http://cbdm.mdc-berlin.de/∼medlineranker/cms/alkemio. © The Author(s) 2014. Published by Oxford University Press on behalf of Nucleic Acids Research.
EAGLE: 'EAGLE'Is an' Algorithmic Graph Library for Exploration
DOE Office of Scientific and Technical Information (OSTI.GOV)
2015-01-16
The Resource Description Framework (RDF) and SPARQL Protocol and RDF Query Language (SPARQL) were introduced about a decade ago to enable flexible schema-free data interchange on the Semantic Web. Today data scientists use the framework as a scalable graph representation for integrating, querying, exploring and analyzing data sets hosted at different sources. With increasing adoption, the need for graph mining capabilities for the Semantic Web has emerged. Today there is no tools to conduct "graph mining" on RDF standard data sets. We address that need through implementation of popular iterative Graph Mining algorithms (Triangle count, Connected component analysis, degree distribution,more » diversity degree, PageRank, etc.). We implement these algorithms as SPARQL queries, wrapped within Python scripts and call our software tool as EAGLE. In RDF style, EAGLE stands for "EAGLE 'Is an' algorithmic graph library for exploration. EAGLE is like 'MATLAB' for 'Linked Data.'« less
Using Syntactic Patterns to Enhance Text Analytics
ERIC Educational Resources Information Center
Meyer, Bradley B.
2017-01-01
Large scale product and service reviews proliferate and are commonly found across the web. The ability to harvest, digest and analyze a large corpus of reviews from online websites is still however a difficult problem. This problem is referred to as "opinion mining." Opinion mining is an important area of research as advances in the…
Web Usage Mining: Application to an Online Educational Digital Library Service
ERIC Educational Resources Information Center
Palmer, Bart C.
2012-01-01
This dissertation was situated in the crossroads of educational data mining (EDM), educational digital libraries (such as the National Science Digital Library; http://nsdl.org), and examination of teacher behaviors while creating online learning resources in an end-user authoring system, the Instructional Architect (IA; http://ia.usu.edu). The…
30 CFR 74.16 - Material required for record.
Code of Federal Regulations, 2010 CFR
2010-07-01
... 30 Mineral Resources 1 2010-07-01 2010-07-01 false Material required for record. 74.16 Section 74.16 Mineral Resources MINE SAFETY AND HEALTH ADMINISTRATION, DEPARTMENT OF LABOR COAL MINE SAFETY AND... deliver a complete sampling device free of charge to NIOSH at the address specified on the NIOSH Web page...
Simple, Scalable, Script-based, Science Processor for Measurements - Data Mining Edition (S4PM-DME)
NASA Astrophysics Data System (ADS)
Pham, L. B.; Eng, E. K.; Lynnes, C. S.; Berrick, S. W.; Vollmer, B. E.
2005-12-01
The S4PM-DME is the Goddard Earth Sciences Distributed Active Archive Center's (GES DAAC) web-based data mining environment. The S4PM-DME replaces the Near-line Archive Data Mining (NADM) system with a better web environment and a richer set of production rules. S4PM-DME enables registered users to submit and execute custom data mining algorithms. The S4PM-DME system uses the GES DAAC developed Simple Scalable Script-based Science Processor for Measurements (S4PM) to automate tasks and perform the actual data processing. A web interface allows the user to access the S4PM-DME system. The user first develops personalized data mining algorithm on his/her home platform and then uploads them to the S4PM-DME system. Algorithms in C and FORTRAN languages are currently supported. The user developed algorithm is automatically audited for any potential security problems before it is installed within the S4PM-DME system and made available to the user. Once the algorithm has been installed the user can promote the algorithm to the "operational" environment. From here the user can search and order the data available in the GES DAAC archive for his/her science algorithm. The user can also set up a processing subscription. The subscription will automatically process new data as it becomes available in the GES DAAC archive. The generated mined data products are then made available for FTP pickup. The benefits of using S4PM-DME are 1) to decrease the downloading time it typically takes a user to transfer the GES DAAC data to his/her system thus off-load the heavy network traffic, 2) to free-up the load on their system, and last 3) to utilize the rich and abundance ocean, atmosphere data from the MODIS and AIRS instruments available from the GES DAAC.
GhostWriter-2.0: Product Reviews with Case-Based Support
NASA Astrophysics Data System (ADS)
Bridge, Derek; Healy, Paul
A lot of user-generated content on the Web takes the form of records of personal experiences. Case-Based Reasoning offers a way of helping one user to reuse another's experiences from the Web. In this paper, we present GhostWriter-2.0, a Case-Based Reasoning system that supports a user who is writing a product review. GhostWriter-2.0 makes suggestions to the user, in the form of short phrases that are mined from other reviews. The purpose of the suggestions is to prompt the user to write a more comprehensive and helpful review than she might otherwise have done. We explain how GhostWriter-2.0's case base is populated with relevant and helpful reviews from Amazon. We show how it extracts and scores phrases in these reviews to decide which to suggest to the user.We report a trial with real users, in which users made greater use of GhostWriter-2.0's suggested phrases than they did of phrases suggested by a system that used a more random form of selection.
30 CFR 777.11 - Format and contents.
Code of Federal Regulations, 2013 CFR
2013-07-01
... 30 Mineral Resources 3 2013-07-01 2013-07-01 false Format and contents. 777.11 Section 777.11 Mineral Resources OFFICE OF SURFACE MINING RECLAMATION AND ENFORCEMENT, DEPARTMENT OF THE INTERIOR SURFACE COAL MINING AND RECLAMATION OPERATIONS PERMITS AND COAL EXPLORATION SYSTEMS UNDER REGULATORY PROGRAMS GENERAL CONTENT REQUIREMENTS FOR PERMIT...
Code of Federal Regulations, 2010 CFR
2010-04-01
... 20 Employees' Benefits 3 2010-04-01 2010-04-01 false Contents. 726.315 Section 726.315 Employees' Benefits EMPLOYMENT STANDARDS ADMINISTRATION, DEPARTMENT OF LABOR FEDERAL COAL MINE HEALTH AND SAFETY ACT OF 1969, AS AMENDED BLACK LUNG BENEFITS; REQUIREMENTS FOR COAL MINE OPERATOR'S INSURANCE Civil Money Penalties § 726.315 Contents. Any petition...
30 CFR 77.201 - Methane content in surface installations.
Code of Federal Regulations, 2010 CFR
2010-07-01
... 30 Mineral Resources 1 2010-07-01 2010-07-01 false Methane content in surface installations. 77.201 Section 77.201 Mineral Resources MINE SAFETY AND HEALTH ADMINISTRATION, DEPARTMENT OF LABOR COAL... UNDERGROUND COAL MINES Surface Installations § 77.201 Methane content in surface installations. The methane...
30 CFR 769.13 - Contents of petitions.
Code of Federal Regulations, 2011 CFR
2011-07-01
... 30 Mineral Resources 3 2011-07-01 2011-07-01 false Contents of petitions. 769.13 Section 769.13 Mineral Resources OFFICE OF SURFACE MINING RECLAMATION AND ENFORCEMENT, DEPARTMENT OF THE INTERIOR AREAS... TYPES OF SURFACE COAL MINING OPERATIONS AND FOR TERMINATION OF PREVIOUS DESIGNATIONS § 769.13 Contents...
30 CFR 769.13 - Contents of petitions.
Code of Federal Regulations, 2010 CFR
2010-07-01
... 30 Mineral Resources 3 2010-07-01 2010-07-01 false Contents of petitions. 769.13 Section 769.13 Mineral Resources OFFICE OF SURFACE MINING RECLAMATION AND ENFORCEMENT, DEPARTMENT OF THE INTERIOR AREAS... TYPES OF SURFACE COAL MINING OPERATIONS AND FOR TERMINATION OF PREVIOUS DESIGNATIONS § 769.13 Contents...
30 CFR 77.201 - Methane content in surface installations.
Code of Federal Regulations, 2011 CFR
2011-07-01
... 30 Mineral Resources 1 2011-07-01 2011-07-01 false Methane content in surface installations. 77.201 Section 77.201 Mineral Resources MINE SAFETY AND HEALTH ADMINISTRATION, DEPARTMENT OF LABOR COAL... UNDERGROUND COAL MINES Surface Installations § 77.201 Methane content in surface installations. The methane...
ESTminer: a Web interface for mining EST contig and cluster databases.
Huang, Yecheng; Pumphrey, Janie; Gingle, Alan R
2005-03-01
ESTminer is a Web application and database schema for interactive mining of expressed sequence tag (EST) contig and cluster datasets. The Web interface contains a query frame that allows the selection of contigs/clusters with specific cDNA library makeup or a threshold number of members. The results are displayed as color-coded tree nodes, where the color indicates the fractional size of each cDNA library component. The nodes are expandable, revealing library statistics as well as EST or contig members, with links to sequence data, GenBank records or user configurable links. Also, the interface allows 'queries within queries' where the result set of a query is further filtered by the subsequent query. ESTminer is implemented in Java/JSP and the package, including MySQL and Oracle schema creation scripts, is available from http://cggc.agtec.uga.edu/Data/download.asp agingle@uga.edu.
Socio-contextual Network Mining for User Assistance in Web-based Knowledge Gathering Tasks
NASA Astrophysics Data System (ADS)
Rajendran, Balaji; Kombiah, Iyakutti
Web-based Knowledge Gathering (WKG) is a specialized and complex information seeking task carried out by many users on the web, for their various learning, and decision-making requirements. We construct a contextual semantic structure by observing the actions of the users involved in WKG task, in order to gain an understanding of their task and requirement. We also build a knowledge warehouse in the form of a master Semantic Link Network (SLX) that accommodates and assimilates all the contextual semantic structures. This master SLX, which is a socio-contextual network, is then mined to provide contextual inputs to the current users through their agents. We validated our approach through experiments and analyzed the benefits to the users in terms of resource explorations and the time saved. The results are positive enough to motivate us to implement in a larger scale.
Mercury and methylmercury contamination related to artisanal gold mining, Suriname
Gray, J.E.; Labson, V.F.; Weaver, J.N.; Krabbenhoft, D.P.
2002-01-01
Elemental Hg-Au amalgamation mining practices are used widely in many developing countries resulting in significant Hg contamination of surrounding ecosystems. We have measured total Hg and methyl-Hg concentrations in sediment and water collected from artisanal Au mines and these are the first Hg speciation data from such mines in Suriname. Total Hg and methyl-Hg contents in mine-waste sediment and water are elevated over local uncontaminated baselines. Total Hg (10-930 ng/L) and methyl-Hg (0.02-3.8 ng/L) are highly elevated in mine waters. Increasing total Hg contents in discharged mine waters correlate with increasing water turbidity indicating that most Hg transport is on suspended particulates. Our Hg results are similar to those found in artisanal Au mines in the Amazon basin, where Hg contamination has led to adverse effects on tropical ecosystems.
NASA Astrophysics Data System (ADS)
Jandova, Katerina; Hyodo, Fujio; Vindušková, Olga; Moradi, Jabbar; Frouz, Jan
2017-04-01
Sediments rich in kerogen ( 19 Ma old, 14C-free) are present in the overburden at post-mining area in Western Bohemia, near Sokolov city, the Czech Republic. There are two successional chronosequences, an alder reclamation and spontaneous succession, consisting of sites that differ in time since heaping. Both chronosequences accumulate recent organic matter over time, although the process is initially faster at reclamation. We hypothesized that (i) radiocarbon age of soil organic matter would be decreasing with time since spoil heaping; (ii) the detrital food web would show the assimilation of fossil carbon by heterotrophic organisms in the initial stages of succession when fossil organic matter is the predominant source of carbon; (iii) the isotopic track of fossil organic matter in the detrital food web would be more prominent at sites with lower vegetation cover and litter production. Nitrogen isotopic ratios of soils were high at the young sites and the decrease in δ15N was correlated with the increase in content of recent organic carbon. Nitrogen isotopic ratios of soil detritivores equalled to that of tree leaves at reclamation but were higher at successional sites. Possibly, other food sources were used apart from tree leaves litter at the latter. Interestingly, soil animals but not primary producers were 14C depleted in the youngest relative to the oldest sites. The depletion in 14C of detritivores relative to primary producers was likely due to the geophagy behaviour of the millipedes at the young sites where fossil organic matter is the largest carbon pool.
Near-line Archive Data Mining at the Goddard Distributed Active Archive Center
NASA Astrophysics Data System (ADS)
Pham, L.; Mack, R.; Eng, E.; Lynnes, C.
2002-12-01
NASA's Earth Observing System (EOS) is generating immense volumes of data, in some cases too much to provide to users with data-intensive needs. As an alternative to moving the data to the user and his/her research algorithms, we are providing a means to move the algorithms to the data. The Near-line Archive Data Mining (NADM) system is the Goddard Earth Sciences Distributed Active Archive Center's (GES DAAC) web data mining portal to the EOS Data and Information System (EOSDIS) data pool, a 50-TB online disk cache. The NADM web portal enables registered users to submit and execute data mining algorithm codes on the data in the EOSDIS data pool. A web interface allows the user to access the NADM system. The users first develops personalized data mining code on their home platform and then uploads them to the NADM system. The C, FORTRAN and IDL languages are currently supported. The user developed code is automatically audited for any potential security problems before it is installed within the NADM system and made available to the user. Once the code has been installed the user is provided a test environment where he/she can test the execution of the software against data sets of the user's choosing. When the user is satisfied with the results, he/she can promote their code to the "operational" environment. From here the user can interactively run his/her code on the data available in the EOSDIS data pool. The user can also set up a processing subscription. The subscription will automatically process new data as it becomes available in the EOSDIS data pool. The generated mined data products are then made available for FTP pickup. The NADM system uses the GES DAAC-developed Simple Scalable Script-based Science Processor (S4P) to automate tasks and perform the actual data processing. Users will also have the option of selecting a DAAC-provided data mining algorithm and using it to process the data of their choice.
Exploring context and content links in social media: a latent space method.
Qi, Guo-Jun; Aggarwal, Charu; Tian, Qi; Ji, Heng; Huang, Thomas S
2012-05-01
Social media networks contain both content and context-specific information. Most existing methods work with either of the two for the purpose of multimedia mining and retrieval. In reality, both content and context information are rich sources of information for mining, and the full power of mining and processing algorithms can be realized only with the use of a combination of the two. This paper proposes a new algorithm which mines both context and content links in social media networks to discover the underlying latent semantic space. This mapping of the multimedia objects into latent feature vectors enables the use of any off-the-shelf multimedia retrieval algorithms. Compared to the state-of-the-art latent methods in multimedia analysis, this algorithm effectively solves the problem of sparse context links by mining the geometric structure underlying the content links between multimedia objects. Specifically for multimedia annotation, we show that an effective algorithm can be developed to directly construct annotation models by simultaneously leveraging both context and content information based on latent structure between correlated semantic concepts. We conduct experiments on the Flickr data set, which contains user tags linked with images. We illustrate the advantages of our approach over the state-of-the-art multimedia retrieval techniques.
ERIC Educational Resources Information Center
Yesiltas, Erkan
2016-01-01
Web pedagogical content knowledge generally takes pedagogical knowledge, content knowledge, and Web knowledge as basis. It is a structure emerging through the interaction of these three components. Content knowledge refers to knowledge of subjects to be taught. Pedagogical knowledge involves knowledge of process, implementation, learning methods,…
Not Your Father's Web Site: Corporate Sites Emerge as New Content Innovators.
ERIC Educational Resources Information Center
O'Leary, Mick
2002-01-01
New economy corporate Web sites have pioneered exciting techniques-rich media, interactivity, personalization, community, and integration of much third-party content. Discusses business-to-business (B2B) Web commerce, with examples of several B2B corporate sites; portal and content elements of these sites; and corporate content outlooks. (AEF)
Clustering and Dimensionality Reduction to Discover Interesting Patterns in Binary Data
NASA Astrophysics Data System (ADS)
Palumbo, Francesco; D'Enza, Alfonso Iodice
The attention towards binary data coding increased consistently in the last decade due to several reasons. The analysis of binary data characterizes several fields of application, such as market basket analysis, DNA microarray data, image mining, text mining and web-clickstream mining. The paper illustrates two different approaches exploiting a profitable combination of clustering and dimensionality reduction for the identification of non-trivial association structures in binary data. An application in the Association Rules framework supports the theory with the empirical evidence.
Gray, J.E.; Greaves, I.A.; Bustos, D.M.; Krabbenhoft, D.P.
2003-01-01
The Palawan Quicksilver mine, Philippines, produced about 2,900 t of mercury during mining of cinnabar ore from 1953 to 1976. More than 2,000,000 t of mine-waste calcines (retorted ore) were produced during mining, much of which were used to construct a jetty in nearby Honda Bay. Since 1995, high Hg contents have been found in several people living near the mine, and 21 of these people were treated for mercury poisoning. Samples of mine-waste calcine contain high total Hg concentrations ranging from 43-660 ??g/g, whereas total Hg concentrations in sediment samples collected from a mine pit lake and local stream vary from 3.7-400 ??g/g. Mine water flowing through the calcines is acidic, pH 3.1-4.3, and total Hg concentrations ranging from 18-31 ??g/l in this water significantly exceed the 1.0-??g/l drinking water standard for Hg recommended by the World Health Organization (WHO). Total Hg contents are generally lower in water samples collected from surrounding domestic wells, the mine pit lake, Honda Bay, and the nearby stream, varying from 0.008-1.4 ??g/l. Methylmercury concentrations in water draining mine calcines range from <0.02-1.4 ng/l, but methylmercury is highest in the pit lake water, ranging from 1.7-3.1 ng/l. Mercury methylation at the Palawan mine is similar to or higher than that found in other mercury mines worldwide. Much of the methylmercury generated in Palawan mine-waste calcines and those in Honda Bay is transferred to water, and then to marine fish and seafood. A food source pathway of Hg to humans is most likely in this coastal, high fish-consuming population.
Geotechnical approaches to coal ash content control in mining of complex structure deposits
NASA Astrophysics Data System (ADS)
Batugin, SA; Gavrilov, VL; Khoyutanov, EA
2017-02-01
Coal deposits having complex structure and nonuniform quality coal reserves require improved processes of production quality control. The paper proposes a method to present coal ash content as components of natural and technological dilution. It is chosen to carry out studies on the western site of Elginsk coal deposit, composed of four coal beds of complex structure. The reported estimates of coal ash content in the beds with respect to five components point at the need to account for such data in confirmation exploration, mine planning and actual mining. Basic means of analysis and control of overall ash content and its components are discussed.
TREC Microblog 2012 Track: Real-Time Algorithm for Microblog Ranking Systems
2012-11-01
such as information about the tweet and the user profile. We collected those tweets by means of web crawler and extract several features from the raw...Mining Text Data. 2012. [5] D. Feltoni. Twittersa: un sistema per l’analisi del sentimento nelle reti sociali. Master’s thesis, Roma Tre University...Morris. Twittersearch: a comparison of microblog search and web search. Proceedings of the fourth ACM international conference on Web search, 2011
Policy-Aware Content Reuse on the Web
NASA Astrophysics Data System (ADS)
Seneviratne, Oshani; Kagal, Lalana; Berners-Lee, Tim
The Web allows users to share their work very effectively leading to the rapid re-use and remixing of content on the Web including text, images, and videos. Scientific research data, social networks, blogs, photo sharing sites and other such applications known collectively as the Social Web have lots of increasingly complex information. Such information from several Web pages can be very easily aggregated, mashed up and presented in other Web pages. Content generation of this nature inevitably leads to many copyright and license violations, motivating research into effective methods to detect and prevent such violations.
ERIC Educational Resources Information Center
Tsai, Yea-Ru; Ouyang, Chen-Sen; Chang, Yukon
2016-01-01
The purpose of this study is to propose a diagnostic approach to identify engineering students' English reading comprehension errors. Student data were collected during the process of reading texts of English for science and technology on a web-based cumulative sentence analysis system. For the analysis, the association-rule, data mining technique…
Application of Learning Analytics Using Clustering Data Mining for Students' Disposition Analysis
ERIC Educational Resources Information Center
Bharara, Sanyam; Sabitha, Sai; Bansal, Abhay
2018-01-01
Learning Analytics (LA) is an emerging field in which sophisticated analytic tools are used to improve learning and education. It draws from, and is closely tied to, a series of other fields of study like business intelligence, web analytics, academic analytics, educational data mining, and action analytics. The main objective of this research…
Rare disease diagnosis: A review of web search, social media and large-scale data-mining approaches.
Svenstrup, Dan; Jørgensen, Henrik L; Winther, Ole
2015-01-01
Physicians and the general public are increasingly using web-based tools to find answers to medical questions. The field of rare diseases is especially challenging and important as shown by the long delay and many mistakes associated with diagnoses. In this paper we review recent initiatives on the use of web search, social media and data mining in data repositories for medical diagnosis. We compare the retrieval accuracy on 56 rare disease cases with known diagnosis for the web search tools google.com, pubmed.gov, omim.org and our own search tool findzebra.com. We give a detailed description of IBM's Watson system and make a rough comparison between findzebra.com and Watson on subsets of the Doctor's dilemma dataset. The recall@10 and recall@20 (fraction of cases where the correct result appears in top 10 and top 20) for the 56 cases are found to be be 29%, 16%, 27% and 59% and 32%, 18%, 34% and 64%, respectively. Thus, FindZebra has a significantly (p < 0.01) higher recall than the other 3 search engines. When tested under the same conditions, Watson and FindZebra showed similar recall@10 accuracy. However, the tests were performed on different subsets of Doctors dilemma questions. Advances in technology and access to high quality data have opened new possibilities for aiding the diagnostic process. Specialized search engines, data mining tools and social media are some of the areas that hold promise.
Rare disease diagnosis: A review of web search, social media and large-scale data-mining approaches
Svenstrup, Dan; Jørgensen, Henrik L; Winther, Ole
2015-01-01
Physicians and the general public are increasingly using web-based tools to find answers to medical questions. The field of rare diseases is especially challenging and important as shown by the long delay and many mistakes associated with diagnoses. In this paper we review recent initiatives on the use of web search, social media and data mining in data repositories for medical diagnosis. We compare the retrieval accuracy on 56 rare disease cases with known diagnosis for the web search tools google.com, pubmed.gov, omim.org and our own search tool findzebra.com. We give a detailed description of IBM's Watson system and make a rough comparison between findzebra.com and Watson on subsets of the Doctor's dilemma dataset. The recall@10 and recall@20 (fraction of cases where the correct result appears in top 10 and top 20) for the 56 cases are found to be be 29%, 16%, 27% and 59% and 32%, 18%, 34% and 64%, respectively. Thus, FindZebra has a significantly (p < 0.01) higher recall than the other 3 search engines. When tested under the same conditions, Watson and FindZebra showed similar recall@10 accuracy. However, the tests were performed on different subsets of Doctors dilemma questions. Advances in technology and access to high quality data have opened new possibilities for aiding the diagnostic process. Specialized search engines, data mining tools and social media are some of the areas that hold promise. PMID:26442199
An Investigation into Web Content Accessibility Guideline Conformance for an Aging Population
ERIC Educational Resources Information Center
Curran, Kevin; Robinson, David
2007-01-01
Poor web site design can cause difficulties for specific groups of users. By applying the Web Content Accessibility Guidelines to a web site, the amount of possible users who can successfully view the content of that site will increase, especially for those who are in the disabled and older adult categories of online users. Older adults are coming…
USDA-ARS?s Scientific Manuscript database
The ARS Microbial Genome Sequence Database (http://199.133.98.43), a web-based database server, was established utilizing the BIGSdb (Bacterial Isolate Genomics Sequence Database) software package, developed at Oxford University, as a tool to manage multi-locus sequence data for the family Streptomy...
Online Persistence in Higher Education Web-Supported Courses
ERIC Educational Resources Information Center
Hershkovitz, Arnon; Nachmias, Rafi
2011-01-01
This research consists of an empirical study of online persistence in Web-supported courses in higher education, using Data Mining techniques. Log files of 58 Moodle websites accompanying Tel Aviv University courses were drawn, recording the activity of 1189 students in 1897 course enrollments during the academic year 2008/9, and were analyzed…
Mining the Human Phenome using Semantic Web Technologies: A Case Study for Type 2 Diabetes
Pathak, Jyotishman; Kiefer, Richard C.; Bielinski, Suzette J.; Chute, Christopher G.
2012-01-01
The ability to conduct genome-wide association studies (GWAS) has enabled new exploration of how genetic variations contribute to health and disease etiology. However, historically GWAS have been limited by inadequate sample size due to associated costs for genotyping and phenotyping of study subjects. This has prompted several academic medical centers to form “biobanks” where biospecimens linked to personal health information, typically in electronic health records (EHRs), are collected and stored on large number of subjects. This provides tremendous opportunities to discover novel genotype-phenotype associations and foster hypothesis generation. In this work, we study how emerging Semantic Web technologies can be applied in conjunction with clinical and genotype data stored at the Mayo Clinic Biobank to mine the phenotype data for genetic associations. In particular, we demonstrate the role of using Resource Description Framework (RDF) for representing EHR diagnoses and procedure data, and enable federated querying via standardized Web protocols to identify subjects genotyped with Type 2 Diabetes for discovering gene-disease associations. Our study highlights the potential of Web-scale data federation techniques to execute complex queries. PMID:23304343
Mining the human phenome using semantic web technologies: a case study for Type 2 Diabetes.
Pathak, Jyotishman; Kiefer, Richard C; Bielinski, Suzette J; Chute, Christopher G
2012-01-01
The ability to conduct genome-wide association studies (GWAS) has enabled new exploration of how genetic variations contribute to health and disease etiology. However, historically GWAS have been limited by inadequate sample size due to associated costs for genotyping and phenotyping of study subjects. This has prompted several academic medical centers to form "biobanks" where biospecimens linked to personal health information, typically in electronic health records (EHRs), are collected and stored on large number of subjects. This provides tremendous opportunities to discover novel genotype-phenotype associations and foster hypothesis generation. In this work, we study how emerging Semantic Web technologies can be applied in conjunction with clinical and genotype data stored at the Mayo Clinic Biobank to mine the phenotype data for genetic associations. In particular, we demonstrate the role of using Resource Description Framework (RDF) for representing EHR diagnoses and procedure data, and enable federated querying via standardized Web protocols to identify subjects genotyped with Type 2 Diabetes for discovering gene-disease associations. Our study highlights the potential of Web-scale data federation techniques to execute complex queries.
"Ordinary People Do This": Rhetorical Examinations of Novice Web Design
ERIC Educational Resources Information Center
Karper, Erin
2005-01-01
Even as weblogs, content management systems, and other forms of automated Web posting and journals are changing the way people create and place content on the Web, new Web pages mushroom overnight. However, many new Web designers produce Web pages that seem to ignore fundamental principles of "good design": full of colored backgrounds, animated…
Integration of Text- and Data-Mining Technologies for Use in Banking Applications
NASA Astrophysics Data System (ADS)
Maslankowski, Jacek
Unstructured data, most of it in the form of text files, typically accounts for 85% of an organization's knowledge stores, but it's not always easy to find, access, analyze or use (Robb 2004). That is why it is important to use solutions based on text and data mining. This solution is known as duo mining. This leads to improve management based on knowledge owned in organization. The results are interesting. Data mining provides to lead with structuralized data, usually powered from data warehouses. Text mining, sometimes called web mining, looks for patterns in unstructured data — memos, document and www. Integrating text-based information with structured data enriches predictive modeling capabilities and provides new stores of insightful and valuable information for driving business and research initiatives forward.
Code of Federal Regulations, 2012 CFR
2012-04-01
... 20 Employees' Benefits 4 2012-04-01 2012-04-01 false Contents. 726.315 Section 726.315 Employees' Benefits OFFICE OF WORKERS' COMPENSATION PROGRAMS, DEPARTMENT OF LABOR FEDERAL COAL MINE HEALTH AND SAFETY ACT OF 1969, AS AMENDED BLACK LUNG BENEFITS; REQUIREMENTS FOR COAL MINE OPERATOR'S INSURANCE Civil Money Penalties § 726.315 Contents. Any...
Code of Federal Regulations, 2014 CFR
2014-04-01
... 20 Employees' Benefits 4 2014-04-01 2014-04-01 false Contents. 726.315 Section 726.315 Employees' Benefits OFFICE OF WORKERS' COMPENSATION PROGRAMS, DEPARTMENT OF LABOR FEDERAL COAL MINE HEALTH AND SAFETY ACT OF 1969, AS AMENDED BLACK LUNG BENEFITS; REQUIREMENTS FOR COAL MINE OPERATOR'S INSURANCE Civil Money Penalties § 726.315 Contents. Any...
Code of Federal Regulations, 2013 CFR
2013-04-01
... 20 Employees' Benefits 4 2013-04-01 2013-04-01 false Contents. 726.315 Section 726.315 Employees' Benefits OFFICE OF WORKERS' COMPENSATION PROGRAMS, DEPARTMENT OF LABOR FEDERAL COAL MINE HEALTH AND SAFETY ACT OF 1969, AS AMENDED BLACK LUNG BENEFITS; REQUIREMENTS FOR COAL MINE OPERATOR'S INSURANCE Civil Money Penalties § 726.315 Contents. Any...
Changes in host-parasitoid food web structure with elevation.
Maunsell, Sarah C; Kitching, Roger L; Burwell, Chris J; Morris, Rebecca J
2015-03-01
Gradients in elevation are increasingly used to investigate how species respond to changes in local climatic conditions. Whilst many studies have shown elevational patterns in species richness and turnover, little is known about how food web structure is affected by elevation. Contrasting responses of predator and prey species to elevation may lead to changes in food web structure. We investigated how the quantitative structure of a herbivore-parasitoid food web changes with elevation in an Australian subtropical rain forest. On four occasions, spread over 1 year, we hand-collected leaf miners at twelve sites, along three elevational gradients (between 493 m and 1159 m a.s.l). A total of 5030 insects, including 603 parasitoids, were reared, and summary food webs were created for each site. We also carried out a replicated manipulative experiment by translocating an abundant leaf-mining weevil Platynotocis sp., which largely escaped parasitism at high elevations (≥ 900 m a.s.l.), to lower, warmer elevations, to test if it would experience higher parasitism pressure. We found strong evidence that the environmental change that occurs with increasing elevation affects food web structure. Quantitative measures of generality, vulnerability and interaction evenness decreased significantly with increasing elevation (and decreasing temperature), whilst elevation did not have a significant effect on connectance. Mined plant composition also had a significant effect on generality and vulnerability, but not on interaction evenness. Several relatively abundant species of leaf miner appeared to escape parasitism at higher elevations, but contrary to our prediction, Platynotocis sp. did not experience greater levels of parasitism when translocated to lower elevations. Our study indicates that leaf-mining herbivores and their parasitoids respond differently to environmental conditions imposed by elevation, thus producing structural changes in their food webs. Increasing temperatures and changes in vegetation communities that are likely to result from climate change may have a restructuring effect on host-parasitoid food webs. Our translocation experiment, however, indicated that leaf miners currently escaping parasitism at high elevations may not automatically experience higher parasitism under warmer conditions and future changes in food web structure may depend on the ability of parasitoids to adapt to novel hosts. © 2014 The Authors. Journal of Animal Ecology © 2014 British Ecological Society.
An active visual search interface for Medline.
Xuan, Weijian; Dai, Manhong; Mirel, Barbara; Wilson, Justin; Athey, Brian; Watson, Stanley J; Meng, Fan
2007-01-01
Searching the Medline database is almost a daily necessity for many biomedical researchers. However, available Medline search solutions are mainly designed for the quick retrieval of a small set of most relevant documents. Because of this search model, they are not suitable for the large-scale exploration of literature and the underlying biomedical conceptual relationships, which are common tasks in the age of high throughput experimental data analysis and cross-discipline research. We try to develop a new Medline exploration approach by incorporating interactive visualization together with powerful grouping, summary, sorting and active external content retrieval functions. Our solution, PubViz, is based on the FLEX platform designed for interactive web applications and its prototype is publicly available at: http://brainarray.mbni.med.umich.edu/Brainarray/DataMining/PubViz.
Text Mining for Adverse Drug Events: the Promise, Challenges, and State of the Art
Harpaz, Rave; Callahan, Alison; Tamang, Suzanne; Low, Yen; Odgers, David; Finlayson, Sam; Jung, Kenneth; LePendu, Paea; Shah, Nigam H.
2014-01-01
Text mining is the computational process of extracting meaningful information from large amounts of unstructured text. Text mining is emerging as a tool to leverage underutilized data sources that can improve pharmacovigilance, including the objective of adverse drug event detection and assessment. This article provides an overview of recent advances in pharmacovigilance driven by the application of text mining, and discusses several data sources—such as biomedical literature, clinical narratives, product labeling, social media, and Web search logs—that are amenable to text-mining for pharmacovigilance. Given the state of the art, it appears text mining can be applied to extract useful ADE-related information from multiple textual sources. Nonetheless, further research is required to address remaining technical challenges associated with the text mining methodologies, and to conclusively determine the relative contribution of each textual source to improving pharmacovigilance. PMID:25151493
Annotations of Mexican bullfighting videos for semantic index
NASA Astrophysics Data System (ADS)
Montoya Obeso, Abraham; Oropesa Morales, Lester Arturo; Fernando Vázquez, Luis; Cocolán Almeda, Sara Ivonne; Stoian, Andrei; García Vázquez, Mireya Saraí; Zamudio Fuentes, Luis Miguel; Montiel Perez, Jesús Yalja; de la O Torres, Saul; Ramírez Acosta, Alejandro Alvaro
2015-09-01
The video annotation is important for web indexing and browsing systems. Indeed, in order to evaluate the performance of video query and mining techniques, databases with concept annotations are required. Therefore, it is necessary generate a database with a semantic indexing that represents the digital content of the Mexican bullfighting atmosphere. This paper proposes a scheme to make complex annotations in a video in the frame of multimedia search engine project. Each video is partitioned using our segmentation algorithm that creates shots of different length and different number of frames. In order to make complex annotations about the video, we use ELAN software. The annotations are done in two steps: First, we take note about the whole content in each shot. Second, we describe the actions as parameters of the camera like direction, position and deepness. As a consequence, we obtain a more complete descriptor of every action. In both cases we use the concepts of the TRECVid 2014 dataset. We also propose new concepts. This methodology allows to generate a database with the necessary information to create descriptors and algorithms capable to detect actions to automatically index and classify new bullfighting multimedia content.
A Clustering Methodology of Web Log Data for Learning Management Systems
ERIC Educational Resources Information Center
Valsamidis, Stavros; Kontogiannis, Sotirios; Kazanidis, Ioannis; Theodosiou, Theodosios; Karakos, Alexandros
2012-01-01
Learning Management Systems (LMS) collect large amounts of data. Data mining techniques can be applied to analyse their web data log files. The instructors may use this data for assessing and measuring their courses. In this respect, we have proposed a methodology for analysing LMS courses and students' activity. This methodology uses a Markov…
Query Classification and Study of University Students' Search Trends
ERIC Educational Resources Information Center
Maabreh, Majdi A.; Al-Kabi, Mohammed N.; Alsmadi, Izzat M.
2012-01-01
Purpose: This study is an attempt to develop an automatic identification method for Arabic web queries and divide them into several query types using data mining. In addition, it seeks to evaluate the impact of the academic environment on using the internet. Design/methodology/approach: The web log files were collected from one of the higher…
Learning System of Web Navigation Patterns through Hypertext Probabilistic Grammars
ERIC Educational Resources Information Center
Cortes Vasquez, Augusto
2015-01-01
One issue of real interest in the area of web data mining is to capture users' activities during connection and extract behavior patterns that help define their preferences in order to improve the design of future pages adapting websites interfaces to individual users. This research is intended to provide, first of all, a presentation of the…
ERIC Educational Resources Information Center
Chen, Chih-Ming; Chang, Chia-Cheng
2014-01-01
Many studies have identified web-based cooperative learning as an increasingly popular educational paradigm with potential to increase learner satisfaction and interactions. However, peer-to-peer interaction often suffers barriers owing to a failure to explore useful social interaction information in web-based cooperative learning environments.…
Surveying the Commons: Current Implementation of Information Commons Web sites
ERIC Educational Resources Information Center
Leeder, Christopher
2009-01-01
This study assessed the content of 72 academic library Information Commons (IC) Web sites using content analysis, quantitative assessment and qualitative surveys of site administrators to analyze current implementation by the academic library community. Results show that IC Web sites vary widely in content, design and functionality, with few…
InterMine Webservices for Phytozome (Rev2)
DOE Office of Scientific and Technical Information (OSTI.GOV)
Carlson, Joseph; Goodstein, David; Rokhsar, Dan
2014-07-10
A datawarehousing framework for information provides a useful infrastructure for providers and users of genomic data. For providers, the infrastructure give them a consistent mechanism for extracting raw data. While for the users, the web services supported by the software allows them to make complex, and often unique, queries of the data. Previously, phytozome.net used BioMart to provide the infrastructure. As the complexity, scale and diversity of the dataset as grown, we decided to implement an InterMine web service on our servers. This change was largely motivated by the ability to have a more complex table structure and richer webmore » reporting mechanism than BioMart. For InterMine to achieve its more complex database schema it requires an XML description of the data and an appropriate loader. Unlimited one-to-many and many-to-many relationship between the tables can be enabled in the schema. We have implemented support for:1.) Genomes and annotations for the data in Phytozome. This set is the 48 organisms currently stored in a back end CHADO datastore. The data loaders are modified versions of the CHADO data adapters from FlyMine. 2.) Interproscan results from all proteins in the Phytozome database. 3.) Clusters of proteins into a grouped heirarchically by similarity. 4.) Cufflinks results from tissue-specific RNA-Seq data of Phytozome organisms. 5.) Diversity data (GATK and SnpEFF results) from a set of individual organism. The last two datatypes are new in this implementation of our web services. We anticipate that the scale of these data will increase considerably in the near future.« less
minepath.org: a free interactive pathway analysis web server.
Koumakis, Lefteris; Roussos, Panos; Potamias, George
2017-07-03
( www.minepath.org ) is a web-based platform that elaborates on, and radically extends the identification of differentially expressed sub-paths in molecular pathways. Besides the network topology, the underlying MinePath algorithmic processes exploit exact gene-gene molecular relationships (e.g. activation, inhibition) and are able to identify differentially expressed pathway parts. Each pathway is decomposed into all its constituent sub-paths, which in turn are matched with corresponding gene expression profiles. The highly ranked, and phenotype inclined sub-paths are kept. Apart from the pathway analysis algorithm, the fundamental innovation of the MinePath web-server concerns its advanced visualization and interactive capabilities. To our knowledge, this is the first pathway analysis server that introduces and offers visualization of the underlying and active pathway regulatory mechanisms instead of genes. Other features include live interaction, immediate visualization of functional sub-paths per phenotype and dynamic linked annotations for the engaged genes and molecular relations. The user can download not only the results but also the corresponding web viewer framework of the performed analysis. This feature provides the flexibility to immediately publish results without publishing source/expression data, and get all the functionality of a web based pathway analysis viewer. © The Author(s) 2017. Published by Oxford University Press on behalf of Nucleic Acids Research.
PLAN2L: a web tool for integrated text mining and literature-based bioentity relation extraction.
Krallinger, Martin; Rodriguez-Penagos, Carlos; Tendulkar, Ashish; Valencia, Alfonso
2009-07-01
There is an increasing interest in using literature mining techniques to complement information extracted from annotation databases or generated by bioinformatics applications. Here we present PLAN2L, a web-based online search system that integrates text mining and information extraction techniques to access systematically information useful for analyzing genetic, cellular and molecular aspects of the plant model organism Arabidopsis thaliana. Our system facilitates a more efficient retrieval of information relevant to heterogeneous biological topics, from implications in biological relationships at the level of protein interactions and gene regulation, to sub-cellular locations of gene products and associations to cellular and developmental processes, i.e. cell cycle, flowering, root, leaf and seed development. Beyond single entities, also predefined pairs of entities can be provided as queries for which literature-derived relations together with textual evidences are returned. PLAN2L does not require registration and is freely accessible at http://zope.bioinfo.cnio.es/plan2l.
Content-Based Personalization Services Integrating Folksonomies
NASA Astrophysics Data System (ADS)
Musto, Cataldo; Narducci, Fedelucio; Lops, Pasquale; de Gemmis, Marco; Semeraro, Giovanni
Basic content-based personalization consists in matching up the attributes of a user profile, in which preferences and interests are stored, with the attributes of a content object. The Web 2.0 (r)evolution has changed the game for personalization, from ‘elitary’ Web 1.0, written by few and read by many, to web content generated by everyone (user-generated content - UGC), since the role of people has evolved from passive consumers of information to that of active contributors.
Web accessibility support for visually impaired users using link content analysis.
Iwata, Hajime; Kobayashi, Naofumi; Tachibana, Kenji; Shirogane, Junko; Fukazawa, Yoshiaki
2013-12-01
Web pages are used for a variety of purposes. End users must understand dynamically changing content and sequentially follow page links to find desired material, requiring significant time and effort. However, for visually impaired users using screen readers, it can be difficult to find links to web pages when link text and alternative text descriptions are inappropriate. Our method supports the discovery of content by analyzing 8 categories of link types, and allows visually impaired users to be aware of the content represented by links in advance. This facilitates end users access to necessary information on web pages. Our method of classifying web page links is therefore effective as a means of evaluating accessibility.
30 CFR 57.6160 - Main facilities.
Code of Federal Regulations, 2010 CFR
2010-07-01
... Mineral Resources MINE SAFETY AND HEALTH ADMINISTRATION, DEPARTMENT OF LABOR METAL AND NONMETAL MINE SAFETY AND HEALTH SAFETY AND HEALTH STANDARDS-UNDERGROUND METAL AND NONMETAL MINES Explosives Storage... facilities will not prevent escape from the mine, or cause detonation of the contents of another storage...
The quality of pediatric orthopaedic information on the internet.
Winship, Brenton; Grisell, Margaret; Yang, Carolyn B; Chen, Rachel X; Bauer, Andrea S
2014-06-01
Many patients use the Internet for health information. However, there are few guarantees to the reliability and accuracy of this information. This study examined the quality and content of the Internet Web pages for 10 common pediatric orthopaedic diagnoses. We identified 10 common diagnoses in pediatric orthopaedics: brachial plexus injury, cerebral palsy, clubfoot, developmental dysplasia of the hip, leg length discrepancy, osteochondroma, polydactyly, scoliosis, spina bifida, and syndactyly. We used 2 of the most popular search engines to identify the top 10 Web sites for each disease. We evaluated the Web sites utilizing both the quality-based Health On the Net (HON) Foundation criteria and our own content-based grading sheets. The custom grading sheets focused on essential information about disease summary, pathogenesis, diagnosis, treatment, and prognosis. Three orthopaedic surgeons graded 98 academic, commercial, nonprofit, and physicians' Web sites for 10 diseases. Academic Web sites scored the highest in content (mean, 60.8% ± 15.5%), whereas commercial Web sites scored the lowest (mean, 46.7% ± 22.2%). Among the diagnoses, osteochondroma Web sites had the highest content scores (mean, 75.8% ± 11.8%), whereas polydactyly Web sites had the lowest content scores (mean, 39.3% ± 15.7%). In contrast, Web sites about developmental dysplasia of the hip had the highest HON scores (65.0 ± 11.1), whereas those about brachial plexus birth palsy scored the lowest (42.6% ± 16.9%). Among the content subgroups, scores were generally higher for disease summary and diagnostics and lower for prognosis. The Internet Web sites reviewed demonstrated a wide range of content and information. We found that nonprofit and academic Web sites were the most reliable sources, whereas commercial and, surprisingly, physician-run Web sites were the least reliable. We advise physicians to talk to their patients about the information they get on the Internet and how it dictates their expectations. We hope this study, combined with further understanding of how our patients use this information, can help improve the Internet content. Physicians should know that their patients may be receiving misleading information from the Internet and be able to discuss this with their patients.
Mishra, Mark V; Bennett, Michele; Vincent, Armon; Lee, Olivia T; Lallas, Costas D; Trabulsi, Edouard J; Gomella, Leonard G; Dicker, Adam P; Showalter, Timothy N
2013-01-01
Qualitative research aimed at identifying patient acceptance of active surveillance (AS) has been identified as a public health research priority. The primary objective of this study was to determine if analysis of a large-sample of anonymous internet conversations (ICs) could be utilized to identify unmet public needs regarding AS. English-language ICs regarding prostate cancer (PC) treatment with AS from 2002-12 were identified using a novel internet search methodology. Web spiders were developed to mine, aggregate, and analyze content from the world-wide-web for ICs centered on AS. Collection of ICs was not restricted to any specific geographic region of origin. NLP was used to evaluate content and perform a sentiment analysis. Conversations were scored as positive, negative, or neutral. A sentiment index (SI) was subsequently calculated according to the following formula to compare temporal trends in public sentiment towards AS: [(# Positive IC/#Total IC)-(#Negative IC/#Total IC) x 100]. A total of 464 ICs were identified. Sentiment increased from -13 to +2 over the study period. The increase sentiment has been driven by increased patient emphasis on quality-of-life factors and endorsement of AS by national medical organizations. Unmet needs identified in these ICs include: a gap between quantitative data regarding long-term outcomes with AS vs. conventional treatments, desire for treatment information from an unbiased specialist, and absence of public role models managed with AS. This study demonstrates the potential utility of online patient communications to provide insight into patient preferences and decision-making. Based on our findings, we recommend that multidisciplinary clinics consider including an unbiased specialist to present treatment options and that future decision tools for AS include quantitative data regarding outcomes after AS.
Handling Dynamic Weights in Weighted Frequent Pattern Mining
NASA Astrophysics Data System (ADS)
Ahmed, Chowdhury Farhan; Tanbeer, Syed Khairuzzaman; Jeong, Byeong-Soo; Lee, Young-Koo
Even though weighted frequent pattern (WFP) mining is more effective than traditional frequent pattern mining because it can consider different semantic significances (weights) of items, existing WFP algorithms assume that each item has a fixed weight. But in real world scenarios, the weight (price or significance) of an item can vary with time. Reflecting these changes in item weight is necessary in several mining applications, such as retail market data analysis and web click stream analysis. In this paper, we introduce the concept of a dynamic weight for each item, and propose an algorithm, DWFPM (dynamic weighted frequent pattern mining), that makes use of this concept. Our algorithm can address situations where the weight (price or significance) of an item varies dynamically. It exploits a pattern growth mining technique to avoid the level-wise candidate set generation-and-test methodology. Furthermore, it requires only one database scan, so it is eligible for use in stream data mining. An extensive performance analysis shows that our algorithm is efficient and scalable for WFP mining using dynamic weights.
ERIC Educational Resources Information Center
Baris, Mehmet Fatih
2015-01-01
Several studies have been conducted on technological, pedagogical content knowledge and web-based education. In this study, the Technological Pedagogical Content Knowledge and Educational Use of Web Technologies (TPCK-W) were analyzed in addition to the self-efficacy and attitudes of 33 teachers from eight different branches carrying out their…
Prediction of fish and sediment mercury in streams using landscape variables and historical mining.
Alpers, Charles N; Yee, Julie L; Ackerman, Joshua T; Orlando, James L; Slotton, Darrel G; Marvin-DiPasquale, Mark C
2016-11-15
Widespread mercury (Hg) contamination of aquatic systems in the Sierra Nevada of California, U.S., is associated with historical use to enhance gold (Au) recovery by amalgamation. In areas affected by historical Au mining operations, including the western slope of the Sierra Nevada and downstream areas in northern California, such as San Francisco Bay and the Sacramento River-San Joaquin River Delta, microbial conversion of Hg to methylmercury (MeHg) leads to bioaccumulation of MeHg in food webs, and increased risks to humans and wildlife. This study focused on developing a predictive model for THg in stream fish tissue based on geospatial data, including land use/land cover data, and the distribution of legacy Au mines. Data on total mercury (THg) and MeHg concentrations in fish tissue and streambed sediment collected during 1980-2012 from stream sites in the Sierra Nevada, California were combined with geospatial data to estimate fish THg concentrations across the landscape. THg concentrations of five fish species (Brown Trout, Rainbow Trout, Sacramento Pikeminnow, Sacramento Sucker, and Smallmouth Bass) within stream sections were predicted using multi-model inference based on Akaike Information Criteria, using geospatial data for mining history and landscape characteristics as well as fish species and length (r(2)=0.61, p<0.001). Including THg concentrations in streambed sediment did not improve the model's fit, however including MeHg concentrations in streambed sediment, organic content (loss on ignition), and sediment grain size resulted in an improved fit (r(2)=0.63, p<0.001). These models can be used to estimate THg concentrations in stream fish based on landscape variables in the Sierra Nevada in areas where direct measurements of THg concentration in fish are unavailable. Published by Elsevier B.V.
Prediction of fish and sediment mercury in streams using landscape variables and historical mining
Alpers, Charles N.; Yee, Julie L.; Ackerman, Joshua T.; Orlando, James L.; Slotton, Darrell G.; Marvin-DiPasquale, Mark C.
2016-01-01
Widespread mercury (Hg) contamination of aquatic systems in the Sierra Nevada of California, U.S., is associated with historical use to enhance gold (Au) recovery by amalgamation. In areas affected by historical Au mining operations, including the western slope of the Sierra Nevada and downstream areas in northern California, such as San Francisco Bay and the Sacramento River–San Joaquin River Delta, microbial conversion of Hg to methylmercury (MeHg) leads to bioaccumulation of MeHg in food webs, and increased risks to humans and wildlife. This study focused on developing a predictive model for THg in stream fish tissue based on geospatial data, including land use/land cover data, and the distribution of legacy Au mines. Data on total mercury (THg) and MeHg concentrations in fish tissue and streambed sediment collected during 1980–2012 from stream sites in the Sierra Nevada, California were combined with geospatial data to estimate fish THg concentrations across the landscape. THg concentrations of five fish species (Brown Trout, Rainbow Trout, Sacramento Pikeminnow, Sacramento Sucker, and Smallmouth Bass) within stream sections were predicted using multi-model inference based on Akaike Information Criteria, using geospatial data for mining history and landscape characteristics as well as fish species and length (r2 = 0.61, p < 0.001). Including THg concentrations in streambed sediment did not improve the model's fit, however including MeHg concentrations in streambed sediment, organic content (loss on ignition), and sediment grain size resulted in an improved fit (r2 = 0.63, p < 0.001). These models can be used to estimate THg concentrations in stream fish based on landscape variables in the Sierra Nevada in areas where direct measurements of THg concentration in fish are unavailable.
[Distribution characteristics of copper in soil and rape around Tongling mining area].
Shen, Chang-Gao; Gao, Chao; Wang, Deng-Feng; Wang, Lei; Chen, Fu-Rong
2007-10-01
Soil and rape samples around Tongling mining area were collected, and their copper (Cu) contents were investigated. The results showed that the upland soil developed on the slope deposit around the mining area as well as the paddy soil distributed in lower reaches was heavily polluted by Cu, while the fluvo-aquic soil further from the mining area was less contaminated. Though the Cu content in paddy soil and upland soil was nearly the same, its bioavailability was higher in paddy soil, due to the Cu pollution of irrigated water. There was a significant correlation between available and total Cu in these three types of soil. The activation rate of soil Cu (percentage of available Cu in total Cu) was 15.0% on average, which was positively correlated with soil total Cu and organic matter while negatively correlated with soil pH and Mn. The average Cu content in rape seed and stalk was 4.0 and 5.8 mg x kg(-1), respectively. The rape Cu content increased obviously with increasing soil available Cu content when the soil available Cu content was relatively low, but the Cu absorption and accumulation by rape decreased gradually when the soil available copper content was higher than 30 mg x kg(-1).
ERIC Educational Resources Information Center
Lee, Min-Hsien; Tsai, Chin-Chung
2010-01-01
Research in the area of educational technology has claimed that Web technology has driven online pedagogy such that teachers need to know how to use Web technology to assist their teaching. This study provides a framework for understanding teachers' Technological Pedagogical Content Knowledge-Web (TPCK-W), while integrating Web technology into…
Presser, Theresa S.
2013-01-01
Investigating the presence and variability of prey and predator species in demographically open systems such as streams also is key to model outcomes given the overall environmental stressors (for example, general landscape change, food-web disruption, recolonization potential) imposed on the composition of biological communities in coal mining and valley-fill affected watersheds
Publications - GMC 273 | Alaska Division of Geological & Geophysical
holes received at the GMC (1 box, holes N1 through N8) of the INEXCO Mining Company Nikolai Project , holes N1 through N8) of the INEXCO Mining Company Nikolai Project, McCarthy, Alaska that consist of core Alaska's Mineral Industry Reports AKGeology.info Rare Earth Elements WebGeochem Engineering Geology Alaska
Analysis of pathology department Web sites and practical recommendations.
Nero, Christopher; Dighe, Anand S
2008-09-01
There are numerous customers for pathology departmental Web sites, including pathology department staff, clinical staff, residency applicants, job seekers, and other individuals outside the department seeking department information. Despite the increasing importance of departmental Web sites as a means of distributing information, no analysis has been done to date of the content and usage of pathology department Web sites. In this study, we analyzed pathology department Web sites to examine the elements present on each site and to evaluate the use of search technology on these sites. Further, we examined the usage patterns of our own departmental Internet and internet Web sites to better understand the users of pathology Web sites. We reviewed selected departmental pathology Web sites and analyzed their content and functionality. Our institution's departmental pathology Web sites were modified to enable detailed information to be stored regarding users and usage patterns, and that information was analyzed. We demonstrate considerable heterogeneity in departmental Web sites with many sites lacking basic content and search features. In addition, we demonstrate that increasing the traffic of a department's informational Web sites may result in reduced phone inquiries to the laboratory. We propose recommendations for pathology department Web sites to maximize promotion of a department's mission. A departmental pathology Web site is an essential communication tool for all pathology departments, and attention to the users and content of the site can have operational impact.
Video mining using combinations of unsupervised and supervised learning techniques
NASA Astrophysics Data System (ADS)
Divakaran, Ajay; Miyahara, Koji; Peker, Kadir A.; Radhakrishnan, Regunathan; Xiong, Ziyou
2003-12-01
We discuss the meaning and significance of the video mining problem, and present our work on some aspects of video mining. A simple definition of video mining is unsupervised discovery of patterns in audio-visual content. Such purely unsupervised discovery is readily applicable to video surveillance as well as to consumer video browsing applications. We interpret video mining as content-adaptive or "blind" content processing, in which the first stage is content characterization and the second stage is event discovery based on the characterization obtained in stage 1. We discuss the target applications and find that using a purely unsupervised approach are too computationally complex to be implemented on our product platform. We then describe various combinations of unsupervised and supervised learning techniques that help discover patterns that are useful to the end-user of the application. We target consumer video browsing applications such as commercial message detection, sports highlights extraction etc. We employ both audio and video features. We find that supervised audio classification combined with unsupervised unusual event discovery enables accurate supervised detection of desired events. Our techniques are computationally simple and robust to common variations in production styles etc.
Narcissism and social networking Web sites.
Buffardi, Laura E; Campbell, W Keith
2008-10-01
The present research examined how narcissism is manifested on a social networking Web site (i.e., Facebook.com). Narcissistic personality self-reports were collected from social networking Web page owners. Then their Web pages were coded for both objective and subjective content features. Finally, strangers viewed the Web pages and rated their impression of the owner on agentic traits, communal traits, and narcissism. Narcissism predicted (a) higher levels of social activity in the online community and (b) more self-promoting content in several aspects of the social networking Web pages. Strangers who viewed the Web pages judged more narcissistic Web page owners to be more narcissistic. Finally, mediational analyses revealed several Web page content features that were influential in raters' narcissistic impressions of the owners, including quantity of social interaction, main photo self-promotion, and main photo attractiveness. Implications of the expression of narcissism in social networking communities are discussed.
The Geomatics Contribution for the Valorisation Project in the Rocca of San Silvestro Landscape Site
NASA Astrophysics Data System (ADS)
Brocchini, D.; Chiabrando, F.; Colucci, E.; Sammartano, G.; Spanò, A.; Teppati Losè, L.; Villa, A.
2017-05-01
This paper proposes an emblematic project where several multi-sensor strategies for spatial data acquisition and management, range based and image based, were combined to create a series of integrated territorial and architectural scale products characterized by a rich multi-content nature. The work presented here was finalized in a test site that is composed by an ensemble of diversified cultural deposits; the objects that were surveyed and modelled range from the landscape with its widespread mining sites, the main tower with its defensive role, the urban configuration of the settlement, the building systems and techniques, a medieval mine. For this reason, the Rocca of San Silvestro represented a perfect test case, due to its complex and multi-stratified character. This archaeological site is a medieval fortified village near the municipality of Campiglia Marittima (LI), Italy. The Rocca is part of an Archaeological Mines Park and is included in the Parchi della Val di Cornia (a system of archaeological parks, natural parks and museums in the south-west of Tuscany). The fundamental role of a deep knowledge about a cultural artefact before the planning of a restoration and valorisation project is globally recognized; the qualitative and quantitative knowledge provided by geomatics techniques is part of this process. The paper will present the different techniques that were used, the products that were obtained and will focus on some mapping and WEB GIS applications and analyses that were performed and considerations that were made.
Meta Data Mining in Earth Remote Sensing Data Archives
NASA Astrophysics Data System (ADS)
Davis, B.; Steinwand, D.
2014-12-01
Modern search and discovery tools for satellite based remote sensing data are often catalog based and rely on query systems which use scene- (or granule-) based meta data for those queries. While these traditional catalog systems are often robust, very little has been done in the way of meta data mining to aid in the search and discovery process. The recently coined term "Big Data" can be applied in the remote sensing world's efforts to derive information from the vast data holdings of satellite based land remote sensing data. Large catalog-based search and discovery systems such as the United States Geological Survey's Earth Explorer system and the NASA Earth Observing System Data and Information System's Reverb-ECHO system provide comprehensive access to these data holdings, but do little to expose the underlying scene-based meta data. These catalog-based systems are extremely flexible, but are manually intensive and often require a high level of user expertise. Exposing scene-based meta data to external, web-based services can enable machine-driven queries to aid in the search and discovery process. Furthermore, services which expose additional scene-based content data (such as product quality information) are now available and can provide a "deeper look" into remote sensing data archives too large for efficient manual search methods. This presentation shows examples of the mining of Landsat and Aster scene-based meta data, and an experimental service using OPeNDAP to extract information from quality band from multiple granules in the MODIS archive.
ERIC Educational Resources Information Center
Small, Ruth V.; Arnone, Marilyn P.
Intended for use by middle or high school teachers and library media specialists, this book describes a World Wide Web evaluation tool developed specifically for use by high school students and designed to provide hands-on experience in critically evaluating the strengths and weaknesses of Web sites. The book uses a workbook format and is…
Web mining for topics defined by complex and precise predicates
NASA Astrophysics Data System (ADS)
Lee, Ching-Cheng; Sampathkumar, Sushma
2004-04-01
The enormous growth of the World Wide Web has made it important to perform resource discovery efficiently for any given topic. Several new techniques have been proposed in the recent years for this kind of topic specific web-mining, and among them a key new technique called focused crawling which is able to crawl topic-specific portions of the web without having to explore all pages. Most existing research on focused crawling considers a simple topic definition that typically consists of one or more keywords connected by an OR operator. However this kind of simple topic definition may result in too many irrelevant pages in which the same keyword appears in a wrong context. In this research we explore new strategies for crawling topic specific portions of the web using complex and precise predicates. A complex predicate will allow the user to precisely specify a topic using Boolean operators such as "AND", "OR" and "NOT". Our work will concentrate on defining a format to specify this kind of a complex topic definition and secondly on devising a crawl strategy to crawl the topic specific portions of the web defined by the complex predicate, efficiently and with minimal overhead. Our new crawl strategy will improve the performance of topic-specific web crawling by reducing the number of irrelevant pages crawled. In order to demonstrate the effectiveness of the above approach, we have built a complete focused crawler called "Eureka" with complex predicate support, and a search engine that indexes and supports end-user searches on the crawled pages.
Scholarly Context Adrift: Three out of Four URI References Lead to Changed Content
Tobin, Richard; Grover, Claire
2016-01-01
Increasingly, scholarly articles contain URI references to “web at large” resources including project web sites, scholarly wikis, ontologies, online debates, presentations, blogs, and videos. Authors reference such resources to provide essential context for the research they report on. A reader who visits a web at large resource by following a URI reference in an article, some time after its publication, is led to believe that the resource’s content is representative of what the author originally referenced. However, due to the dynamic nature of the web, that may very well not be the case. We reuse a dataset from a previous study in which several authors of this paper were involved, and investigate to what extent the textual content of web at large resources referenced in a vast collection of Science, Technology, and Medicine (STM) articles published between 1997 and 2012 has remained stable since the publication of the referencing article. We do so in a two-step approach that relies on various well-established similarity measures to compare textual content. In a first step, we use 19 web archives to find snapshots of referenced web at large resources that have textual content that is representative of the state of the resource around the time of publication of the referencing paper. We find that representative snapshots exist for about 30% of all URI references. In a second step, we compare the textual content of representative snapshots with that of their live web counterparts. We find that for over 75% of references the content has drifted away from what it was when referenced. These results raise significant concerns regarding the long term integrity of the web-based scholarly record and call for the deployment of techniques to combat these problems. PMID:27911955
Text mining for adverse drug events: the promise, challenges, and state of the art.
Harpaz, Rave; Callahan, Alison; Tamang, Suzanne; Low, Yen; Odgers, David; Finlayson, Sam; Jung, Kenneth; LePendu, Paea; Shah, Nigam H
2014-10-01
Text mining is the computational process of extracting meaningful information from large amounts of unstructured text. It is emerging as a tool to leverage underutilized data sources that can improve pharmacovigilance, including the objective of adverse drug event (ADE) detection and assessment. This article provides an overview of recent advances in pharmacovigilance driven by the application of text mining, and discusses several data sources-such as biomedical literature, clinical narratives, product labeling, social media, and Web search logs-that are amenable to text mining for pharmacovigilance. Given the state of the art, it appears text mining can be applied to extract useful ADE-related information from multiple textual sources. Nonetheless, further research is required to address remaining technical challenges associated with the text mining methodologies, and to conclusively determine the relative contribution of each textual source to improving pharmacovigilance.
Görzig, Anke
2016-08-01
Possible links of cyberbullying with suicide and psychological problems have recently received considerable attention. Suicide-related behaviors have also been linked with viewing of associated web content. Studies on traditional bullying indicate that the roles of bullying involvement (bullies, victims, and bully-victims) matter in terms of associations with specific suicide-related behaviors and psychological problems. Yet, related research in the area of cyberbullying is lacking. The current study investigates the association of cyberbullying roles with viewing of specific suicide-related web content and psychological problems. Data from N = 19,406 (50 percent girls) 11-16-year-olds (M = 13.54, SD = 1.68) of a representative sample of Internet-using children in Europe were analyzed. Self-reports were obtained for cyberbullying role, viewing of web content related to self-harm, and suicide, as well as the emotional, peer, and conduct problem subscales of the Strengths and Difficulties Questionnaire (SDQ). Multinomial logistic regression analyses revealed that compared with those not involved in cyberbullying, viewing of web content related to suicide was higher for cybervictims and cyberbully-victims, but not for cyberbullies. Viewing of web content related to self-harm was higher for all cyberbullying roles, especially for cyberbully-victims. Rates of emotional problems were higher among cybervictims and cyberbully-victims, rates of peer problems were higher for cybervictims, and rates of conduct problems were higher for all cyberbullying roles. Moreover, the links between cyberbullying role and viewing of suicide-related web content were independent of psychological problems. The results can be useful to more precisely target efforts toward the specific problems of each cyberbullying role. The outcomes on viewing of web content also indicate an opportunity to enhance the presence of health service providers on Internet platforms.
A Radioelement Analysis of the Northern Black Hills, South Dakota, U.S.A
NASA Astrophysics Data System (ADS)
Young, Dylan Wade
The uranium, thorium, and potassium contents from 736 samples, within a 15-km radius of the Homestake Gold Mine and Sanford Underground Research Facility in the Northern Black Hills indicate the geoneutrino background may be higher than average for the continental crust. The radioactive element contents of igneous, metamorphic, and sedimentary rocks were determined by gamma ray spectrometry. Many rocks show hydrothermal and metamorphic alteration within the last ten Ma of the Tertiary period. Young alkali rich igneous rocks, such as rhyolite, phonolite and other volcanic rocks, have lower than average Th:U ratios. The radioelement content of 215 igneous rocks were determined. The radioelement contents of 143 metamorphic rocks were determined. This study also shows that metamorphic rocks were found to have low variable U:Th content when compared to content in igneous rocks. Sedimentary rocks, in general, have low U, Th, and K content. The radioelement content of 236 sedimentary rocks were determined. Rocks present within the Homestake Gold Mine, are highly altered by hydrothermal and metamorphic activity, enriching U, and in some areas, Th content. The Homestake Gold Mine lies almost entirely within metamorphic rocks. Igneous rocks occur in the mine as veins and dikes. The dominant igneous rock present is rhyolite. Metamorphic rocks present inside the HGM, were divided by formation; Ellison Fm, Poorman Fm, Yates Unit [lower Poorman Fm], Homestake Fm, and Flagrock Fm. The finding of high radioelement content in the rocks suggests that the antineutrinos background at the HGM will need to be considered and calibrated for, in future experiments conducted at the Sanford Underground Research Facility. A geoneutrino luminosity of 1.26x105 (mg-1s -1) was calculated from the samples analyzed within the Homestake Gold Mine. A total geoneutrino luminosity of 4.44x105 (mg -1s=1) was calculated from the sum of all analyses conducted in the Northern Black Hills.
20 CFR 726.305 - Contents of notice.
Code of Federal Regulations, 2010 CFR
2010-04-01
... 20 Employees' Benefits 3 2010-04-01 2010-04-01 false Contents of notice. 726.305 Section 726.305 Employees' Benefits EMPLOYMENT STANDARDS ADMINISTRATION, DEPARTMENT OF LABOR FEDERAL COAL MINE HEALTH AND SAFETY ACT OF 1969, AS AMENDED BLACK LUNG BENEFITS; REQUIREMENTS FOR COAL MINE OPERATOR'S INSURANCE Civil...
Characteristics of Food Industry Web Sites and "Advergames" Targeting Children
ERIC Educational Resources Information Center
Culp, Jennifer; Bell, Robert A.; Cassady, Diana
2010-01-01
Objective: To assess the content of food industry Web sites targeting children by describing strategies used to prolong their visits and foster brand loyalty; and to document health-promoting messages on these Web sites. Design: A content analysis was conducted of Web sites advertised on 2 children's networks, Cartoon Network and Nickelodeon. A…
Library Web Sites in Pakistan: An Analysis of Content
ERIC Educational Resources Information Center
Qutab, Saima; Mahmood, Khalid
2009-01-01
Purpose: The purpose of this paper is to investigate library web sites in Pakistan, to analyse their content and navigational strengths and weaknesses and to give recommendations for developing better web sites and quality assessment studies. Design/methodology/approach: Survey of web sites of 52 academic, special, public and national libraries in…
Growing and navigating the small world Web by local content
Menczer, Filippo
2002-01-01
Can we model the scale-free distribution of Web hypertext degree under realistic assumptions about the behavior of page authors? Can a Web crawler efficiently locate an unknown relevant page? These questions are receiving much attention due to their potential impact for understanding the structure of the Web and for building better search engines. Here I investigate the connection between the linkage and content topology of Web pages. The relationship between a text-induced distance metric and a link-based neighborhood probability distribution displays a phase transition between a region where linkage is not determined by content and one where linkage decays according to a power law. This relationship is used to propose a Web growth model that is shown to accurately predict the distribution of Web page degree, based on textual content and assuming only local knowledge of degree for existing pages. A qualitatively similar phase transition is found between linkage and semantic distance, with an exponential decay tail. Both relationships suggest that efficient paths can be discovered by decentralized Web navigation algorithms based on textual and/or categorical cues. PMID:12381792
Growing and navigating the small world Web by local content
NASA Astrophysics Data System (ADS)
Menczer, Filippo
2002-10-01
Can we model the scale-free distribution of Web hypertext degree under realistic assumptions about the behavior of page authors? Can a Web crawler efficiently locate an unknown relevant page? These questions are receiving much attention due to their potential impact for understanding the structure of the Web and for building better search engines. Here I investigate the connection between the linkage and content topology of Web pages. The relationship between a text-induced distance metric and a link-based neighborhood probability distribution displays a phase transition between a region where linkage is not determined by content and one where linkage decays according to a power law. This relationship is used to propose a Web growth model that is shown to accurately predict the distribution of Web page degree, based on textual content and assuming only local knowledge of degree for existing pages. A qualitatively similar phase transition is found between linkage and semantic distance, with an exponential decay tail. Both relationships suggest that efficient paths can be discovered by decentralized Web navigation algorithms based on textual and/or categorical cues.
Growing and navigating the small world Web by local content.
Menczer, Filippo
2002-10-29
Can we model the scale-free distribution of Web hypertext degree under realistic assumptions about the behavior of page authors? Can a Web crawler efficiently locate an unknown relevant page? These questions are receiving much attention due to their potential impact for understanding the structure of the Web and for building better search engines. Here I investigate the connection between the linkage and content topology of Web pages. The relationship between a text-induced distance metric and a link-based neighborhood probability distribution displays a phase transition between a region where linkage is not determined by content and one where linkage decays according to a power law. This relationship is used to propose a Web growth model that is shown to accurately predict the distribution of Web page degree, based on textual content and assuming only local knowledge of degree for existing pages. A qualitatively similar phase transition is found between linkage and semantic distance, with an exponential decay tail. Both relationships suggest that efficient paths can be discovered by decentralized Web navigation algorithms based on textual and/or categorical cues.
NASA Astrophysics Data System (ADS)
Bos, Nathan Daniel
This dissertation investigates the emerging affordance of the World Wide Web as a place for high school students to become authors and publishers of information. Two empirical studies lay groundwork for student publishing by examining learning issues related to audience adaptation in writing, motivation and engagement with hypermedia, design, problem-solving, and critical evaluation. Two models of student publishing on the World Wide Web were investigated over the course of two 11spth grade project-based science curriculums. In the first curricular model, students worked in pairs to design informative hypermedia projects about infectious diseases that were published on the Web. Four case studies were written, drawing on both product- and process-related data sources. Four theoretically important findings are illustrated through these cases: (1) multimedia, especially graphics, seemed to catalyze some students' design processes by affecting the sequence of their design process and by providing a connection between the science content and their personal interest areas, (2) hypermedia design can demand high levels of analysis and synthesis of science content, (3) students can learn to think about science content representation through engagement with challenging design tasks, and (4) students' consideration of an outside audience can be facilitated by teacher-given design principles. The second Web-publishing model examines how students critically evaluate scientific resources on the Web, and how students can contribute to the Web's organization and usability by publishing critical reviews. Students critically evaluated Web resources using a four-part scheme: summarization of content, content, evaluation of credibility, evaluation of organizational structure, and evaluation of appearance. Content analyses comparing students' reviews and reviewed Web documents showed that students were proficient at summarizing content of Web documents, identifying their publishing source, and evaluating their organizational features; however, students struggled to identify scientific evidence, bias, or sophisticated use of media in Web pages. Shortcomings were shown to be partly due to deficiencies in the Web pages themselves and partly due to students' inexperience with the medium or lack of critical evaluation skills. Future directions of this idea are discussed, including discussion of how students' reviews have been integrated into a current digital library development project.
Stress monitoring versus microseismic ruptures in an active deep mine
NASA Astrophysics Data System (ADS)
Tonnellier, Alice; Bouffier, Christian; Bigarré, Pascal; Nyström, Anders; Österberg, Anders; Fjellström, Peter
2015-04-01
Nowadays, underground mining industry has developed high-technology mass mining methods to optimise the productivity at deep levels. Such massive extraction induces high-level stress redistribution generating seismic events around the mining works, threatening safety and economics. For this reason mining irregular deep ore bodies calls for steadily enhanced scientific practises and technologies to guarantee the mine environment to be safer and stable for the miners and the infrastructures. INERIS, within the framework of the FP7 European project I2Mine and in partnership with the Swedish mining company Boliden, has developed new methodologies in order to monitor both quasi-static stress changes and ruptures in a seismic prone area. To this purpose, a unique local permanent microseismic and stress monitoring network has been installed into the deep-working Garpenberg mine situated to the north of Uppsala (Sweden). In this mine, ore is extracted using sublevel stoping with paste fill production/distribution system and long-hole drilling method. This monitoring network has been deployed between about 1100 and 1250 meter depth. It consists in six 1-component and five 3-component microseismic probes (14-Hz geophones) deployed in the Lappberget area, in addition to three 3D stress monitoring cells that focus on a very local exploited area. Objective is three-fold: to quantify accurately quasi-static stress changes and freshly-induced stress gradients with drift development in the orebody, to study quantitatively those stress changes versus induced detected and located microseismic ruptures, and possibly to identify quasi-static stress transfer from those seismic ruptures. Geophysical and geotechnical data are acquired continuously and automatically transferred to INERIS datacenter through the web. They are made available on a secured web cloud monitoring infrastructure called e.cenaris and completed with mine data. Such interface enables the visualisation of the monitoring data coming from the mine in quasi-real time and facilitates information exchanges and decision making for experts and stakeholders. On the basis of these data acquisition and sharing, preliminary analysis has been started to highlight whether stress variations and seismic sources behaviour might be directly bound with mine working evolution and could improve the knowledge on the equilibrium states inside the mine. Knowing such parameters indeed will be a potential solution to understand better the response of deep mining activities to the exploitation solicitations and to develop, if possible, methods to prevent from major hazards such as rock bursts and other ground failure phenomena.
Song, Juyoung; Song, Tae Min; Seo, Dong-Chul; Jin, Jae Hyun
2016-12-01
To investigate online search activity of suicide-related words in South Korean adolescents through data mining of social media Web sites as the suicide rate in South Korea is one of the highest in the world. Out of more than 2.35 billion posts for 2 years from January 1, 2011 to December 31, 2012 on 163 social media Web sites in South Korea, 99,693 suicide-related documents were retrieved by Crawler and analyzed using text mining and opinion mining. These data were further combined with monthly employment rate, monthly rental prices index, monthly youth suicide rate, and monthly number of reported bully victims to fit multilevel models as well as structural equation models. The link from grade pressure to suicide risk showed the largest standardized path coefficient (beta = .357, p < .001) in structural models and a significant random effect (p < .01) in multilevel models. Depression was a partial mediator between suicide risk and grade pressure, low body image, victims of bullying, and concerns about disease. The largest total effect was observed in the grade pressure to depression to suicide risk. The multilevel models indicate about 27% of the variance in the daily suicide-related word search activity is explained by month-to-month variations. A lower employment rate, a higher rental prices index, and more bullying were associated with an increased suicide-related word search activity. Academic pressure appears to be the biggest contributor to Korean adolescents' suicide risk. Real-time suicide-related word search activity monitoring and response system needs to be developed. Copyright © 2016 Society for Adolescent Health and Medicine. Published by Elsevier Inc. All rights reserved.
Trippi, Michael H.; Ruppert, Leslie F.; Attanasi, E.D.; Milici, Robert C.; Freeman, P.A.
2014-01-01
Data from 157 counties in the Appalachian basin of average sulfur content of coal mined for electrical power generation from 1983 through 2005 show a general decrease in the number of counties where coal mining has occurred and a decrease in the number of counties where higher sulfur coals (>2 percent sulfur) were mined. Calculated potential SO2 emissions (assuming no post-combustion SO2 removal) show a corresponding decrease over the same period of time.
30 CFR 937.777 - General content requirements for permit applications.
Code of Federal Regulations, 2013 CFR
2013-07-01
... 30 Mineral Resources 3 2013-07-01 2013-07-01 false General content requirements for permit applications. 937.777 Section 937.777 Mineral Resources OFFICE OF SURFACE MINING RECLAMATION AND ENFORCEMENT, DEPARTMENT OF THE INTERIOR PROGRAMS FOR THE CONDUCT OF SURFACE MINING OPERATIONS WITHIN EACH STATE OREGON...
30 CFR 937.777 - General content requirements for permit applications.
Code of Federal Regulations, 2014 CFR
2014-07-01
... 30 Mineral Resources 3 2014-07-01 2014-07-01 false General content requirements for permit applications. 937.777 Section 937.777 Mineral Resources OFFICE OF SURFACE MINING RECLAMATION AND ENFORCEMENT, DEPARTMENT OF THE INTERIOR PROGRAMS FOR THE CONDUCT OF SURFACE MINING OPERATIONS WITHIN EACH STATE OREGON...
Gray, John E.; Stillings, Lisa L.
2003-01-01
Mercury and methylmercury concentrations were measured in mine wastes, stream sediments, and stream waters collected both proximal and distal from abandoned mercury mines to evaluate mercury contamination and mercury methylation in the Humboldt River system. The climate in the study area is arid, and due to the lack of mine-water runoff, water-leaching laboratory experiments were used to evaluate the potential of mine wastes to release mercury. Mine-waste calcine contains mercury concentrations as high as 14,000 ?g/g. Stream-sediment samples collected within 1 km of the mercury mines studied contain mercury concentrations as high as 170 ?g/g, but sediments collected from the Humboldt River and regional baseline sites have much lower mercury contents, less than 0.44 ?g/g. Similarly, methylmercury concentrations in mine-waste calcine are locally as high as 96 ng/g, but methylmercury contents in stream sediments collected down-stream from the mines and from the Humboldt River are lower (<0.05-0.95 ng/g). Stream-water samples collected below two mines studied contain mercury concentrations ranging from 6 to 2,000 ng/L, whereas mercury contents in Humboldt River and Rye Patch Reservoir water were generally lower, ranging from 2.1 to 9.0 ng/L. Methylmercury concentrations in Humboldt River system water were the lowest in this study (<0.02- 0.27 ng/L). Although mercury and methylmercury concentrations were elevated in some mine-waste calcine and mercury concentrations were locally high in mine-waste leachate samples, data show significant dilution of mercury and lower mercury methylation down gradient from the mines, especially in the sediments and water collected from the Humboldt River, which is more than 8 km from any mercury mines. Data show only minor, local transference of mercury and methylmercury from mine-waste calcine to stream sediment, and then onto the water column, and indicate little transference of mercury from the mine sites to the Humboldt River system.
One EPA Web Principles that Guide Content Development
The principles of One EPA Web can be applied to better meet the needs and expectations of our audiences, fit their information-seeking behavior, and help them accomplish tasks. Learn about the five paths forward for transforming web content.
Data Mining for Web Site Evaluation: An Exploration of Site Usage by Graduate Social Work Students
ERIC Educational Resources Information Center
Santhiveeran, Janaki
2006-01-01
This paper evaluates the actual use of a course Website by graduate social work students. The study utilized data mining techniques to discover meaningful trends by using the data from server logs. The course Website was accessed 24,730 times by all 49 graduate students during a semester. The students utilized the course Website 23 hours a day, 7…
Cañada, Andres; Rabal, Obdulia; Oyarzabal, Julen; Valencia, Alfonso
2017-01-01
Abstract A considerable effort has been devoted to retrieve systematically information for genes and proteins as well as relationships between them. Despite the importance of chemical compounds and drugs as a central bio-entity in pharmacological and biological research, only a limited number of freely available chemical text-mining/search engine technologies are currently accessible. Here we present LimTox (Literature Mining for Toxicology), a web-based online biomedical search tool with special focus on adverse hepatobiliary reactions. It integrates a range of text mining, named entity recognition and information extraction components. LimTox relies on machine-learning, rule-based, pattern-based and term lookup strategies. This system processes scientific abstracts, a set of full text articles and medical agency assessment reports. Although the main focus of LimTox is on adverse liver events, it enables also basic searches for other organ level toxicity associations (nephrotoxicity, cardiotoxicity, thyrotoxicity and phospholipidosis). This tool supports specialized search queries for: chemical compounds/drugs, genes (with additional emphasis on key enzymes in drug metabolism, namely P450 cytochromes—CYPs) and biochemical liver markers. The LimTox website is free and open to all users and there is no login requirement. LimTox can be accessed at: http://limtox.bioinfo.cnio.es PMID:28531339
Doukas, Charalampos; Goudas, Theodosis; Fischer, Simon; Mierswa, Ingo; Chatziioannou, Aristotle; Maglogiannis, Ilias
2010-01-01
This paper presents an open image-mining framework that provides access to tools and methods for the characterization of medical images. Several image processing and feature extraction operators have been implemented and exposed through Web Services. Rapid-Miner, an open source data mining system has been utilized for applying classification operators and creating the essential processing workflows. The proposed framework has been applied for the detection of salient objects in Obstructive Nephropathy microscopy images. Initial classification results are quite promising demonstrating the feasibility of automated characterization of kidney biopsy images.
A novel web informatics approach for automated surveillance of cancer mortality trends✩
Tourassi, Georgia; Yoon, Hong-Jun; Xu, Songhua
2016-01-01
Cancer surveillance data are collected every year in the United States via the National Program of Cancer Registries (NPCR) and the Surveillance, Epidemiology and End Results (SEER) Program of the National Cancer Institute (NCI). General trends are closely monitored to measure the nation's progress against cancer. The objective of this study was to apply a novel web informatics approach for enabling fully automated monitoring of cancer mortality trends. The approach involves automated collection and text mining of online obituaries to derive the age distribution, geospatial, and temporal trends of cancer deaths in the US. Using breast and lung cancer as examples, we mined 23,850 cancer-related and 413,024 general online obituaries spanning the timeframe 2008–2012. There was high correlation between the web-derived mortality trends and the official surveillance statistics reported by NCI with respect to the age distribution (ρ = 0.981 for breast; ρ = 0.994 for lung), the geospatial distribution (ρ = 0.939 for breast; ρ = 0.881 for lung), and the annual rates of cancer deaths (ρ = 0.661 for breast; ρ = 0.839 for lung). Additional experiments investigated the effect of sample size on the consistency of the web-based findings. Overall, our study findings support web informatics as a promising, cost-effective way to dynamically monitor spatiotemporal cancer mortality trends. PMID:27044930
A novel web informatics approach for automated surveillance of cancer mortality trends
DOE Office of Scientific and Technical Information (OSTI.GOV)
Tourassi, Georgia; Yoon, Hong -Jun; Xu, Songhua
Cancer surveillance data are collected every year in the United States via the National Program of Cancer Registries (NPCR) and the Surveillance, Epidemiology and End Results (SEER) Program of the National Cancer Institute (NCI). General trends are closely monitored to measure the nation’s progress against cancer. The objective of this study was to apply a novel web informatics approach for enabling fully automated monitoring of cancer mortality trends. The approach involves automated collection and text mining of online obituaries to derive the age distribution, geospatial, and temporal trends of cancer deaths in the US. Using breast and lung cancer asmore » examples, we mined 23,850 cancer-related and 413,024 general online obituaries spanning the timeframe 2008–2012. There was high correlation between the web-derived mortality trends and the official surveillance statistics reported by NCI with respect to the age distribution (ρ = 0.981 for breast; ρ = 0.994 for lung), the geospatial distribution (ρ = 0.939 for breast; ρ = 0.881 for lung), and the annual rates of cancer deaths (ρ = 0.661 for breast; ρ = 0.839 for lung). Additional experiments investigated the effect of sample size on the consistency of the web-based findings. Altogether, our study findings support web informatics as a promising, cost-effective way to dynamically monitor spatiotemporal cancer mortality trends.« less
A novel web informatics approach for automated surveillance of cancer mortality trends
Tourassi, Georgia; Yoon, Hong -Jun; Xu, Songhua
2016-04-01
Cancer surveillance data are collected every year in the United States via the National Program of Cancer Registries (NPCR) and the Surveillance, Epidemiology and End Results (SEER) Program of the National Cancer Institute (NCI). General trends are closely monitored to measure the nation’s progress against cancer. The objective of this study was to apply a novel web informatics approach for enabling fully automated monitoring of cancer mortality trends. The approach involves automated collection and text mining of online obituaries to derive the age distribution, geospatial, and temporal trends of cancer deaths in the US. Using breast and lung cancer asmore » examples, we mined 23,850 cancer-related and 413,024 general online obituaries spanning the timeframe 2008–2012. There was high correlation between the web-derived mortality trends and the official surveillance statistics reported by NCI with respect to the age distribution (ρ = 0.981 for breast; ρ = 0.994 for lung), the geospatial distribution (ρ = 0.939 for breast; ρ = 0.881 for lung), and the annual rates of cancer deaths (ρ = 0.661 for breast; ρ = 0.839 for lung). Additional experiments investigated the effect of sample size on the consistency of the web-based findings. Altogether, our study findings support web informatics as a promising, cost-effective way to dynamically monitor spatiotemporal cancer mortality trends.« less
ERIC Educational Resources Information Center
Mills, Lane B.
2007-01-01
While many people are just getting comfortable with the Internet and e-mail, Web 2.0 technologies are already changing the playing field for education. Though definitions of Web 2.0 vary, the one constant is that Internet users are now content providers rather than content receivers. The top-down approach of the Web has been replaced with users…
A Framework for Transparently Accessing Deep Web Sources
ERIC Educational Resources Information Center
Dragut, Eduard Constantin
2010-01-01
An increasing number of Web sites expose their content via query interfaces, many of them offering the same type of products/services (e.g., flight tickets, car rental/purchasing). They constitute the so-called "Deep Web". Accessing the content on the Deep Web has been a long-standing challenge for the database community. For a user interested in…
Adverse Drug Event Discovery Using Biomedical Literature: A Big Data Neural Network Adventure
Badger, Jonathan; LaRose, Eric; Shirzadi, Ehsan; Mahnke, Andrea; Mayer, John; Ye, Zhan; Page, David; Peissig, Peggy
2017-01-01
Background The study of adverse drug events (ADEs) is a tenured topic in medical literature. In recent years, increasing numbers of scientific articles and health-related social media posts have been generated and shared daily, albeit with very limited use for ADE study and with little known about the content with respect to ADEs. Objective The aim of this study was to develop a big data analytics strategy that mines the content of scientific articles and health-related Web-based social media to detect and identify ADEs. Methods We analyzed the following two data sources: (1) biomedical articles and (2) health-related social media blog posts. We developed an intelligent and scalable text mining solution on big data infrastructures composed of Apache Spark, natural language processing, and machine learning. This was combined with an Elasticsearch No-SQL distributed database to explore and visualize ADEs. Results The accuracy, precision, recall, and area under receiver operating characteristic of the system were 92.7%, 93.6%, 93.0%, and 0.905, respectively, and showed better results in comparison with traditional approaches in the literature. This work not only detected and classified ADE sentences from big data biomedical literature but also scientifically visualized ADE interactions. Conclusions To the best of our knowledge, this work is the first to investigate a big data machine learning strategy for ADE discovery on massive datasets downloaded from PubMed Central and social media. This contribution illustrates possible capacities in big data biomedical text analysis using advanced computational methods with real-time update from new data published on a daily basis. PMID:29222076
SA-Search: a web tool for protein structure mining based on a Structural Alphabet
Guyon, Frédéric; Camproux, Anne-Claude; Hochez, Joëlle; Tufféry, Pierre
2004-01-01
SA-Search is a web tool that can be used to mine for protein structures and extract structural similarities. It is based on a hidden Markov model derived Structural Alphabet (SA) that allows the compression of three-dimensional (3D) protein conformations into a one-dimensional (1D) representation using a limited number of prototype conformations. Using such a representation, classical methods developed for amino acid sequences can be employed. Currently, SA-Search permits the performance of fast 3D similarity searches such as the extraction of exact words using a suffix tree approach, and the search for fuzzy words viewed as a simple 1D sequence alignment problem. SA-Search is available at http://bioserv.rpbs.jussieu.fr/cgi-bin/SA-Search. PMID:15215446
SA-Search: a web tool for protein structure mining based on a Structural Alphabet.
Guyon, Frédéric; Camproux, Anne-Claude; Hochez, Joëlle; Tufféry, Pierre
2004-07-01
SA-Search is a web tool that can be used to mine for protein structures and extract structural similarities. It is based on a hidden Markov model derived Structural Alphabet (SA) that allows the compression of three-dimensional (3D) protein conformations into a one-dimensional (1D) representation using a limited number of prototype conformations. Using such a representation, classical methods developed for amino acid sequences can be employed. Currently, SA-Search permits the performance of fast 3D similarity searches such as the extraction of exact words using a suffix tree approach, and the search for fuzzy words viewed as a simple 1D sequence alignment problem. SA-Search is available at http://bioserv.rpbs.jussieu.fr/cgi-bin/SA-Search.
Interactive text mining with Pipeline Pilot: a bibliographic web-based tool for PubMed.
Vellay, S G P; Latimer, N E Miller; Paillard, G
2009-06-01
Text mining has become an integral part of all research in the medical field. Many text analysis software platforms support particular use cases and only those. We show an example of a bibliographic tool that can be used to support virtually any use case in an agile manner. Here we focus on a Pipeline Pilot web-based application that interactively analyzes and reports on PubMed search results. This will be of interest to any scientist to help identify the most relevant papers in a topical area more quickly and to evaluate the results of query refinement. Links with Entrez databases help both the biologist and the chemist alike. We illustrate this application with Leishmaniasis, a neglected tropical disease, as a case study.
NASA Astrophysics Data System (ADS)
Ma, Kevin C.; Forsyth, Sydney; Amezcua, Lilyana; Liu, Brent J.
2017-03-01
We have designed and developed a multiple sclerosis eFolder system for patient data storage, image viewing, and automatic lesion quantification results to allow patient tracking. The web-based system aims to be integrated in DICOM-compliant clinical and research environments to aid clinicians in patient treatments and data analysis. The system quantifies lesion volumes, identify and register lesion locations to track shifts in volume and quantity of lesions in a longitudinal study. We aim to evaluate the two most important features of the system, data mining and longitudinal lesion tracking, to demonstrate the MS eFolder's capability in improving clinical workflow efficiency and outcome analysis for research. In order to evaluate data mining capabilities, we have collected radiological and neurological data from 72 patients, 36 Caucasian and 36 Hispanic matched by gender, disease duration, and age. Data analysis on those patients based on ethnicity is performed, and analysis results are displayed by the system's web-based user interface. The data mining module is able to successfully separate Hispanic and Caucasian patients and compare their disease profiles. For longitudinal lesion tracking, we have collected 4 longitudinal cases and simulated different lesion growths over the next year. As a result, the eFolder is able to detect changes in lesion volume and identifying lesions with the most changes. Data mining and lesion tracking evaluation results show high potential of eFolder's usefulness in patientcare and informatics research for multiple sclerosis.
Ecogeochemistry of the subsurface food web at pH 0-2.5 in Iron Mountain, California, U.S.A.
Robbins, E.I.; Rodgers, T.M.; Alpers, Charles N.; Nordstrom, D. Kirk
2000-01-01
Pyrite oxidation in the underground mining environment of Iron Mountain, California, has created the most acidic pH values ever reported in aquatic systems. Sulfate values as high as 120 000 mg l-1 and iron as high as 27 600 mg l-1 have been measured in the mine water, which also carries abundant other dissolved metals including Al, Zn, Cu, Cd, Mn, Sb and Pb. Extreme acidity and high metal concentrations apparently do not preclude the presence of an underground acidophilic food web, which has developed with bacterial biomass at the base and heliozoans as top predators. Slimes, oil-like films, flexible and inflexible stalactites, sediments, water and precipitates were found to have distinctive communities. A variety of filamentous and non-filamentous bacteria grew in slimes in water having pH values < 1.0. Fungal hyphae colonize stalactites dripping pH 1.0 water; they may help to form these drip structures. Motile hypotrichous ciliates and bdelloid rotifers are particularly abundant in slimes having a pH of 1.5. Holdfasts of the iron bacterium Leptothrix discophora attach to biofilms covering pools of standing water having a pH of 2.5 in the mine. The mine is not a closed environment - people, forced air flow and massive flushing during high intensity rainfall provide intermittent contact between the surface and underground habitats, so the mine ecosystem probably is not a restricted one.
Ecogeochemistry of the subsurface food web at pH 0–2.5 in Iron Mountain, California, U.S.A.
Robbins, Eleanora I.; Rodgers , Teresa M.; Alpers, Charles N.; Nordstrom, D. Kirk
2000-01-01
Pyrite oxidation in the underground mining environment of Iron Mountain, California, has created the most acidic pH values ever reported in aquatic systems. Sulfate values as high as 120 000 mg l−1 and iron as high as 27 600 mg l−1 have been measured in the mine water, which also carries abundant other dissolved metals including Al, Zn, Cu, Cd, Mn, Sb and Pb. Extreme acidity and high metal concentrations apparently do not preclude the presence of an underground acidophilic food web, which has developed with bacterial biomass at the base and heliozoans as top predators. Slimes, oil-like films, flexible and inflexible stalactites, sediments, water and precipitates were found to have distinctive communities. A variety of filamentous and non-filamentous bacteria grew in slimes in water having pH values <1.0. Fungal hyphae colonize stalactites dripping pH 1.0 water; they may help to form these drip structures. Motile hypotrichous ciliates and bdelloid rotifers are particularly abundant in slimes having a pH of 1.5. Holdfasts of the iron bacterium Leptothrix discophora attach to biofilms covering pools of standing water having a pH of 2.5 in the mine. The mine is not a closed environment – people, forced air flow and massive flushing during high intensity rainfall provide intermittent contact between the surface and underground habitats, so the mine ecosystem probably is not a restricted one.
Mining large heterogeneous data sets in drug discovery.
Wild, David J
2009-10-01
Increasingly, effective drug discovery involves the searching and data mining of large volumes of information from many sources covering the domains of chemistry, biology and pharmacology amongst others. This has led to a proliferation of databases and data sources relevant to drug discovery. This paper provides a review of the publicly-available large-scale databases relevant to drug discovery, describes the kinds of data mining approaches that can be applied to them and discusses recent work in integrative data mining that looks for associations that pan multiple sources, including the use of Semantic Web techniques. The future of mining large data sets for drug discovery requires intelligent, semantic aggregation of information from all of the data sources described in this review, along with the application of advanced methods such as intelligent agents and inference engines in client applications.
Assessment and distribution of antimony in soils around three coal mines, Anhui, China
Qi, C.; Liu, Gaisheng; Kang, Y.; Lam, P.K.S.; Chou, C.
2011-01-01
Thirty-three soil samples were collected from the Luling, Liuer, and Zhangji coal mines in the Huaibei and Huainan areas of Anhui Province, China. The samples were analyzed for antimony (Sb) by inductively coupled plasmaoptical emission spectrometry (ICP-OES) method. The average Sb content in the 33 samples was 4 mg kg-1, which is lower than in coals from this region (6.2 mg kg-1). More than 75% of the soils sampled showed a significant degree of Sb pollution (enrichment factors [EFs] 5-20). The soils collected near the gob pile and coal preparation plant were higher in Sb content than those collected from residential areas near the mines. The gob pile and tailings from the preparation plant were high in mineral matter content and high in Sb. They are the sources of Sb pollution in surface soils in the vicinity of coal mines. The spatial dispersion of Sb in surface soil in the mine region shows that Sb pollution could reach out as far as 350 m into the local environment conditions. Crops in rice paddies may adsorb some Sb and reduce the Sb content in soils from paddyfields. Vertical distribution of Sb in two soil profiles indicates that Sb is normally relatively immobile in soils. ?? 2011 Air & Waste Management Association.
Morrison, James J; Hostetter, Jason; Wang, Kenneth; Siegel, Eliot L
2015-02-01
Real-time mining of large research trial datasets enables development of case-based clinical decision support tools. Several applicable research datasets exist including the National Lung Screening Trial (NLST), a dataset unparalleled in size and scope for studying population-based lung cancer screening. Using these data, a clinical decision support tool was developed which matches patient demographics and lung nodule characteristics to a cohort of similar patients. The NLST dataset was converted into Structured Query Language (SQL) tables hosted on a web server, and a web-based JavaScript application was developed which performs real-time queries. JavaScript is used for both the server-side and client-side language, allowing for rapid development of a robust client interface and server-side data layer. Real-time data mining of user-specified patient cohorts achieved a rapid return of cohort cancer statistics and lung nodule distribution information. This system demonstrates the potential of individualized real-time data mining using large high-quality clinical trial datasets to drive evidence-based clinical decision-making.
Biomedical data mining in clinical routine: expanding the impact of hospital information systems.
Müller, Marcel; Markó, Kornel; Daumke, Philipp; Paetzold, Jan; Roesner, Arnold; Klar, Rüdiger
2007-01-01
In this paper we want to describe how the promising technology of biomedical data mining can improve the use of hospital information systems: a large set of unstructured, narrative clinical data from a dermatological university hospital like discharge letters or other dermatological reports were processed through a morpho-semantic text retrieval engine ("MorphoSaurus") and integrated with other clinical data using a web-based interface and brought into daily clinical routine. The user evaluation showed a very high user acceptance - this system seems to meet the clinicians' requirements for a vertical data mining in the electronic patient records. What emerges is the need for integration of biomedical data mining into hospital information systems for clinical, scientific, educational and economic reasons.
pubmed.mineR: an R package with text-mining algorithms to analyse PubMed abstracts.
Rani, Jyoti; Shah, A B Rauf; Ramachandran, Srinivasan
2015-10-01
The PubMed literature database is a valuable source of information for scientific research. It is rich in biomedical literature with more than 24 million citations. Data-mining of voluminous literature is a challenging task. Although several text-mining algorithms have been developed in recent years with focus on data visualization, they have limitations such as speed, are rigid and are not available in the open source. We have developed an R package, pubmed.mineR, wherein we have combined the advantages of existing algorithms, overcome their limitations, and offer user flexibility and link with other packages in Bioconductor and the Comprehensive R Network (CRAN) in order to expand the user capabilities for executing multifaceted approaches. Three case studies are presented, namely, 'Evolving role of diabetes educators', 'Cancer risk assessment' and 'Dynamic concepts on disease and comorbidity' to illustrate the use of pubmed.mineR. The package generally runs fast with small elapsed times in regular workstations even on large corpus sizes and with compute intensive functions. The pubmed.mineR is available at http://cran.rproject. org/web/packages/pubmed.mineR.
Capturing Trust in Social Web Applications
NASA Astrophysics Data System (ADS)
O'Donovan, John
The Social Web constitutes a shift in information flow from the traditional Web. Previously, content was provided by the owners of a website, for consumption by the end-user. Nowadays, these websites are being replaced by Social Web applications which are frameworks for the publication of user-provided content. Traditionally, Web content could be `trusted' to some extent based on the site it originated from. Algorithms such as Google's PageRank were (and still are) used to compute the importance of a website, based on analysis of underlying link topology. In the Social Web, analysis of link topology merely tells us about the importance of the information framework which hosts the content. Consumers of information still need to know about the importance/reliability of the content they are reading, and therefore about the reliability of the producers of that content. Research into trust and reputation of the producers of information in the Social Web is still very much in its infancy. Every day, people are forced to make trusting decisions about strangers on the Web based on a very limited amount of information. For example, purchasing a product from an eBay seller with a `reputation' of 99%, downloading a file from a peer-to-peer application such as Bit-Torrent, or allowing Amazon.com tell you what products you will like. Even something as simple as reading comments on a Web-blog requires the consumer to make a trusting decision about the quality of that information. In all of these example cases, and indeed throughout the Social Web, there is a pressing demand for increased information upon which we can make trusting decisions. This chapter examines the diversity of sources from which trust information can be harnessed within Social Web applications and discusses a high level classification of those sources. Three different techniques for harnessing and using trust from a range of sources are presented. These techniques are deployed in two sample Social Web applications—a recommender system and an online auction. In all cases, it is shown that harnessing an increased amount of information upon which to make trust decisions greatly enhances the user experience with the Social Web application.
Oh, Pok-Ja; Kim, Il-Ok; Shin, Sung-Rae; Jung, Hoe-Kyung
2004-10-01
This study was to develop Web-based multimedia content for Physical Examination and Health Assessment. The multimedia content was developed based on Jung's teaching and learning structure plan model, using the following 5 processes : 1) Analysis Stage, 2) Planning Stage, 3) Storyboard Framing and Production Stage, 4) Program Operation Stage, and 5) Final Evaluation Stage. The web based multimedia content consisted of an intro movie, main page and sub pages. On the main page, there were 6 menu bars that consisted of Announcement center, Information of professors, Lecture guide, Cyber lecture, Q&A, and Data centers, and a site map which introduced 15 week lectures. In the operation of web based multimedia content, HTML, JavaScript, Flash, and multimedia technology (Audio and Video) were utilized and the content consisted of text content, interactive content, animation, and audio & video. Consultation with the experts in context, computer engineering, and educational technology was utilized in the development of these processes. Web-based multimedia content is expected to offer individualized and tailored learning opportunities to maximize and facilitate the effectiveness of the teaching and learning process. Therefore, multimedia content should be utilized concurrently with the lecture in the Physical Examination and Health Assessment classes as a vital teaching aid to make up for the weakness of the face-to- face teaching-learning method.
Chu, Zhaoxia; Wang, Xingming; Wang, Yunmin; Liu, Guijian; Dong, Zhongbing; Lu, Xianwen; Chen, Guangzhou; Zha, Fugeng
2017-12-21
Copper mine tailings pose many threats to the surrounding environment and human health, and thus, their remediation is fundamental. Coal spoil is the waste by-product of coal mining and characterized by low levels of metals, high content of organic matter, and many essential microelements. This study was designed to evaluate the role of coal spoil on heavy uptake and physiological responses of Lolium perenne L. grown in copper mine tailings amended with coal spoil at rates of 0, 0.5, 1, 5, 10, and 20%. The results showed that applying coal spoil to copper mine tailings decreased the diethylenetriaminepentaacetic acid (DTPA)-extractable Cd, Cu, Pb, and Zn contents in tailings and reduced those metal contents in both roots and shoots of the plant. However, application of coal spoil increased the DTPA-extractable Cr concentration in tailings and also increased Cr uptake and accumulation by Lolium perenne L. The statistical analysis of physiological parameters indicated that chlorophyll and carotenoid increased at the lower amendments of coal spoil followed by a decrease compared to their respective controls. Protein content was enhanced at all the coal spoil amendments. When treated with coal spoil, the activities of superoxide dismutases (SOD), peroxidase (POD), and catalase (CAT) responded differently. CAT activity was inhibited, but POD activity was increased with increasing amendment ratio of coal spoil. SOD activity increased up to 1% coal spoil followed by a decrease. Overall, the addition of coal spoil decreased the oxidative stress in Lolium perenne L., reflected by the reduction in malondialdehyde (MDA) contents in the plant. It is concluded that coal spoil has the potential to stabilize most metals studied in copper mine tailings and ameliorate the harmful effects in Lolium perenne L. through changing the physiological attributes of the plant grown in copper mine tailings.
Don't Be Afraid to Explore Web 2.0
ERIC Educational Resources Information Center
Thompson, John
2008-01-01
Web 2.0 is a hot topic. The term "Web 2.0" refers to the next generation of Internet applications that allow the average Internet user to collaborate and share information online. Web 2.0 sites allow anyone to contribute content and to participate with other users in editing and even combining or remixing existing content with other material to…
ERIC Educational Resources Information Center
Henry, Anna E.; Story, Mary
2009-01-01
Objective: To identify food and beverage brand Web sites featuring designated children's areas, assess marketing techniques present on those industry Web sites, and determine nutritional quality of branded food items marketed to children. Design: Systematic content analysis of food and beverage brand Web sites and nutrient analysis of food and…
The Web and Accountability. Leaders Sharing--For Tech Leaders
ERIC Educational Resources Information Center
Deubel, Patricia
2005-01-01
A Web site can be the unifying force among schools with in a district. Many approaches work. Anyone who examines a few school district Web sites will appreciate the variation in content and design each provides. For example, Des Moines Iowa Public Schools, a district in which many in the community speak a second language, provides Web content in…
Meza-Figueroa, Diana; Maier, Raina M.; de la O-Villanueva, Margarita; Gómez-Alvarez, Agustín; Moreno-Zazueta, Alan; Rivera, Jacinto; Campillo, Alberto; Grandlic, Christopher; Anaya, Ricardo; Palafox-Reyes, Juan
2009-01-01
Past mining activities in northern Mexico left a legacy of delerict landscapes devoid of vegetation and seasonal formation of salt efflorescence. Metal content was measured in mine tailings, efflorescent salts, soils, road dust and residential soils to investigate contamination. Climatic effects such as heavy wind and rainfall events can have great impact on the dispersion of metals in semi-arid areas, since soils are typically sparsely vegetated. Geochemical analysis of this site revealed that even though total metal content in mine tailings was relatively low (e.g. Cu = 1000 mg kg-1), metals including Mn, Ba, Zn, and Cu were all found at significantly higher levels in efflorescence salts formed by evaporation on the tailings impoundment surface following the rainy season (e.g. Cu=68000 mg kg-1). Such efflorescent fine-grained salts are susceptible to wind erosion resulting in increased metal spread to nearby residential soils. Our results highlight the importance of seasonally dependent salt-formation and wind erosion in determining risk levels associated with potential inhalation or ingestion of airborne particulates originating from contaminated sites such as tailings impoundments. In low metal-content mine tailings located in arid and semi-arid environments, efflorescence salts could represent a human health risk and a challenge for plant establishment in mine tailings. PMID:19500816
Meza-Figueroa, Diana; Maier, Raina M; de la O-Villanueva, Margarita; Gómez-Alvarez, Agustín; Moreno-Zazueta, Alan; Rivera, Jacinto; Campillo, Alberto; Grandlic, Christopher J; Anaya, Ricardo; Palafox-Reyes, Juan
2009-09-01
Past mining activities in northern Mexico left a legacy of delerict landscapes devoid of vegetation and seasonal formation of salt efflorescence. Metal content was measured in mine tailings, efflorescent salts, soils, road dust, and residential soils to investigate contamination. Climatic effects such as heavy wind and rainfall events can have great impact on the dispersion of metals in semi-arid areas, since soils are typically sparsely vegetated. Geochemical analysis of this site revealed that even though total metal content in mine tailings was relatively low (e.g. Cu= 1000 mg kg(-1)), metals including Mn, Ba, Zn, and Cu were all found at significantly higher levels in efflorescence salts formed by evaporation on the tailings impoundment surface following the rainy season (e.g. Cu= 68,000 mg kg(-1)). Such efflorescent fine-grained salts are susceptible to wind erosion resulting in increased metal spread to nearby residential soils. Our results highlight the importance of seasonally dependent salt-formation and wind erosion in determining risk levels associated with potential inhalation or ingestion of airborne particulates originating from contaminated sites such as tailings impoundments. In low metal-content mine tailings located in arid and semi-arid environments, efflorescence salts could represent a human health risk and a challenge for plant establishment in mine tailings.
NASA Astrophysics Data System (ADS)
Chant, Ian J.; Staines, Geoff
1997-07-01
United Nations Peacekeeping forces around the world need to transport food, personnel and medical supplies through disputed regions were land mines are in active use as road blocks and terror weapons. A method of fast, effective land mine detection is needed to combat this threat to road transport. The technique must operate from a vehicle travelling at a reasonable velocity and give warning far enough ahead for the vehicle to stop in time to avoid the land mine. There is particular interest in detecting low- metallic content land mines. One possible solutionis the use of ultra-wide-band (UWB) radar. The Australian Defence Department is investigating the feasibility of using UWB radar for land mine detection from a vehicle. A 3 GHz UWB system has been used to collect target response from a series of inert land mines and mine-like objects placed on the ground and buried in the ground. The targets measured were a subset of those in the target set described in Wong et al. with the addition of inert land mines corresponding to some of the surrogate targets in this set. The results are encouraging for the detection of metallic land mines and the larger non-metallic land mines. Smaller low-metallic- content anti-personnel land mines are less likely to be detected.
Frank, M S; Dreyer, K
2001-06-01
We describe a virtual web site hosting technology that enables educators in radiology to emblazon and make available for delivery on the world wide web their own interactive educational content, free from dependencies on in-house resources and policies. This suite of technologies includes a graphically oriented software application, designed for the computer novice, to facilitate the input, storage, and management of domain expertise within a database system. The database stores this expertise as choreographed and interlinked multimedia entities including text, imagery, interactive questions, and audio. Case-based presentations or thematic lectures can be authored locally, previewed locally within a web browser, then uploaded at will as packaged knowledge objects to an educator's (or department's) personal web site housed within a virtual server architecture. This architecture can host an unlimited number of unique educational web sites for individuals or departments in need of such service. Each virtual site's content is stored within that site's protected back-end database connected to Internet Information Server (Microsoft Corp, Redmond WA) using a suite of Active Server Page (ASP) modules that incorporate Microsoft's Active Data Objects (ADO) technology. Each person's or department's electronic teaching material appears as an independent web site with different levels of access--controlled by a username-password strategy--for teachers and students. There is essentially no static hypertext markup language (HTML). Rather, all pages displayed for a given site are rendered dynamically from case-based or thematic content that is fetched from that virtual site's database. The dynamically rendered HTML is displayed within a web browser in a Socratic fashion that can assess the recipient's current fund of knowledge while providing instantaneous user-specific feedback. Each site is emblazoned with the logo and identification of the participating institution. Individuals with teacher-level access can use a web browser to upload new content as well as manage content already stored on their virtual site. Each virtual site stores, collates, and scores participants' responses to the interactive questions posed on line. This virtual web site strategy empowers the educator with an end-to-end solution for creating interactive educational content and hosting that content within the educator's personalized and protected educational site on the world wide web, thus providing a valuable outlet that can magnify the impact of his or her talents and contributions.
Education problems and Web-based teaching: how it impacts dental educators?
Clark, G T
2001-01-01
This article looks at six problems that vex educators and how web-based teaching might help solve them. These problems include: (1) limited access to educational content, (2) need for asynchronous access to educational content, (3) depth and diversity of educational content, (4) training in complex problem solving, (5) promotion of lifelong learning behaviors and (6) achieving excellence in education. The advantages and disadvantage of web-based educational content for each problem are discussed. The article suggests that when a poorly organized course with inaccurate and irrelevant content is placed online, it solves no problems. However some of the above issues can be partially or fully solved by hosting well-constructed teaching modules on the web. This article also reviews the literature investigating the efficacy of off-site education as compared to that provided on-site. The conclusion of this review is that teleconference-based and web-based delivery of educational content can be as effective as traditional classroom-based teaching assuming the technologic problems sometimes associated with delivering teaching content to off-site locations do not interfere in the learning process. A suggested hierarchy for rating and comparing e-learning concepts and methods is presented for consideration.
Facilitating Student-Generated Content Using Web 2.0 Technologies
ERIC Educational Resources Information Center
Lee, Eunbae
2011-01-01
Web 2.0 technologies have created a trend of user-generated content by supporting media production, collaboration, communication, and dissemination. User-generated content is translated into student-generated content (SGC) in education. SGC engages learners in an authentic project that fosters students' autonomy, creativity, and real-world…
Code of Federal Regulations, 2011 CFR
2011-07-01
... 30 Mineral Resources 3 2011-07-01 2011-07-01 false Scope. 779.1 Section 779.1 Mineral Resources OFFICE OF SURFACE MINING RECLAMATION AND ENFORCEMENT, DEPARTMENT OF THE INTERIOR SURFACE COAL MINING AND... environmental resources contents of applications for surface mining activities. ...
Code of Federal Regulations, 2010 CFR
2010-07-01
... 30 Mineral Resources 3 2010-07-01 2010-07-01 false Scope. 779.1 Section 779.1 Mineral Resources OFFICE OF SURFACE MINING RECLAMATION AND ENFORCEMENT, DEPARTMENT OF THE INTERIOR SURFACE COAL MINING AND... environmental resources contents of applications for surface mining activities. ...
Visual aspects of perception of multimedia messages on the web through the "eye tracker" method.
Svilicić, Niksa
2010-09-01
Since the dawn of civilisation visual communication played a role in everyday life. In the early times there were simply shaped drawings of animals, pictograms explaining hunting tactics or strategies of attacking the enemies. Through evolution visual expression becomes an important component of communication process on several levels, from the existential and economic level to the artistic level. However, there was always a question of the level of user reception of such visual information in the medium transmitting the information. Does physical positioning of information in the medium contribute to the efficiency of the message? Do the same rules of content positioning apply for traditional (offline) and online media (Internet)? Rapid development of information technology and Internet in almost all segments of contemporary life calls for defining the rules of designing and positioning multimedia online contents on web sites. Recent research indicates beyond doubt that the physical positioning of an online content on a web site significantly determines the quality of user's perception of such content. By employing the "Eye tracking" method it is possible to objectively analyse the level of user perception of a multimedia content on a web site. What is the first thing observed by the user after opening the web site and how does he/she visually search the online content? By which methods can this be investigated subjectively and objectively? How can the survey results be used to improve the creation of web sites and to optimise the positioning of relevant contents on the site? The answers to these questions will significantly improve the presentation of multimedia interactive contents on the Web.
XMM-Newton Mobile Web Application
NASA Astrophysics Data System (ADS)
Ibarra, A.; Kennedy, M.; Rodríguez, P.; Hernández, C.; Saxton, R.; Gabriel, C.
2013-10-01
We present the first XMM-Newton web mobile application, coded using new web technologies such as HTML5, the Query mobile framework, and D3 JavaScript data-driven library. This new web mobile application focuses on re-formatted contents extracted directly from the XMM-Newton web, optimizing the contents for mobile devices. The main goals of this development were to reach all kind of handheld devices and operating systems, while minimizing software maintenance. The application therefore has been developed as a web mobile implementation rather than a more costly native application. New functionality will be added regularly.
Mason, Alicia; Wright, Kevin B
2011-02-01
This exploratory study analyzed the content of medical tourism Web sites in an attempt to examine how they convey information about benefits and risks of medical procedures, how they frame credibility, and the degree to which these Web sites include interactive features for consumers. Drawing upon framing theory, the researchers content analyzed a sample of 66 medical tourism Web sites throughout the world. The results indicated that medical tourism Web sites largely promote the benefits of medical procedures while downplaying the risks, and relatively little information regarding the credibility of these services appears. In addition, the presentation of benefits/risks, credibility, and Web site interactivity were found to differ by region and type of facility. The authors discuss the implications of these findings concerning the framing of medical tourism Web site content, future directions for research, and limitations.
Resource Management Scheme Based on Ubiquitous Data Analysis
Lee, Heung Ki; Jung, Jaehee
2014-01-01
Resource management of the main memory and process handler is critical to enhancing the system performance of a web server. Owing to the transaction delay time that affects incoming requests from web clients, web server systems utilize several web processes to anticipate future requests. This procedure is able to decrease the web generation time because there are enough processes to handle the incoming requests from web browsers. However, inefficient process management results in low service quality for the web server system. Proper pregenerated process mechanisms are required for dealing with the clients' requests. Unfortunately, it is difficult to predict how many requests a web server system is going to receive. If a web server system builds too many web processes, it wastes a considerable amount of memory space, and thus performance is reduced. We propose an adaptive web process manager scheme based on the analysis of web log mining. In the proposed scheme, the number of web processes is controlled through prediction of incoming requests, and accordingly, the web process management scheme consumes the least possible web transaction resources. In experiments, real web trace data were used to prove the improved performance of the proposed scheme. PMID:25197692
ERIC Educational Resources Information Center
Yu, Chong Ho; Jannasch-Pennell, Angel; DiGangi, Samuel
2011-01-01
The objective of this article is to illustrate that text mining and qualitative research are epistemologically compatible. First, like many qualitative research approaches, such as grounded theory, text mining encourages open-mindedness and discourages preconceptions. Contrary to the popular belief that text mining is a linear and fully automated…
SeMPI: a genome-based secondary metabolite prediction and identification web server.
Zierep, Paul F; Padilla, Natàlia; Yonchev, Dimitar G; Telukunta, Kiran K; Klementz, Dennis; Günther, Stefan
2017-07-03
The secondary metabolism of bacteria, fungi and plants yields a vast number of bioactive substances. The constantly increasing amount of published genomic data provides the opportunity for an efficient identification of gene clusters by genome mining. Conversely, for many natural products with resolved structures, the encoding gene clusters have not been identified yet. Even though genome mining tools have become significantly more efficient in the identification of biosynthetic gene clusters, structural elucidation of the actual secondary metabolite is still challenging, especially due to as yet unpredictable post-modifications. Here, we introduce SeMPI, a web server providing a prediction and identification pipeline for natural products synthesized by polyketide synthases of type I modular. In order to limit the possible structures of PKS products and to include putative tailoring reactions, a structural comparison with annotated natural products was introduced. Furthermore, a benchmark was designed based on 40 gene clusters with annotated PKS products. The web server of the pipeline (SeMPI) is freely available at: http://www.pharmaceutical-bioinformatics.de/sempi. © The Author(s) 2017. Published by Oxford University Press on behalf of Nucleic Acids Research.
Tozzi, Alberto Eugenio; Buonuomo, Paola Sabrina; Ciofi degli Atti, Marta Luisa; Carloni, Emanuela; Meloni, Marco; Gamba, Fiorenza
2010-01-01
Information available on the Internet about immunizations may influence parents' perception about human papillomavirus (HPV) immunization and their attitude toward vaccinating their daughters. We hypothesized that the quality of information on HPV available on the Internet may vary with language and with the level of knowledge of parents. To this end we compared the quality of a sample of Web pages in Italian with a sample of Web pages in English. Five reviewers assessed the quality of Web pages retrieved with popular search engines using criteria adapted from the Good Information Practice Essential Criteria for Vaccine Safety Web Sites recommended by the World Health Organization. Quality of Web pages was assessed in the domains of accessibility, credibility, content, and design. Scores in these domains were compared through nonparametric statistical tests. We retrieved and reviewed 74 Web sites in Italian and 117 in English. Most retrieved Web pages (33.5%) were from private agencies. Median scores were higher in Web pages in English compared with those in Italian in the domain of accessibility (p < .01), credibility (p < .01), and content (p < .01). The highest credibility and content scores were those of Web pages from governmental agencies or universities. Accessibility scores were positively associated with content scores (p < .01) and with credibility scores (p < .01). A total of 16.2% of Web pages in Italian opposed HPV immunization compared with 6.0% of those in English (p < .05). Quality of information and number of Web pages opposing HPV immunization may vary with the Web site language. High-quality Web pages on HPV, especially from public health agencies and universities, should be easily accessible and retrievable with common Web search engines. Copyright 2010 Society for Adolescent Medicine. Published by Elsevier Inc. All rights reserved.
KnowledgePuzzle: A Browsing Tool to Adapt the Web Navigation Process to the Learner's Mental Model
ERIC Educational Resources Information Center
AlAgha, Iyad
2012-01-01
This article presents KnowledgePuzzle, a browsing tool for knowledge construction from the web. It aims to adapt the structure of web content to the learner's information needs regardless of how the web content is originally delivered. Learners are provided with a meta-cognitive space (e.g., a concept mapping tool) that enables them to plan…
Web-Based Media Contents Editor for UCC Websites
NASA Astrophysics Data System (ADS)
Kim, Seoksoo
The purpose of this research is to "design web-based media contents editor for establishing UCC(User Created Contents)-based websites." The web-based editor features user-oriented interfaces and increased convenience, significantly different from previous off-line editors. It allows users to edit media contents online and can be effectively used for online promotion activities of enterprises and organizations. In addition to development of the editor, the research aims to support the entry of enterprises and public agencies to the online market by combining the technology with various UCC items.
QuakeSim: a Web Service Environment for Productive Investigations with Earth Surface Sensor Data
NASA Astrophysics Data System (ADS)
Parker, J. W.; Donnellan, A.; Granat, R. A.; Lyzenga, G. A.; Glasscoe, M. T.; McLeod, D.; Al-Ghanmi, R.; Pierce, M.; Fox, G.; Grant Ludwig, L.; Rundle, J. B.
2011-12-01
The QuakeSim science gateway environment includes a visually rich portal interface, web service access to data and data processing operations, and the QuakeTables ontology-based database of fault models and sensor data. The integrated tools and services are designed to assist investigators by covering the entire earthquake cycle of strain accumulation and release. The Web interface now includes Drupal-based access to diverse and changing content, with new ability to access data and data processing directly from the public page, as well as the traditional project management areas that require password access. The system is designed to make initial browsing of fault models and deformation data particularly engaging for new users. Popular data and data processing include GPS time series with data mining techniques to find anomalies in time and space, experimental forecasting methods based on catalogue seismicity, faulted deformation models (both half-space and finite element), and model-based inversion of sensor data. The fault models include the CGS and UCERF 2.0 faults of California and are easily augmented with self-consistent fault models from other regions. The QuakeTables deformation data include the comprehensive set of UAVSAR interferograms as well as a growing collection of satellite InSAR data.. Fault interaction simulations are also being incorporated in the web environment based on Virtual California. A sample usage scenario is presented which follows an investigation of UAVSAR data from viewing as an overlay in Google Maps, to selection of an area of interest via a polygon tool, to fast extraction of the relevant correlation and phase information from large data files, to a model inversion of fault slip followed by calculation and display of a synthetic model interferogram.
Feris, Kevin; Ramsey, Philip; Frazar, Chris; Moore, Johnnie N.; Gannon, James E.; Holben, William E.
2003-01-01
The hyporheic zone of a river is nonphotic, has steep chemical and redox gradients, and has a heterotrophic food web based on the consumption of organic carbon entrained from downwelling surface water or from upwelling groundwater. The microbial communities in the hyporheic zone are an important component of these heterotrophic food webs and perform essential functions in lotic ecosystems. Using a suite of methods (denaturing gradient gel electrophoresis, 16S rRNA phylogeny, phospholipid fatty acid analysis, direct microscopic enumeration, and quantitative PCR), we compared the microbial communities inhabiting the hyporheic zone of six different river sites that encompass a wide range of sediment metal loads resulting from large base-metal mining activity in the region. There was no correlation between sediment metal content and the total hyporheic microbial biomass present within each site. However, microbial community structure showed a significant linear relationship with the sediment metal loads. The abundances of four phylogenetic groups (groups I, II, III, and IV) most closely related to α-, β-, and γ-proteobacteria and the cyanobacteria, respectively, were determined. The sediment metal content gradient was positively correlated with group III abundance and negatively correlated with group II abundance. No correlation was apparent with regard to group I or IV abundance. This is the first documentation of a relationship between fluvially deposited heavy-metal contamination and hyporheic microbial community structure. The information presented here may be useful in predicting long-term effects of heavy-metal contamination in streams and provides a basis for further studies of metal effects on hyporheic microbial communities. PMID:12957946
Calypso: a user-friendly web-server for mining and visualizing microbiome-environment interactions.
Zakrzewski, Martha; Proietti, Carla; Ellis, Jonathan J; Hasan, Shihab; Brion, Marie-Jo; Berger, Bernard; Krause, Lutz
2017-03-01
Calypso is an easy-to-use online software suite that allows non-expert users to mine, interpret and compare taxonomic information from metagenomic or 16S rDNA datasets. Calypso has a focus on multivariate statistical approaches that can identify complex environment-microbiome associations. The software enables quantitative visualizations, statistical testing, multivariate analysis, supervised learning, factor analysis, multivariable regression, network analysis and diversity estimates. Comprehensive help pages, tutorials and videos are provided via a wiki page. The web-interface is accessible via http://cgenome.net/calypso/ . The software is programmed in Java, PERL and R and the source code is available from Zenodo ( https://zenodo.org/record/50931 ). The software is freely available for non-commercial users. l.krause@uq.edu.au. Supplementary data are available at Bioinformatics online. © The Author 2016. Published by Oxford University Press.
Impacts of gold mine waste disposal on a tropical pelagic ecosystem.
Brewer, D T; Morello, E B; Griffiths, S; Fry, G; Heales, D; Apte, S C; Venables, W N; Rothlisberg, P C; Moeseneder, C; Lansdell, M; Pendrey, R; Coman, F; Strzelecki, J; Jarolimek, C V; Jung, R F; Richardson, A J
2012-12-01
We used a comparative approach to investigate the impact of the disposal of gold mine tailings into the ocean near the Lihir mine (Niolam Island, Papua New Guinea). We found abundance and diversity of zooplankton, micronekton and pelagic fish to be similar or higher in the mine region compared to the reference site. We also found relatively high trace metal concentrations in lower trophic level groups, especially zooplankton, near the mine discharge, but few differences in tissue concentrations of micronekton, baitfish and pelagic fish between the two regions. Biomagnification of some trace metals by micronekton, and of mercury by fish was evident in both regions. We conclude that ocean mine waste disposal at Niolam Island has a local impact on the smaller and less mobile pelagic communities in terms of trace metal concentrations, but has little effect on the abundance and biodiversity of the local food web. Crown Copyright © 2012. Published by Elsevier Ltd. All rights reserved.
NASA Astrophysics Data System (ADS)
Meseguer, S.; Sanfeliu, T.; Jordán, M. M.
2009-02-01
The Oliete basin (Early Cretaceous, NE Teruel, Spain) is one of the most important areas for the supply of mine spoils used as ball clays for the production of white and red stoneware in the Spanish ceramic industry of wall and floor tiles. This study corresponds to the second part of the paper published recently by Meseguer et al. (Environ Geol 2008) about the use of mine spoils from Teruel coal mining district. The present study shows a statistical data analysis from chemical data (major, minor and trace elements). The performed statistical analysis of chemical data included descriptive statistics and cluster analysis (with ANOVA and Scheffé methods). The cluster analysis of chemical data provided three main groups: C3 with the highest mean SiO2 content (66%) and lowest mean Al2O3 content (20%); C2 with lower SiO2 content (48%) and higher mean Al2O3 content (28%); and C1 with medium values for the SiO2 and Al2O3 mean content. The main applications of these materials are refractory, white and red ceramics, stoneware, heavy ceramics (including red earthenware, bricks and roof tiles), and components of white Portland cement and aluminous cement. Clays from group 2 are used in refractories (with higher kaolinite content, and constrictions to CaO + MgO and K2O + Na2O contents). All materials can be used in fine ceramics (white or red, according to the Fe2O3 + TiO2 content).
Accredited hand surgery fellowship Web sites: analysis of content and accessibility.
Trehan, Samir K; Morrell, Nathan T; Akelman, Edward
2015-04-01
To assess the accessibility and content of accredited hand surgery fellowship Web sites. A list of all accredited hand surgery fellowships was obtained from the online database of the American Society for Surgery of the Hand (ASSH). Fellowship program information on the ASSH Web site was recorded. All fellowship program Web sites were located via Google search. Fellowship program Web sites were analyzed for accessibility and content in 3 domains: program overview, application information/recruitment, and education. At the time of this study, there were 81 accredited hand surgery fellowships with 169 available positions. Thirty of 81 programs (37%) had a functional link on the ASSH online hand surgery fellowship directory; however, Google search identified 78 Web sites. Three programs did not have a Web site. Analysis of content revealed that most Web sites contained contact information, whereas information regarding the anticipated clinical, research, and educational experiences during fellowship was less often present. Furthermore, information regarding past and present fellows, salary, application process/requirements, call responsibilities, and case volume was frequently lacking. Overall, 52 of 81 programs (64%) had the minimal online information required for residents to independently complete the fellowship application process. Hand fellowship program Web sites could be accessed either via the ASSH online directory or Google search, except for 3 programs that did not have Web sites. Although most fellowship program Web sites contained contact information, other content such as application information/recruitment and education, was less frequently present. This study provides comparative data regarding the clinical and educational experiences outlined on hand fellowship program Web sites that are of relevance to residents, fellows, and academic hand surgeons. This study also draws attention to various ways in which the hand surgery fellowship application process can be made more user-friendly and efficient. Copyright © 2015 American Society for Surgery of the Hand. Published by Elsevier Inc. All rights reserved.
Future View: Web Navigation based on Learning User's Browsing Strategy
NASA Astrophysics Data System (ADS)
Nagino, Norikatsu; Yamada, Seiji
In this paper, we propose a Future View system that assists user's usual Web browsing. The Future View will prefetch Web pages based on user's browsing strategies and present them to a user in order to assist Web browsing. To learn user's browsing strategy, the Future View uses two types of learning classifier systems: a content-based classifier system for contents change patterns and an action-based classifier system for user's action patterns. The results of learning is applied to crawling by Web robots, and the gathered Web pages are presented to a user through a Web browser interface. We experimentally show effectiveness of navigation using the Future View.
State of remediation and metal toxicity in the Tri-State Mining District, USA.
Johnson, Aaron W; Gutiérrez, Mélida; Gouzie, Douglas; McAliley, L Rex
2016-02-01
Mining operations in the Tri-State Mining District of Kansas, Missouri and Oklahoma (TSMD), once one of the major lead and zinc mining areas in the world, had completely ceased by 1970. As mining companies moved out, the land was left with underground tunnels and mine shafts and the surface with abandoned tailings piles, which progressively contaminated groundwater and soil. Despite remedial actions undertaken in the 1980's, areas within the TSMD still contain Cd, Pb, and Zn concentrations exceeding safe levels. Because of the large area and highly dispersed occurrence of wastes, environmental studies generally have been confined either to a stream basin or to a single state. Studies also have differed in their approach and analytical methodologies. An overview of the totality of the TSMD and its present state of contamination is presented here. Data show that metal content in sediments have the following common features: (1) a wide range of Pb and Zn concentrations, up to three orders of magnitude, (2) median values for Cd, Pb and Zn content in sediments and soils were similar among studies, (3) median values for most studies were at or above the guidelines recommended for aquatic habitats, and (4) highest content of Pb and Zn were closely associated with the geographical location of former mining and smelting centers. The above observations imply that mine wastes remain a problem and further remediation is needed. Cost-effective remedial alternatives for this area's geology, climate, and land use, are discussed. Copyright © 2015 Elsevier Ltd. All rights reserved.
Contextual advertisement placement in printed media
NASA Astrophysics Data System (ADS)
Liu, Sam; Joshi, Parag
2010-02-01
Advertisements today provide the necessary revenue model supporting the WWW ecosystem. Targeted or contextual ad insertion plays an important role in optimizing the financial return of this model. Nearly all the current ads that appear on web sites are geared for display purposes such as banner and "pay-per-click". Little attention, however, is focused on deriving additional ad revenues when the content is repurposed for alternative mean of presentation, e.g. being printed. Although more and more content is moving to the Web, there are still many occasions where printed output of web content is desirable, such as maps and articles; thus printed ad insertion can potentially be lucrative. In this paper, we describe a contextual ad insertion network aimed to realize new revenue for print service providers for web printing. We introduce a cloud print service that enables contextual ads insertion, with respect to the main web page content, when a printout of the page is requested. To encourage service utilization, it would provide higher quality printouts than what is possible from current browser print drivers, which generally produce poor outputs, e.g. ill formatted pages. At this juncture we will limit the scope to only article-related web pages although the concept can be extended to arbitrary web pages. The key components of this system include (1) the extraction of article from web pages, (2) the extraction of semantics from article, (3) querying the ad database for matching advertisement or coupon, and (4) joint content and ad layout for print outputs.
Code of Federal Regulations, 2010 CFR
2010-07-01
... 30 Mineral Resources 3 2010-07-01 2010-07-01 false Scope. 783.1 Section 783.1 Mineral Resources OFFICE OF SURFACE MINING RECLAMATION AND ENFORCEMENT, DEPARTMENT OF THE INTERIOR SURFACE COAL MINING AND... for the environmental resources contents of applications for permits for underground mining activities. ...
Code of Federal Regulations, 2011 CFR
2011-07-01
... 30 Mineral Resources 3 2011-07-01 2011-07-01 false Scope. 783.1 Section 783.1 Mineral Resources OFFICE OF SURFACE MINING RECLAMATION AND ENFORCEMENT, DEPARTMENT OF THE INTERIOR SURFACE COAL MINING AND... for the environmental resources contents of applications for permits for underground mining activities. ...
Stochastic production phase design for an open pit mining complex with multiple processing streams
NASA Astrophysics Data System (ADS)
Asad, Mohammad Waqar Ali; Dimitrakopoulos, Roussos; van Eldert, Jeroen
2014-08-01
In a mining complex, the mine is a source of supply of valuable material (ore) to a number of processes that convert the raw ore to a saleable product or a metal concentrate for production of the refined metal. In this context, expected variation in metal content throughout the extent of the orebody defines the inherent uncertainty in the supply of ore, which impacts the subsequent ore and metal production targets. Traditional optimization methods for designing production phases and ultimate pit limit of an open pit mine not only ignore the uncertainty in metal content, but, in addition, commonly assume that the mine delivers ore to a single processing facility. A stochastic network flow approach is proposed that jointly integrates uncertainty in supply of ore and multiple ore destinations into the development of production phase design and ultimate pit limit. An application at a copper mine demonstrates the intricacies of the new approach. The case study shows a 14% higher discounted cash flow when compared to the traditional approach.
Young, Bradley L; Cantrell, Colin K; Patt, Joshua C; Ponce, Brent A
2018-06-01
Accessible, adequate online information is important to fellowship applicants. Program web sites can affect which programs applicants apply to, subsequently altering interview costs incurred by both parties and ultimately impacting rank lists. Web site analyses have been performed for all orthopaedic subspecialties other than those involved in the combined adult reconstruction and musculoskeletal (MSK) oncology fellowship match. A complete list of active programs was obtained from the official adult reconstruction and MSK oncology society web sites. Web site accessibility was assessed using a structured Google search. Accessible web sites were evaluated based on 21 previously reported content criteria. Seventy-four adult reconstruction programs and 11 MSK oncology programs were listed on the official society web sites. Web sites were identified and accessible for 58 (78%) adult reconstruction and 9 (82%) MSK oncology fellowship programs. No web site contained all content criteria and more than half of both adult reconstruction and MSK oncology web sites failed to include 12 of the 21 criteria. Several programs participating in the combined Adult Reconstructive Hip and Knee/Musculoskeletal Oncology Fellowship Match did not have accessible web sites. Of the web sites that were accessible, none contained comprehensive information and the majority lacked information that has been previously identified as being important to perspective applicants.
The effect and value of a WebQuest activity on weather in a 5th grade classroom
NASA Astrophysics Data System (ADS)
Oliver, Deborah
WebQuests are increasing in popularity across the country, yet it remains unclear whether WebQuests confer a significant benefit in student content learning. In addition, the perceptions of teachers regarding the classroom value and efficacy of WebQuests in teaching higher level thinking skills are still unclear. The goals of the study were (a) to determine the effect of WebQuests on elementary students' content area knowledge gains; (b) to investigate teacher perceptions of students' higher order thinking skills while engaged in a WebQuest activity and the value the of the WebQuest, as perceived by teacher. To accomplish the above research goals, a quasi-experimental design was used in this study. The subjects were four teachers and classes at an elementary school in southern California. Results of the study showed that the WebQuest did increase content knowledge in fifth grade students, but not significantly more than traditional teaching as measured by a researcher-modified WebQuest For Teacher Questionnaire (WQFT) (Zheng, Perez, Williamson & Flygare, 2007) and teacher interviews. Teachers responded positively to the value of the WebQuest in their daily teaching. Teachers also indicated that their students engaged in higher level thinking skills while engaged in the WebQuest activity. Keywords: WebQuest, higher level thinking, learning
2009-06-01
capabilities: web-based, relational/multi-dimensional, client/server, and metadata (data about data) inclusion (pp. 39-40). Text mining, on the other...and Organizational Systems ( CASOS ) (Carley, 2005). Although AutoMap can be used to conduct text-mining, it was utilized only for its visualization...provides insight into how the GMCOI is using the terms, and where there might be redundant terms and need for de -confliction and standardization
FlyMine: an integrated database for Drosophila and Anopheles genomics
Lyne, Rachel; Smith, Richard; Rutherford, Kim; Wakeling, Matthew; Varley, Andrew; Guillier, Francois; Janssens, Hilde; Ji, Wenyan; Mclaren, Peter; North, Philip; Rana, Debashis; Riley, Tom; Sullivan, Julie; Watkins, Xavier; Woodbridge, Mark; Lilley, Kathryn; Russell, Steve; Ashburner, Michael; Mizuguchi, Kenji; Micklem, Gos
2007-01-01
FlyMine is a data warehouse that addresses one of the important challenges of modern biology: how to integrate and make use of the diversity and volume of current biological data. Its main focus is genomic and proteomics data for Drosophila and other insects. It provides web access to integrated data at a number of different levels, from simple browsing to construction of complex queries, which can be executed on either single items or lists. PMID:17615057
ERIC Educational Resources Information Center
Obilade, Titilola T.; Burton, John K.
2015-01-01
This textual content analysis set out to determine the extent to which the theories, principles, and guidelines in 4 standard books of instructional design and technology were also addressed in 4 popular books on web design. The standard books on instructional design and the popular books on web design were chosen by experts in the fields. The…
Cañada, Andres; Capella-Gutierrez, Salvador; Rabal, Obdulia; Oyarzabal, Julen; Valencia, Alfonso; Krallinger, Martin
2017-07-03
A considerable effort has been devoted to retrieve systematically information for genes and proteins as well as relationships between them. Despite the importance of chemical compounds and drugs as a central bio-entity in pharmacological and biological research, only a limited number of freely available chemical text-mining/search engine technologies are currently accessible. Here we present LimTox (Literature Mining for Toxicology), a web-based online biomedical search tool with special focus on adverse hepatobiliary reactions. It integrates a range of text mining, named entity recognition and information extraction components. LimTox relies on machine-learning, rule-based, pattern-based and term lookup strategies. This system processes scientific abstracts, a set of full text articles and medical agency assessment reports. Although the main focus of LimTox is on adverse liver events, it enables also basic searches for other organ level toxicity associations (nephrotoxicity, cardiotoxicity, thyrotoxicity and phospholipidosis). This tool supports specialized search queries for: chemical compounds/drugs, genes (with additional emphasis on key enzymes in drug metabolism, namely P450 cytochromes-CYPs) and biochemical liver markers. The LimTox website is free and open to all users and there is no login requirement. LimTox can be accessed at: http://limtox.bioinfo.cnio.es. © The Author(s) 2017. Published by Oxford University Press on behalf of Nucleic Acids Research.
Demir, Yasemin; Gozum, Sebahat
2015-09-01
This study was designed to evaluate the quality, content, usability, and efficacy of a Web site prepared for the purpose of improving the caregiving capability of family members who provide care for stroke survivors at home. The DISCERN score for the Web site was found to be 4.35 over 5. The first section that assesses reliability of the Web site was 4.38 over 5; mean score of the second section that measures the quality of the provided information on treatment/care options was 4.30, and mean score of the third section that gives a general evaluation of the material was 4.1. The Web site content achieved an average score of 3.47 over 4 after evaluation by experts. The Web site system usability score was found to be 79.4 over 100. The Web site was utilized mostly for exercises in bed (76.3%; n = 29), use of medications, and patient safety (68.4%; n = 26). It was determined that those who were younger and employed and had no previous experience of nursing any patient utilized relatively more from the section of patient nutrition and oral care and married family caregivers from the body hygiene section. The Web site quality and content were judged to be good and reliable to use. The Web site was efficiently used by caregivers.
Crafting a positive professional digital profile to augment your practice
Kraakevik, Jeff
2016-01-01
Abstract A digital profile is the sum content about a person on the Internet. A digital profile can be composed of personal or professional information shared on public Web sites posted personally or by others. One of the most effective ways to build a positive professional digital profile is through social media. It is increasingly important to maintain a positive digital profile as others mine the Internet to find out about a professional prior to meeting him or her. As the digital environment continues to grow, it will become increasingly difficult to neglect a professional digital profile without potential negative consequences. There are many benefits to creating a digital presence and using the tools available to learn about neurology and interact with other professionals and patients in ways that were not possible in the past. The spread of social media to a large part of the population makes it unlikely to go away. PMID:29443275
Crafting a positive professional digital profile to augment your practice.
Kraakevik, Jeff
2016-02-01
A digital profile is the sum content about a person on the Internet. A digital profile can be composed of personal or professional information shared on public Web sites posted personally or by others. One of the most effective ways to build a positive professional digital profile is through social media. It is increasingly important to maintain a positive digital profile as others mine the Internet to find out about a professional prior to meeting him or her. As the digital environment continues to grow, it will become increasingly difficult to neglect a professional digital profile without potential negative consequences. There are many benefits to creating a digital presence and using the tools available to learn about neurology and interact with other professionals and patients in ways that were not possible in the past. The spread of social media to a large part of the population makes it unlikely to go away.
EPA Monthly Key Performance Indicator Dashboards 2017
Each month, the Web Analytics Program posts updated Key Performance Indicator (KPI) dashboards that correspond to three Web performance goals: content consumption, content discovery, and audience engagement.
Quality of Web-based information on cocaine addiction.
Khazaal, Yasser; Chatton, Anne; Cochand, Sophie; Zullino, Daniele
2008-08-01
To evaluate the quality of web-based information on cocaine use and addiction and to investigate potential content quality indicators. Three keywords: cocaine, cocaine addiction and cocaine dependence were entered into two popular World Wide Web search engines. Websites were assessed with a standardized proforma designed to rate sites on the basis of accountability, presentation, interactivity, readability and content quality. "Health on the Net" (HON) quality label, and DISCERN scale scores aiding people without content expertise to assess quality of written health publication were used to verify their efficiency as quality indicators. Of the 120 websites identified, 61 were included. Most were commercial sites. The results of the study indicate low scores on each of the measures including content quality. A global score (the sum of accountability, interactivity, content quality and aesthetic criteria) appeared as a good content quality indicator. While cocaine education websites for patients are widespread, their global quality is poor. There is a need for better evidence-based information about cocaine use and addiction on the web. The poor and variable quality of web-based information and its possible impact on physician-patient relationship argue for a serious provider for patient talk about the health information found on Internet. Internet sites could improve their content using the global score as a quality indicator.
Web 1.0 to Web 3.0 Evolution: Reviewing the Impacts on Tourism Development and Opportunities
NASA Astrophysics Data System (ADS)
Eftekhari, M. Hossein; Barzegar, Zeynab; Isaai, M. T.
The most important event following the establishmenet of the Internet network was the Web introduced by Tim Berners-Lee. Websites give their owners features that allow sharing with which they can publish their content with users and visitors. In the last 5 years, we have seen some changes in the use of web. Users want to participate in content sharing and they like to interact with each other. This is known as Web 2.0. In the last year, Web 2.0 has reached maturity and now we need a smart web which will be accordingly be called Web 3.0. Web 3.0 is based on semantic web definition. Changing the way of using the web has had a clear impact on E-Tourism and its development and also on business models. In this paper, we review the definitions and describe the impacts of web evolution on E-Tourism.
Rodríguez Martín, José Antonio; Gutiérrez, Carmen; Escuer, Miguel; García-González, Ma Teresa; Campos-Herrera, Raquel; Aguila, Nancy
2014-03-01
The Cartagena-La Union mining district, exploited since the end of the 3rd century BC, was one of the world's largest lead producers in the 19th century. Although activity ceased in 1991, today mining residues pose a huge pollution problem. This study characterises lead contents (total and DPTA) and other soil parameters (N, P, K, pH, SOM, CaCO3, granulometric fraction, etc.) using multivariate geostatistical methods in relation to nematode diversity. In this work, trophic groups and metabolic footprints of soil nematodes were measured using 193 samples from the mining, natural and agricultural areas in this district. We explored the relationship between soil health and nematode communities. High lead concentrations were quantified: mean 8,500 mg kg(-1) for total and 340 mg kg(-1) for DPTA in this mining area. Although nematode diversity was broad (81 taxa), their diversity, abundance and metabolic footprints significantly reduced in the mining area. Significant differences in the nematode community structure were observed, and the relative abundance of predators was sensitive to mine and agricultural activities, whilst omnivores reduced only in the agricultural area, and bacterial feeders exhibited a differential response to both anthropogenic disturbances. The total abundance of nematodes, trophic groups and c-p groups correlated negatively with soil Pb contents, and a positive relationship was found with SOM and N, P and K contents. Copyright © 2013 Elsevier B.V. All rights reserved.
Märten, Arno; Berger, Dietrich; Köhler, Mirko; Merten, Dirk
2015-12-01
We reconstructed the contamination history of an area influenced by 40 years of uranium mining and subsequent remediation actions using dendroanalysis (i.e., the determination of the elemental content of tree rings). The uranium content in the tree rings of four individual oak trees (Quercus sp.) was determined by laser ablation with inductively coupled plasma mass spectrometry (LA-ICP-MS). This technique allows the investigation of trace metals in solid samples with a spatial resolution of 250 μm and a detection limit below 0.01 μg/g for uranium. The investigations show that in three of the four oaks sampled, there were temporally similar uranium concentrations. These were approximately 2 orders of magnitude higher (0.15 to 0.4 μg/g) than those from before the period of active mining (concentrations below 0.01 μg/g). After the mining was terminated and the area was restored, the uranium contents in the wood decreased by approximately 1 order of magnitude. The similar radial uranium distribution patterns of the three trees were confirmed by correlation analysis. In combination with the results of soil analyses, it was determined that there was a heterogeneous contamination in the forest investigated. This could be confirmed by pre-remediation soil uranium contents from literature. The uranium contents in the tree rings of the oaks investigated reflect the contamination history of the study area. This study demonstrates that the dendrochemical analysis of oak tree rings is a suitable technique for investigating past and recent uranium contamination in mining areas.
Data Mining Meets HCI: Making Sense of Large Graphs
2012-07-01
graph algo- rithms, won the Open Source Software World Challenge, Silver Award. We have released Pegasus as free , open-source software, downloaded by...METIS [77], spectral clustering [108], and the parameter- free “Cross-associations” (CA) [26]. Belief Propagation can also be used for clus- tering, as...number of tools have been developed to support “ landscape ” views of information. These include WebBook and Web- Forager [23], which use a book metaphor
From in-situ coal to fly ash: A study of coal mines and power plants from Indiana
Mastalerz, Maria; Hower, J.C.; Drobniak, A.; Mardon, S.M.; Lis, G.
2004-01-01
This paper presents data on the properties of coal and fly ash from two coal mines and two power plants that burn single-source coal from two mines in Indiana. One mine is in the low-sulfur (5%) Springfield Coal Member of the Petersburg Formation (Pennsylvanian). Both seams have comparable ash contents (???11%). Coals sampled at the mines (both raw and washed fractions) were analyzed for proximate/ultimate/sulfur forms/heating value, major oxides, trace elements and petrographic composition. The properties of fly ash from these coals reflect the properties of the feed coal, as well as local combustion and post-combustion conditions. Sulfur and spinel content, and As, Pb and Zn concentrations of the fly ash are the parameters that most closely reflect the properties of the source coal. ?? 2004 Elsevier B.V. All rights reserved.
Deriving novel relationships from the scientific literature is an important adjunct to datamining activities for complex datasets in genomics and high-throughput screening activities. Automated text-mining algorithms can be used to extract relevant content from the literature and...
An experiment with content distribution methods in touchscreen mobile devices.
Garcia-Lopez, Eva; Garcia-Cabot, Antonio; de-Marcos, Luis
2015-09-01
This paper compares the usability of three different content distribution methods (scrolling, paging and internal links) in touchscreen mobile devices as means to display web documents. Usability is operationalized in terms of effectiveness, efficiency and user satisfaction. These dimensions are then measured in an experiment (N = 23) in which users are required to find words in regular-length web documents. Results suggest that scrolling is statistically better in terms of efficiency and user satisfaction. It is also found to be more effective but results were not significant. Our findings are also compared with existing literature to propose the following guideline: "try to use vertical scrolling in web pages for mobile devices instead of paging or internal links, except when the content is too large, then paging is recommended". With an ever increasing number of touchscreen web-enabled mobile devices, this new guideline can be relevant for content developers targeting the mobile web as well as institutions trying to improve the usability of their content for mobile platforms. Copyright © 2015 Elsevier Ltd and The Ergonomics Society. All rights reserved.
The Library Web Site: Collaborative Content Creation and Management
ERIC Educational Resources Information Center
Slater, Robert
2008-01-01
Oakland University's Kresge Library first launched its Web site in 1996. The initial design and subsequent contributions were originally managed by a single Webmaster. In 2002, the library restructured its Web content creation and management to a distributed, collaborative method with the goal of increasing the amount, accuracy, and timeliness of…
Assessment of Web Content Accessibility Levels in Spanish Official Online Education Environments
ERIC Educational Resources Information Center
Roig-Vila, Rosabel; Ferrández, Sergio; Ferri-Miralles, Imma
2014-01-01
Diversity-based designing, or the goal of ensuring that web-based information is accessible to as many diverse users as possible, has received growing international acceptance in recent years, with many countries introducing legislation to enforce it. This paper analyses web content accessibility levels in Spanish education portals according to…
Ancient Uses of Meteoritic Metals as Precedent for Modern In-Situ Asteroid Mining
NASA Astrophysics Data System (ADS)
Mardon, Austin A.; Fawcett, Brett; Krispin, Daniel
2016-05-01
Given the strain on earth's supply of metal and the meteoritic content of meteorites, a prudent course would be to pursue in-situ asteroid mining of meteors for metal. There is a precedent for this going back to ancient Egypt; humans have always used the meteoritic content of meteorites to fashion everything from weapons to cosmetics.
Jin, Shu-Lan; Huang, Yi-Zong; Wang, Fei; Xu, Feng; Wang, Xiao-Ling; Gao, Zhu; Hu, Ying; Qiao Min; Li, Jin; Xiang, Meng
2015-03-01
Rare earth elements content in farmland soils and crops of the surrounding copper mining and smelting plant in Jiangxi province was studied. The results showed that copper mining and smelting could increase the content of rare earth elements in soils and crops. Rare earth elements content in farmland soils of the surrounding Yinshan Lead Zinc Copper Mine and Guixi Smelting Plant varied from 112.42 to 397.02 mg x kg(-1) and 48.81 to 250.06 mg x kg(-1), and the average content was 254.84 mg x kg(-1) and 144.21 mg x kg(-1), respectively. The average contents of rare earth elements in soils in these two areas were 1.21 times and 0.68 times of the background value in Jiangxi province, 1.36 times and 0.77 times of the domestic background value, 3.59 times and 2.03 times of the control samples, respectively. Rare earth elements content in 10 crops of the surrounding Guixi Smelting Plant varied from 0.35 to 2.87 mg x kg(-1). The contents of rare earth elements in the leaves of crops were higher than those in stem and root. The contents of rare earth elements in Tomato, lettuce leaves and radish leaves were respectively 2.87 mg x kg(-1), 1.58 mg x kg(-1) and 0.80 mg x kg(-1), which were well above the hygienic standard limit of rare earth elements in vegetables and fruits (0.70 mg x kg(-1)). According to the health risk assessment method recommended by America Environmental Protection Bureau (USEPA), we found that the residents' lifelong average daily intake of rare earth elements was 17.72 mg x (kg x d)(-1), lower than the critical value of rare earth elements damage to human health. The results suggested that people must pay attention to the impact of rare earth elements on the surrounding environment when they mine and smelt copper ore in Jiangxi.
Legaz-García, María del Carmen; Martínez-Costa, Catalina; Menárguez-Tortosa, Marcos; Fernández-Breis, Jesualdo Tomás
2012-01-01
Linking Electronic Healthcare Records (EHR) content to educational materials has been considered a key international recommendation to enable clinical engagement and to promote patient safety. This would suggest citizens to access reliable information available on the web and to guide them properly. In this paper, we describe an approach in that direction, based on the use of dual model EHR standards and standardized educational contents. The recommendation method will be based on the semantic coverage of the learning content repository for a particular archetype, which will be calculated by applying semantic web technologies like ontologies and semantic annotations.
Turning Interoperability Operational with GST
NASA Astrophysics Data System (ADS)
Schaeben, Helmut; Gabriel, Paul; Gietzel, Jan; Le, Hai Ha
2013-04-01
GST - Geosciences in space and time is being developed and implemented as hub to facilitate the exchange of spatially and temporally indexed multi-dimensional geoscience data and corresponding geomodels amongst partners. It originates from TUBAF's contribution to the EU project "ProMine" and its perspective extensions are TUBAF's contribution to the actual EU project "GeoMol". As of today, it provides basic components of a geodata infrastructure as required to establish interoperability with respect to geosciences. Generally, interoperability means the facilitation of cross-border and cross-sector information exchange, taking into account legal, organisational, semantic and technical aspects, cf. Interoperability Solutions for European Public Administrations (ISA), cf. http://ec.europa.eu/isa/. Practical interoperability for partners of a joint geoscience project, say European Geological Surveys acting in a border region, means in particular provision of IT technology to exchange spatially and maybe additionally temporally indexed multi-dimensional geoscience data and corresponding models, i.e. the objects composing geomodels capturing the geometry, topology, and various geoscience contents. Geodata Infrastructure (GDI) and interoperability are objectives of several inititatives, e.g. INSPIRE, OneGeology-Europe, and most recently EGDI-SCOPE to name just the most prominent ones. Then there are quite a few markup languages (ML) related to geographical or geological information like GeoSciML, EarthResourceML, BoreholeML, ResqML for reservoir characterization, earth and reservoir models, and many others featuring geoscience information. Several Web Services are focused on geographical or geoscience information. The Open Geospatial Consortium (OGC) promotes specifications of a Web Feature Service (WFS), a Web Map Service (WMS), a Web Coverage Serverice (WCS), a Web 3D Service (W3DS), and many more. It will be clarified how GST is related to these initiatives, especially how it complies with existing or developing standards or quasi-standards and how it applies and extents services towards interoperability in the Earth sciences.
Henry, Anna E; Story, Mary
2009-01-01
To identify food and beverage brand Web sites featuring designated children's areas, assess marketing techniques present on those industry Web sites, and determine nutritional quality of branded food items marketed to children. Systematic content analysis of food and beverage brand Web sites and nutrient analysis of food and beverages advertised on these Web sites. The World Wide Web. One-hundred thirty Internet Web sites of food and beverage brands with top media expenditures based on the America's Top 2000 Brands section of Brandweek magazine's annual "Superbrands" report. A standardized content analysis rating form to determine marketing techniques used on the food and beverage brand Web sites. Nutritional analysis of food brands was conducted. Of 130 Web sites analyzed, 48% featured designated children's areas. These Web sites featured a variety of Internet marketing techniques, including advergaming on 85% of the Web sites and interactive programs on 92% of the Web sites. Branded spokescharacters and tie-ins to other products were featured on the majority of the Web sites, as well. Few food brands (13%) with Web sites that market to children met the nutrition criteria set by the National Alliance for Nutrition and Activity. Nearly half of branded Web sites analyzed used designated children's areas to market food and beverages to children, 87% of which were of low nutritional quality. Nutrition professionals should advocate the use of advertising techniques to encourage healthful food choices for children.
WebVR: an interactive web browser for virtual environments
NASA Astrophysics Data System (ADS)
Barsoum, Emad; Kuester, Falko
2005-03-01
The pervasive nature of web-based content has lead to the development of applications and user interfaces that port between a broad range of operating systems and databases, while providing intuitive access to static and time-varying information. However, the integration of this vast resource into virtual environments has remained elusive. In this paper we present an implementation of a 3D Web Browser (WebVR) that enables the user to search the internet for arbitrary information and to seamlessly augment this information into virtual environments. WebVR provides access to the standard data input and query mechanisms offered by conventional web browsers, with the difference that it generates active texture-skins of the web contents that can be mapped onto arbitrary surfaces within the environment. Once mapped, the corresponding texture functions as a fully integrated web-browser that will respond to traditional events such as the selection of links or text input. As a result, any surface within the environment can be turned into a web-enabled resource that provides access to user-definable data. In order to leverage from the continuous advancement of browser technology and to support both static as well as streamed content, WebVR uses ActiveX controls to extract the desired texture skin from industry strength browsers, providing a unique mechanism for data fusion and extensibility.
Rapid Evaluation of Radioactive Contamination in Rare Earth Mine Mining
NASA Astrophysics Data System (ADS)
Wang, N.
2017-12-01
In order to estimate the current levels of environmental radioactivity in Bayan Obo rare earth mine and to study the rapid evaluation methods of radioactivity contamination in the rare earth mine, the surveys of the in-situ gamma-ray spectrometry and gamma dose rate measurement were carried out around the mining area and living area. The in-situ gamma-ray spectrometer was composed of a scintillation detector of NaI(Tl) (Φ75mm×75mm) and a multichannel analyzer. Our survey results in Bayan Obo Mine display: (1) Thorium-232 is the radioactive contamination source of this region, and uranium-238 and potassium - 40 is at the background level. (2) The average content of thorium-232 in the slag of the tailings dam in Bayan Obo is as high as 276 mg/kg, which is 37 times as the global average value of thorium content. (3) We found that the thorium-232 content in the soil in the living area near the mining is higher than that in the local soil in Guyang County. The average thorium-232 concentrations in the mining areas of the Bayan Obo Mine and the living areas of the Bayan Obo Town were 18.7±7.5 and 26.2±9.1 mg/kg, respectively. (4) It was observed that thorium-232 was abnormal distributed in the contaminated area near the tailings dam. Our preliminary research results show that the in-situ gamma-ray spectrometry is an effective approach of fast evaluating rare earths radioactive pollution, not only can the scene to determine the types of radioactive contamination source, but also to measure the radioactivity concentration of thorium and uranium in soil. The environmental radioactive evaluation of rare earth ore and tailings dam in open-pit mining is also needed. The research was supported by National Natural Science Foundation of China (No. 41674111).
Content and Accessibility of Shoulder and Elbow Fellowship Web Sites in the United States.
Young, Bradley L; Oladeji, Lasun O; Cichos, Kyle; Ponce, Brent
2016-01-01
Increasing numbers of training physicians are using the Internet to gather information about graduate medical education programs. The content and accessibility of web sites that provide this information have been demonstrated to influence applicants' decisions. Assessments of orthopedic fellowship web sites including sports medicine, pediatrics, hand and spine have found varying degrees of accessibility and material. The purpose of this study was to evaluate the accessibility and content of the American Shoulder and Elbow Surgeons (ASES) fellowship web sites (SEFWs). A complete list of ASES programs was obtained from a database on the ASES web site. The accessibility of each SEFWs was assessed by the existence of a functioning link found in the database and through Google®. Then, the following content areas of each SEFWs were evaluated: fellow education, faculty/previous fellow information, and recruitment. At the time of the study, 17 of the 28 (60.7%) ASES programs had web sites accessible through Google®, and only five (17.9%) had functioning links in the ASES database. Nine programs lacked a web site. Concerning web site content, the majority of SEFWs contained information regarding research opportunities, research requirements, case descriptions, meetings and conferences, teaching responsibilities, attending faculty, the application process, and a program description. Fewer than half of the SEFWs provided information regarding rotation schedules, current fellows, previous fellows, on-call expectations, journal clubs, medical school of current fellows, residency of current fellows, employment of previous fellows, current research, and previous research. A large portion of ASES fellowship programs lacked functioning web sites, and even fewer provided functioning links through the ASES database. Valuable information for potential applicants was largely inadequate across present SEFWs.
Non-visual Web Browsing: Beyond Web Accessibility
Ramakrishnan, I.V.; Ashok, Vikas
2017-01-01
People with vision impairments typically use screen readers to browse the Web. To facilitate non-visual browsing, web sites must be made accessible to screen readers, i.e., all the visible elements in the web site must be readable by the screen reader. But even if web sites are accessible, screen-reader users may not find them easy to use and/or easy to navigate. For example, they may not be able to locate the desired information without having to listen to a lot of irrelevant contents. These issues go beyond web accessibility and directly impact web usability. Several techniques have been reported in the accessibility literature for making the Web usable for screen reading. This paper is a review of these techniques. Interestingly, the review reveals that understanding the semantics of the web content is the overarching theme that drives these techniques for improving web usability. PMID:29202137
Non-visual Web Browsing: Beyond Web Accessibility.
Ramakrishnan, I V; Ashok, Vikas; Billah, Syed Masum
2017-07-01
People with vision impairments typically use screen readers to browse the Web. To facilitate non-visual browsing, web sites must be made accessible to screen readers, i.e., all the visible elements in the web site must be readable by the screen reader. But even if web sites are accessible, screen-reader users may not find them easy to use and/or easy to navigate. For example, they may not be able to locate the desired information without having to listen to a lot of irrelevant contents. These issues go beyond web accessibility and directly impact web usability. Several techniques have been reported in the accessibility literature for making the Web usable for screen reading. This paper is a review of these techniques. Interestingly, the review reveals that understanding the semantics of the web content is the overarching theme that drives these techniques for improving web usability.
Analyzing Student Inquiry Data Using Process Discovery and Sequence Classification
ERIC Educational Resources Information Center
Emond, Bruno; Buffett, Scott
2015-01-01
This paper reports on results of applying process discovery mining and sequence classification mining techniques to a data set of semi-structured learning activities. The main research objective is to advance educational data mining to model and support self-regulated learning in heterogeneous environments of learning content, activities, and…
Adverse Drug Event Discovery Using Biomedical Literature: A Big Data Neural Network Adventure.
P Tafti, Ahmad; Badger, Jonathan; LaRose, Eric; Shirzadi, Ehsan; Mahnke, Andrea; Mayer, John; Ye, Zhan; Page, David; Peissig, Peggy
2017-12-08
The study of adverse drug events (ADEs) is a tenured topic in medical literature. In recent years, increasing numbers of scientific articles and health-related social media posts have been generated and shared daily, albeit with very limited use for ADE study and with little known about the content with respect to ADEs. The aim of this study was to develop a big data analytics strategy that mines the content of scientific articles and health-related Web-based social media to detect and identify ADEs. We analyzed the following two data sources: (1) biomedical articles and (2) health-related social media blog posts. We developed an intelligent and scalable text mining solution on big data infrastructures composed of Apache Spark, natural language processing, and machine learning. This was combined with an Elasticsearch No-SQL distributed database to explore and visualize ADEs. The accuracy, precision, recall, and area under receiver operating characteristic of the system were 92.7%, 93.6%, 93.0%, and 0.905, respectively, and showed better results in comparison with traditional approaches in the literature. This work not only detected and classified ADE sentences from big data biomedical literature but also scientifically visualized ADE interactions. To the best of our knowledge, this work is the first to investigate a big data machine learning strategy for ADE discovery on massive datasets downloaded from PubMed Central and social media. This contribution illustrates possible capacities in big data biomedical text analysis using advanced computational methods with real-time update from new data published on a daily basis. ©Ahmad P Tafti, Jonathan Badger, Eric LaRose, Ehsan Shirzadi, Andrea Mahnke, John Mayer, Zhan Ye, David Page, Peggy Peissig. Originally published in JMIR Medical Informatics (http://medinform.jmir.org), 08.12.2017.
Proactive Supply Chain Performance Management with Predictive Analytics
Stefanovic, Nenad
2014-01-01
Today's business climate requires supply chains to be proactive rather than reactive, which demands a new approach that incorporates data mining predictive analytics. This paper introduces a predictive supply chain performance management model which combines process modelling, performance measurement, data mining models, and web portal technologies into a unique model. It presents the supply chain modelling approach based on the specialized metamodel which allows modelling of any supply chain configuration and at different level of details. The paper also presents the supply chain semantic business intelligence (BI) model which encapsulates data sources and business rules and includes the data warehouse model with specific supply chain dimensions, measures, and KPIs (key performance indicators). Next, the paper describes two generic approaches for designing the KPI predictive data mining models based on the BI semantic model. KPI predictive models were trained and tested with a real-world data set. Finally, a specialized analytical web portal which offers collaborative performance monitoring and decision making is presented. The results show that these models give very accurate KPI projections and provide valuable insights into newly emerging trends, opportunities, and problems. This should lead to more intelligent, predictive, and responsive supply chains capable of adapting to future business environment. PMID:25386605
Proactive supply chain performance management with predictive analytics.
Stefanovic, Nenad
2014-01-01
Today's business climate requires supply chains to be proactive rather than reactive, which demands a new approach that incorporates data mining predictive analytics. This paper introduces a predictive supply chain performance management model which combines process modelling, performance measurement, data mining models, and web portal technologies into a unique model. It presents the supply chain modelling approach based on the specialized metamodel which allows modelling of any supply chain configuration and at different level of details. The paper also presents the supply chain semantic business intelligence (BI) model which encapsulates data sources and business rules and includes the data warehouse model with specific supply chain dimensions, measures, and KPIs (key performance indicators). Next, the paper describes two generic approaches for designing the KPI predictive data mining models based on the BI semantic model. KPI predictive models were trained and tested with a real-world data set. Finally, a specialized analytical web portal which offers collaborative performance monitoring and decision making is presented. The results show that these models give very accurate KPI projections and provide valuable insights into newly emerging trends, opportunities, and problems. This should lead to more intelligent, predictive, and responsive supply chains capable of adapting to future business environment.
Hancock, David; Wilson, Michael; Velarde, Giles; Morrison, Norman; Hayes, Andrew; Hulme, Helen; Wood, A Joseph; Nashar, Karim; Kell, Douglas B; Brass, Andy
2005-11-03
maxdLoad2 is a relational database schema and Java application for microarray experimental annotation and storage. It is compliant with all standards for microarray meta-data capture; including the specification of what data should be recorded, extensive use of standard ontologies and support for data exchange formats. The output from maxdLoad2 is of a form acceptable for submission to the ArrayExpress microarray repository at the European Bioinformatics Institute. maxdBrowse is a PHP web-application that makes contents of maxdLoad2 databases accessible via web-browser, the command-line and web-service environments. It thus acts as both a dissemination and data-mining tool. maxdLoad2 presents an easy-to-use interface to an underlying relational database and provides a full complement of facilities for browsing, searching and editing. There is a tree-based visualization of data connectivity and the ability to explore the links between any pair of data elements, irrespective of how many intermediate links lie between them. Its principle novel features are: the flexibility of the meta-data that can be captured, the tools provided for importing data from spreadsheets and other tabular representations, the tools provided for the automatic creation of structured documents, the ability to browse and access the data via web and web-services interfaces. Within maxdLoad2 it is very straightforward to customise the meta-data that is being captured or change the definitions of the meta-data. These meta-data definitions are stored within the database itself allowing client software to connect properly to a modified database without having to be specially configured. The meta-data definitions (configuration file) can also be centralized allowing changes made in response to revisions of standards or terminologies to be propagated to clients without user intervention.maxdBrowse is hosted on a web-server and presents multiple interfaces to the contents of maxd databases. maxdBrowse emulates many of the browse and search features available in the maxdLoad2 application via a web-browser. This allows users who are not familiar with maxdLoad2 to browse and export microarray data from the database for their own analysis. The same browse and search features are also available via command-line and SOAP server interfaces. This both enables scripting of data export for use embedded in data repositories and analysis environments, and allows access to the maxd databases via web-service architectures. maxdLoad2 http://www.bioinf.man.ac.uk/microarray/maxd/ and maxdBrowse http://dbk.ch.umist.ac.uk/maxdBrowse are portable and compatible with all common operating systems and major database servers. They provide a powerful, flexible package for annotation of microarray experiments and a convenient dissemination environment. They are available for download and open sourced under the Artistic License.
15 CFR 970.203 - Exploration plan.
Code of Federal Regulations, 2013 CFR
2013-01-01
... DATA SERVICE DEEP SEABED MINING REGULATIONS FOR EXPLORATION LICENSES Applications Contents § 970.203...; (iii) Designing and testing system components onshore and at sea; (iv) Designing and testing mining...
15 CFR 970.207 - Antitrust information.
Code of Federal Regulations, 2014 CFR
2014-01-01
... ENVIRONMENTAL DATA SERVICE DEEP SEABED MINING REGULATIONS FOR EXPLORATION LICENSES Applications Contents § 970... license, provided that said agreement relates to deep seabed hard mineral resource exploration or mining...
15 CFR 970.203 - Exploration plan.
Code of Federal Regulations, 2014 CFR
2014-01-01
... DATA SERVICE DEEP SEABED MINING REGULATIONS FOR EXPLORATION LICENSES Applications Contents § 970.203...; (iii) Designing and testing system components onshore and at sea; (iv) Designing and testing mining...
15 CFR 970.203 - Exploration plan.
Code of Federal Regulations, 2012 CFR
2012-01-01
... DATA SERVICE DEEP SEABED MINING REGULATIONS FOR EXPLORATION LICENSES Applications Contents § 970.203...; (iii) Designing and testing system components onshore and at sea; (iv) Designing and testing mining...
15 CFR 970.207 - Antitrust information.
Code of Federal Regulations, 2012 CFR
2012-01-01
... ENVIRONMENTAL DATA SERVICE DEEP SEABED MINING REGULATIONS FOR EXPLORATION LICENSES Applications Contents § 970... license, provided that said agreement relates to deep seabed hard mineral resource exploration or mining...
15 CFR 970.207 - Antitrust information.
Code of Federal Regulations, 2013 CFR
2013-01-01
... ENVIRONMENTAL DATA SERVICE DEEP SEABED MINING REGULATIONS FOR EXPLORATION LICENSES Applications Contents § 970... license, provided that said agreement relates to deep seabed hard mineral resource exploration or mining...
15 CFR 970.207 - Antitrust information.
Code of Federal Regulations, 2010 CFR
2010-01-01
... ENVIRONMENTAL DATA SERVICE DEEP SEABED MINING REGULATIONS FOR EXPLORATION LICENSES Applications Contents § 970... license, provided that said agreement relates to deep seabed hard mineral resource exploration or mining...
15 CFR 970.203 - Exploration plan.
Code of Federal Regulations, 2010 CFR
2010-01-01
... DATA SERVICE DEEP SEABED MINING REGULATIONS FOR EXPLORATION LICENSES Applications Contents § 970.203...; (iii) Designing and testing system components onshore and at sea; (iv) Designing and testing mining...
15 CFR 970.207 - Antitrust information.
Code of Federal Regulations, 2011 CFR
2011-01-01
... ENVIRONMENTAL DATA SERVICE DEEP SEABED MINING REGULATIONS FOR EXPLORATION LICENSES Applications Contents § 970... license, provided that said agreement relates to deep seabed hard mineral resource exploration or mining...
15 CFR 970.203 - Exploration plan.
Code of Federal Regulations, 2011 CFR
2011-01-01
... DATA SERVICE DEEP SEABED MINING REGULATIONS FOR EXPLORATION LICENSES Applications Contents § 970.203...; (iii) Designing and testing system components onshore and at sea; (iv) Designing and testing mining...
Ji, Li; Yang, Li Xue
2017-12-01
Phenolic acids are one of the most important factors that influence microbial community structure. Investigating the dynamic changes of phenolic acids and their relationship with the microbial community structure in plantation soils with different tree species could contribute to better understanding and revealing the mechanisms of microbial community changes under afforestation restoration in coal-mining subsidence areas. In this study, plantations of three conifer and one deciduous species (Pinus koraiensis, Larix gmelinii, Pinus sylvestris var. mongolica, and Populus ussuriensis) were established on abandoned coal-mining subsidence areas in Baoshan District, Shuangyashan City. The contents of soil phenols, 11 types of phenolic acids, and microbial communities in all plots were determined. The results showed that the contents of soil complex phenol in plantations were significantly higher than that of abandoned land overall. Specifically, soils in larch and poplar plantations had higher contents of complex phenol, while soils in larch and Korean pine plantations had greater contents of total phenol. Moreover, soil in the P. koraiensis plantation had a higher content of water-soluble phenol compared with abandoned lands. The determination of 11 phenolic acids indicated that the contents of ferulic acid, abietic acid, β-sitosterol, oleanolic acid, shikimic acid, linoleic acid, and stearic acid were higher in plantation soils. Although soil phenol contents were not related with soil microbial biomass, the individual phenolic acids showed a significant relationship with soil microbes. Ferulic acid, abietic acid, and β-sitosterol showed significant promoting effects on soil microbial biomass, and they showed positive correlations with fungi and fungi/bacteria ratio. These three phenolic acids had higher contents in the poplar plantation, suggesting that poplar affo-restation had a beneficial effect on soil quality in coal-mining subsidence areas.
Emergence and growth of plant species in coal mine soil
DOE Office of Scientific and Technical Information (OSTI.GOV)
Day, A.D.; Mitchell, G.F.; Tucker, T.C.
1979-01-01
Experiments were conducted in the laboratory and greenhouse in Arizona with the following objectives: to evaluate the chemical properties of undisturbed soil, surface-mined coal land (coal mine soil) on the Black Mesa Coal Mine, and Gila loam soil; and to study the emergence of seven plant species in the greenhouse in Gila loam soil and coal mine soil. The pH of coal mine soil (6.2) was lower than the pH of undisturbed soil (7.5) or Gila loam (7.6). The total soluble salts in coal mine soil (3241) and undisturbed soil (4592) were much higher than in Gila loam (378); however,more » coal mine soil was lower in total soluble salts than undisturbed soil. The nitrogen content of coal mine soil was higher than the nitrogen content of undisturbed soil or gila loam. Emergence percentages for seven plant species grown in coal mine soil were similar to emergence percentages for the same species grown in Gila loam. Alfalfa (Medicago sativa L.), barley (Hordeum vulgare L.), and wheat (Triticum aestivum L. em Thell.) had from 84 to 93% emergence in coal mine soil. Indian ricegrass (Oryzopsis hymenoides Roem. and Shult), fourwing saltbush (Atriplex canescens Pursh), yellow sweetclover (Melilotus officinalis Lam.), and winterfat (Euroti lanata Pursh.) emerged <35% in coal mine soil and Gila loam. Plant growth data from forage species grown in the greenhouse indicate that coal mine soil has a lower fertility level than does Gila loam soil. When supplied with optimum soil moisture and plant nutrients, coal mine soil produced approximately the same yields of forage from alfalfa, barley, and wheat as were produced in Gila loam under the same soil-moisture and fertility conditions.« less
Enhancing e-Learning Content by Using Semantic Web Technologies
ERIC Educational Resources Information Center
García-González, Herminio; Gayo, José Emilio Labra; del Puerto Paule-Ruiz, María
2017-01-01
We describe a new educational tool that relies on Semantic Web technologies to enhance lessons content. We conducted an experiment with 32 students whose results demonstrate better performance when exposed to our tool in comparison with a plain native tool. Consequently, this prototype opens new possibilities in lessons content enhancement.
10 Budget-Savvy Content Management Strategies
ERIC Educational Resources Information Center
Hillis, David J.
2004-01-01
Facing an overall budget reduction of 10 percent, most colleges or universities would postpone investing in a Web content management system. However, for California State University Monterey Bay (CSUMB), a large budget cut made Web content management even more important. CSUMB found an innovative way to purchase and implement a new Content…
Content Management for a Content-Rich Website.
ERIC Educational Resources Information Center
Honeysett, Nik
Over the last year the J. Paul Getty Trust's Web presence has evolved from a group of disparate, independently maintained Web sites into a homogeneous consistently branded one. This transformation recently culminated with the implementation of a leading Content Management System (CMS). There were and are many process-changes and challenges in…
Frost, Thomas P.; Box, Stephen E.
2009-01-01
This reconnaissance study was undertaken at the request of the USDA Forest Service, Region 4, to assess the geochemistry, in particular the mercury and selenium contents, of mining-impacted sediments in the Yankee Fork of the Salmon River in Custer County Idaho. The Yankee Fork has been the site of hard-rock and placer mining, primarily for gold and silver, starting in the 1880s. Major dredge placer mining from the 1930s to 1950s in the Yankee Fork disturbed about a 10-kilometer reach. Mercury was commonly used in early hard-rock mining and placer operations for amalgamation and recovery of gold. During the late 1970s, feasibility studies were done on cyanide-heap leach recovery of gold from low-grade ores of the Sunbeam and related deposits. In the mid-1990s a major open-pit bulk-vat leach operation was started at the Grouse Creek Mine. This operation shut down when gold values proved to be lower than expected. Mercury in stream sediments in the Yankee Fork ranges from below 0.02 ppm to 7 ppm, with the highest values associated with old mill locations and lode and placer mines. Selenium ranges from below the detection limit for this study of 0.2 ppm to 4 ppm in Yankee Fork sediment samples. The generally elevated selenium content in the sediment samples reflect the generally high selenium contents in the volcanic rocks that underlie the Yankee Fork and the presence of gold and silver selenides in some of the veins that were exploited in the early phases of mining.
Advanced Query and Data Mining Capabilities for MaROS
NASA Technical Reports Server (NTRS)
Wang, Paul; Wallick, Michael N.; Allard, Daniel A.; Gladden, Roy E.; Hy, Franklin H.
2013-01-01
The Mars Relay Operational Service (MaROS) comprises a number of tools to coordinate, plan, and visualize various aspects of the Mars Relay network. These levels include a Web-based user interface, a back-end "ReSTlet" built in Java, and databases that store the data as it is received from the network. As part of MaROS, the innovators have developed and implemented a feature set that operates on several levels of the software architecture. This new feature is an advanced querying capability through either the Web-based user interface, or through a back-end REST interface to access all of the data gathered from the network. This software is not meant to replace the REST interface, but to augment and expand the range of available data. The current REST interface provides specific data that is used by the MaROS Web application to display and visualize the information; however, the returned information from the REST interface has typically been pre-processed to return only a subset of the entire information within the repository, particularly only the information that is of interest to the GUI (graphical user interface). The new, advanced query and data mining capabilities allow users to retrieve the raw data and/or to perform their own data processing. The query language used to access the repository is a restricted subset of the structured query language (SQL) that can be built safely from the Web user interface, or entered as freeform SQL by a user. The results are returned in a CSV (Comma Separated Values) format for easy exporting to third party tools and applications that can be used for data mining or user-defined visualization and interpretation. This is the first time that a service is capable of providing access to all cross-project relay data from a single Web resource. Because MaROS contains the data for a variety of missions from the Mars network, which span both NASA and ESA, the software also establishes an access control list (ACL) on each data record in the database repository to enforce user access permissions through a multilayered approach.
MINING ENVIRONMENTAL TOXICOLOGY INFORMATION WEB RESOURCES
Environmental toxicology is the study of the ecological effects of anthropogenic substances released into the environment. It is a relatively diverse field addressing impacts to aquatic and terrestrial organisms and communities. The determination of potential risk associated with...
AMP: A platform for managing and mining data in the treatment of Autism Spectrum Disorder.
Linstead, Erik; Burns, Ryan; Duy Nguyen; Tyler, David
2016-08-01
We introduce AMP (Autism Management Platform), an integrated health care information system for capturing, analyzing, and managing data associated with the diagnosis and treatment of Autism Spectrum Disorder in children. AMP's mobile application simplifies the means by which parents, guardians, and clinicians can collect and share multimedia data with one another, facilitating communication and reducing data redundancy, while simplifying retrieval. Additionally, AMP provides an intelligent web interface and analytics platform which allow physicians and specialists to aggregate and mine patient data in real-time, as well as give relevant feedback to automatically learn data filtering preferences over time. Together AMP's mobile app, web client, and analytics engine implement a rich set of features that streamline the data collection and analysis process in the context of a secure and easy-to-use system so that data may be more effectively leveraged to guide treatment.
ERIC Educational Resources Information Center
Hall, Richard H.; Hanna, Patrick
2004-01-01
The purpose of this experiment was to examine the effect of web page text/background colour combination on readability, retention, aesthetics, and behavioural intention. One hundred and thirty-six participants studied two Web pages, one with educational content and one with commercial content, in one of four colour-combination conditions. Major…
Trigger Videos on the Web: Impact of Audiovisual Design
ERIC Educational Resources Information Center
Verleur, Ria; Heuvelman, Ard; Verhagen, Plon W.
2011-01-01
Audiovisual design might impact emotional responses, as studies from the 1970s and 1980s on movie and television content show. Given today's abundant presence of web-based videos, this study investigates whether audiovisual design will impact web-video content in a similar way. The study is motivated by the potential influence of video-evoked…
Code of Federal Regulations, 2011 CFR
2011-10-01
... standards, and resolve any related issues. (c) Based on those discussions, the Project Officer shall provide... communication must meet the accessibility standards in 36 CFR 1194.22, “Web-based intranet and Internet... standards for HHS Web site content and communications materials. 311.7001 Section 311.7001 Federal...
Code of Federal Regulations, 2013 CFR
2013-10-01
... standards, and resolve any related issues. (c) Based on those discussions, the Project Officer shall provide... communication must meet the accessibility standards in 36 CFR 1194.22, “Web-based intranet and Internet... standards for HHS Web site content and communications materials. 311.7001 Section 311.7001 Federal...
Code of Federal Regulations, 2012 CFR
2012-10-01
... standards, and resolve any related issues. (c) Based on those discussions, the Project Officer shall provide... communication must meet the accessibility standards in 36 CFR 1194.22, “Web-based intranet and Internet... standards for HHS Web site content and communications materials. 311.7001 Section 311.7001 Federal...
Code of Federal Regulations, 2014 CFR
2014-10-01
... standards, and resolve any related issues. (c) Based on those discussions, the Project Officer shall provide... communication must meet the accessibility standards in 36 CFR 1194.22, “Web-based intranet and Internet... standards for HHS Web site content and communications materials. 311.7001 Section 311.7001 Federal...
Predicting Audience Demographics of Web Sites Using Local Cues
ERIC Educational Resources Information Center
Kim, Iljoo
2011-01-01
The size and dynamism of the Web poses challenges for all its stakeholders, which include producers/consumers of content, and advertisers who want to place advertisements next to relevant content. A critical piece of information for the stakeholders is the demographics of the consumers who are likely to visit a given web site. However, predicting…
Methane drainage at the Minerales Monclova mines in the Sabinas coal basin, Coahuila, Mexico
DOE Office of Scientific and Technical Information (OSTI.GOV)
Brunner, D.J.; Ponce, J.R.
Minerales Monclova S.A. De C.V. (MIMOSA) operates five underground longwall mines in the Gassy Los Olmos Coals of the Sabinas Basin in the state of Coahuila in Northern Mexico. Because of high in-situ gas contents and high cleat and natural fracture permeability, MIMOSA has had to incorporate a system of methane drainage in advance of mining in order to safely and cost effectively exploit their reserves. In the early 1990s Resource Enterprises (REI) conducted reservoir characterization tests, numerical simulations, and Coal Mine Methane (CMM) production tests at a nearby mine property in the same basin. Using this information REI approachedmore » MIMOSA and recommended the mine-wide implementation of a degasification system that involves long in-seam directionally drilled boreholes. REI was contracted to conduct the drilling, and to date has drilled over 26,000 m (85,000 ft) of in-seam borehole in advance of mining developments, reducing gas contents significantly below in-situ values. This paper discusses the basis for the degasification program recommended at the MIMOSA mines, and presents the impact of its mine-wide application on MIMOSA's mining operations over the last six years. The paper focuses on the degasification system's impacts on methane emissions into mine workings, coal production, and ventilation demands. It also presents lessons learned by the degasification planners in implementing in-seam methane drainage. The paper presents actual CMM production data, measurements of methane emissions and advance rates at development sections, and mine methane liberations.« less
Uncovering text mining: A survey of current work on web-based epidemic intelligence
Collier, Nigel
2012-01-01
Real world pandemics such as SARS 2002 as well as popular fiction like the movie Contagion graphically depict the health threat of a global pandemic and the key role of epidemic intelligence (EI). While EI relies heavily on established indicator sources a new class of methods based on event alerting from unstructured digital Internet media is rapidly becoming acknowledged within the public health community. At the heart of automated information gathering systems is a technology called text mining. My contribution here is to provide an overview of the role that text mining technology plays in detecting epidemics and to synthesise my existing research on the BioCaster project. PMID:22783909
SalanderMaps: A rapid overview about felt earthquakes through data mining of web-accesses
NASA Astrophysics Data System (ADS)
Kradolfer, Urs
2013-04-01
While seismological observatories detect and locate earthquakes based on measurements of the ground motion, they neither know a priori whether an earthquake has been felt by the public nor is it known, where it has been felt. Such information is usually gathered by evaluating feedback reported by the public through on-line forms on the web. However, after a felt earthquake in Switzerland, many people visit the webpages of the Swiss Seismological Service (SED) at the ETH Zurich and each such visit leaves traces in the logfiles on our web-servers. Data mining techniques, applied to these logfiles and mining publicly available data bases on the internet open possibilities to obtain previously unknown information about our virtual visitors. In order to provide precise information to authorities and the media, it would be desirable to rapidly know from which locations these web-accesses origin. The method 'Salander' (Seismic Activitiy Linked to Area codes - Nimble Detection of Earthquake Rumbles) will be introduced and it will be explained, how the IP-addresses (each computer or router directly connected to the internet has a unique IP-address; an example would be 129.132.53.5) of a sufficient amount of our virtual visitors were linked to their geographical area. This allows us to unprecedentedly quickly know whether and where an earthquake was felt in Switzerland. It will also be explained, why the method Salander is superior to commercial so-called geolocation products. The corresponding products of the Salander method, animated SalanderMaps, which are routinely generated after each earthquake with a magnitude of M>2 in Switzerland (http://www.seismo.ethz.ch/prod/salandermaps/, available after March 2013), demonstrate how the wavefield of earthquakes propagates through Switzerland and where it was felt. Often, such information is available within less than 60 seconds after origin time, and we always get a clear picture within already five minutes after origin time. Furthermore, the method allows to detect earthquakes solely on the analysis of accesses to our web-servers. Analyzing more than 170 million web-accesses since 2003, all seismic events within or near Switzerland with magnitudes M>4 and most felt events with magnitudes between 3 and 4 were detected. The current system is very robust, as we only had one false alarm while re-processing the web-access logfiles of the past almost 10 years. We anticipate that this method will produce even faster results in the future as the number of both commercial and private internet users is - according to the statistics of our logfiles - still increasing.
Protano, Giuseppe; Nannoni, Francesco
2018-05-01
A geochemical study was carried out at the former Abbadia San Salvatore (ASS) mining site of the Monte Amiata ore district (Italy). Hg, As and Sb total contents and fractionation using a sequential extraction procedure were determined in soil and mining waste samples. Ore processing activities provided a different contribution to Hg contamination and concentration in soil fractions, influencing its behaviour as volatility and availability. Soils of roasting zone showed the highest Hg contamination levels mainly due to the deposition of Hg released as Hg 0 by furnaces during cinnabar roasting. High Hg contents were also measured in waste from the lower part of mining dump due to the presence of cinnabar. The fractionation pattern suggested that Hg was largely as volatile species in both uncontaminated and contaminated soils and mining waste, and concentrations of these Hg species increased as contamination increased. These findings were in agreement with the fact that the ASS mining site is characterized by high Hg concentrations in the air and the presence of Hg 0 liquid droplets in soil. Volatile Hg species were also prevalent in uncontaminated soils likely because the Monte Amiata region is an area characterized by anomalous fluxes of gaseous Hg from natural and anthropogenic inputs. At the ASS mining site soils were also contaminated by Sb, while As contents were comparable with its local background in soil. In all soil and waste samples Sb and As were preferentially in residual fraction. Copyright © 2018 Elsevier Ltd. All rights reserved.
Moran, Anthony R; Hettiarachchi, Hiroshan
2011-07-01
Clayey soil found in coal mines in Appalachian Ohio is often sold to landfills for constructing Recompacted Soil Liners (RSL) in landfills. Since clayey soils possess low hydraulic conductivity, the suitability of mined clay for RSL in Ohio is first assessed by determining its clay content. When soil samples are tested in a laboratory, the same engineering properties are typically expected for the soils originated from the same source, provided that the testing techniques applied are standard, but mined clay from Appalachian Ohio has shown drastic differences in particle size distribution depending on the sampling and/or laboratory processing methods. Sometimes more than a 10 percent decrease in the clay content is observed in the samples collected at the stockpiles, compared to those collected through reverse circulation drilling. This discrepancy poses a challenge to geotechnical engineers who work on the prequalification process of RSL material as it can result in misleading estimates of the hydraulic conductivity of the samples. This paper describes a laboratory investigation conducted on mined clay from Appalachian Ohio to determine how and why the standard sampling and/or processing methods can affect the grain-size distributions. The variation in the clay content was determined to be due to heavy concentrations of shale fragments in the clayey soils. It was also concluded that, in order to obtain reliable grain size distributions from the samples collected at a stockpile of mined clay, the material needs to be processed using a soil grinder. Otherwise, the samples should be collected through drilling.
Moran, Anthony R.; Hettiarachchi, Hiroshan
2011-01-01
Clayey soil found in coal mines in Appalachian Ohio is often sold to landfills for constructing Recompacted Soil Liners (RSL) in landfills. Since clayey soils possess low hydraulic conductivity, the suitability of mined clay for RSL in Ohio is first assessed by determining its clay content. When soil samples are tested in a laboratory, the same engineering properties are typically expected for the soils originated from the same source, provided that the testing techniques applied are standard, but mined clay from Appalachian Ohio has shown drastic differences in particle size distribution depending on the sampling and/or laboratory processing methods. Sometimes more than a 10 percent decrease in the clay content is observed in the samples collected at the stockpiles, compared to those collected through reverse circulation drilling. This discrepancy poses a challenge to geotechnical engineers who work on the prequalification process of RSL material as it can result in misleading estimates of the hydraulic conductivity of the samples. This paper describes a laboratory investigation conducted on mined clay from Appalachian Ohio to determine how and why the standard sampling and/or processing methods can affect the grain-size distributions. The variation in the clay content was determined to be due to heavy concentrations of shale fragments in the clayey soils. It was also concluded that, in order to obtain reliable grain size distributions from the samples collected at a stockpile of mined clay, the material needs to be processed using a soil grinder. Otherwise, the samples should be collected through drilling. PMID:21845150
Online activities to optimize in person learning
NASA Astrophysics Data System (ADS)
Stelzer, Tim
Students' unprecedented access to content on the web is providing a unique opportunity to transform the role lectures in education, moving the focus from content delivery to helping students synthesize the content into knowledge. We have introduced a variety of activities to facilitate this transformation at the University of Illinois, including web-based preflight assessments of student understanding before lecture, peer instruction (clickers) to assess and facilitate student understanding during lecture, and web-based multimedia pre-lectures designed to provide students with content before lecture. In this talk I will discuss the pedagogical motivation for introducing these activities, and the impact they have had at the University of Illinois. .
Fate and Trophic Transfer of Rare Earth Elements in Temperate Lake Food Webs.
Amyot, Marc; Clayden, Meredith G; MacMillan, Gwyneth A; Perron, Tania; Arscott-Gauvin, Alexandre
2017-06-06
Many mining projects targeting rare earth elements (REE) are in development in North America, but the background concentrations and trophic transfer of these elements in natural environments have not been well characterized. We sampled abiotic and food web components in 14 Canadian temperate lakes unaffected by mines to assess the natural ecosystem fate of REE. Individual REE and total REE concentrations (sum of individual element concentrations, ΣREE) were strongly related with each other throughout different components of lake food webs. Dissolved organic carbon and dissolved oxygen in the water column, as well as ΣREE in sediments, were identified as potential drivers of aqueous ΣREE. Log 10 of median bioaccumulation factors ranged from 1.3, 3.7, 4.0, and 4.4 L/kg (wet weight) for fish muscle, zooplankton, predatory invertebrates, and nonpredatory invertebrates, respectively. [ΣREE] in fish, benthic macroinvertebrates, and zooplankton declined as a function of their trophic position, as determined by functional feeding groups and isotopic signatures of nitrogen (δ 15 N), indicating that REE were subject to trophic dilution. Low concentrations of REE in freshwater fish muscle compared to their potential invertebrate prey suggest that fish fillet consumption is unlikely to be a significant source of REE to humans in areas unperturbed by mining activities. However, other fish predators (e.g., piscivorous birds and mammals) may accumulate REE from whole fish as they are more concentrated than muscle. Overall, this study provides key information on the baseline concentrations and trophic patterns for REE in freshwater temperate lakes in Quebec, Canada.
Speciation of arsenic in bulk and rhizosphere soils from artisanal cooperative mines in Bolivia.
Acosta, Jose A; Arocena, Joselito M; Faz, Angel
2015-11-01
Soils near artisanal and small-scale gold mines (ASGM) have high arsenic (As) contents due to the presence of arsenopyrite in gold ores and accelerated accumulations due to mine wastes disposal practices and other mining activities. We determined the content and speciation to understand the fate and environmental risks of As accumulations in 24 bulk and 12 rhizosphere soil samples collected in the Virgen Del Rosario and the Rayo Rojo cooperative mines in the highlands of Bolivia. Mean total As contents in bulk and rhizosphere soils ranged from 13 to 64 mg kg(-1) and exceeded the soil environmental quality guidelines of Canada. Rhizosphere soils always contained at least twice the As contents in the bulk soil. Elemental mapping using 4×5 μm synchrotron-generated X-ray micro-beam revealed As accumulations in areas enriched with Fe. Results of As-X-ray Absorption Near Edge Spectroscopy (As-XANES) showed that only As(V) species was detectable in all samples regardless of As contents, size fractions and types of vegetation. Although the toxicity of As(V) is less than As(III), we suggest that As uptake of commonly-grazed vegetation by alpaca and llama must be determined to fully understand the environmental risks of high As in soils near ASGM in Bolivia. In addition, knowledge on the speciation of the As bio-accessible fraction will provide another useful information to better understand the fate and transfer of As from soils into the food chain in environments associated with the ASGM in Bolivia and other parts of the world. Copyright © 2014 Elsevier Ltd. All rights reserved.
Distribution and mobility of arsenic in soils of a mining area (Western Spain).
García-Sánchez, A; Alonso-Rojo, P; Santos-Francés, F
2010-09-01
High levels of total and bioavailable As in soils in mining areas may lead to the potential contamination of surface water and groundwater, being toxic to human, plants, and animals. The soils in the studied area (Province of Salamanca, Spain) recorded a total As concentration that varied from 5.5mg/kg to 150mg/kg, and water-soluble As ranged from 0.004mg/kg to 0.107mg/kg, often exceeding the guideline limits for agricultural soil (50mg/kg total As, 0.04mg/kg water-soluble As). The range of As concentration in pond water was <0.001microg/l-60microg/l, with 40% of samples exceeding the maximum permissible level (10microg/l) for drinking water. Estimated bioavailable As in soil varied from 0.045mg/kg to 0.760mg/kg, around six times higher than water-soluble As fraction, which may pose a high potential risk in regard to its entry into food chain. Soil column leaching tests show an As potential mobility constant threatening water contamination by continuous leaching. The vertical distribution of As through soil profiles suggests a deposition mechanism of this element on the top-soils that involves the wind or water transport of mine tailings. A similar vertical distribution of As and organic matter (OM) contents in soil profiles, as well as, significant correlations between As concentrations and OM and N contents, suggests that type and content of soil OM are major factors for determining the content, distribution, and mobilization of As in the soil. Due to the low supergenic mobility of this element in mining environments, the soil pollution degree in the studied area is moderate, in spite of the elevated As contents in mine tailings. Copyright 2010 Elsevier B.V. All rights reserved.
Dempsey, Patrick G; Pollard, Jonisha; Porter, William L; Mayton, Alan; Heberger, John R; Gallagher, Sean; Reardon, Leanna; Drury, Colin G
2017-12-01
The development and testing of ergonomics and safety audits for small and bulk bag filling, haul truck and maintenance and repair operations in coal preparation and mineral processing plants found at surface mine sites is described. The content for the audits was derived from diverse sources of information on ergonomics and safety deficiencies including: analysis of injury, illness and fatality data and reports; task analysis; empirical laboratory studies of particular tasks; field studies and observations at mine sites; and maintenance records. These diverse sources of information were utilised to establish construct validity of the modular audits that were developed for use by mine safety personnel. User and interrater reliability testing was carried out prior to finalising the audits. The audits can be implemented using downloadable paper versions or with a free mobile NIOSH-developed Android application called ErgoMine. Practitioner Summary: The methodology used to develop ergonomics audits for three types of mining operations is described. Various sources of audit content are compared and contrasted to serve as a guide for developing ergonomics audits for other occupational contexts.
Odumo, Benjamin Okang'; Carbonell, Gregoria; Angeyo, Hudson Kalambuka; Patel, Jayanti Purshottam; Torrijos, Manuel; Rodríguez Martín, José Antonio
2014-11-01
This work considered the environmental impact of artisanal mining gold activity in the Migori-Transmara area (Kenya). From artisanal gold mining, mercury is released to the environment, thus contributing to degradation of soil and water bodies. High mercury contents have been quantified in soil (140 μg kg(-1)), sediment (430 μg kg(-1)) and tailings (8,900 μg kg(-1)), as expected. The results reveal that the mechanism for transporting mercury to the terrestrial ecosystem is associated with wet and dry depositions. Lichens and mosses, used as bioindicators of pollution, are related to the proximity to mining areas. The further the distance from mining areas, the lower the mercury levels. This study also provides risk maps to evaluate potential negative repercussions. We conclude that the Migori-Transmara region can be considered a strongly polluted area with high mercury contents. The technology used to extract gold throughout amalgamation processes causes a high degree of mercury pollution around this gold mining area. Thus, alternative gold extraction methods should be considered to reduce mercury levels that can be released to the environment.
NASA Astrophysics Data System (ADS)
López-Berdonces, Miguel Angel; María Esbrí, José; Fernández-Calderón, Sergio; Naharro, Elena; García-Noguero, Eva Maria; Higueras, Pablo
2014-05-01
El Borracho mine was active since Roman times, but with its higher production period on 19th Century. Mine closure occured without restoration works and nowadays the mining area is dedicated to deer hunting activities. In order to evaluate heavy metals distribution on mining tailings and surrounding soils of the studied area, 40 samples of dumps, soils and sediments were taken. Samples from the mine tailings were collected with an Eijkelkamp soil core sampler for undisturbed samples, with a vertical constant spacing of 25 cm. With this procedure, a total of 21 samples were taken in two points at main dump. Samples of Oak-tree leaves and moss were taken to evaluate metal transfer to biota. Analytical determinations have included soil parameters (pH, conductivity, organic matter content), and total metal contents in geological and biological samples by EDXRF. Analytical determinations shows higher metal contents in dumps, especially in surficial samples, 17,700 mg kg-1 and 470 mg kg-1 in average of Pb and Zn respectively, and lower contents in soils, 5,200 mg kg-1 and 300 mg kg-1, and sediments, 3,500 mg kg-1 and 120 mg kg-1. Metal contents in tailings profiles shows higher levels of Pb, Zn and Cu at 3.5 meters depth, a zone with lower grainsize and higher moisture. Differences in efficiency of extraction techniques and metal remobilization inside the dump can be an explanation for this enrichment level. Metal contents in agricultural soils exceeded maximum allowed levels by European Community (300 mg kg-1 for Pb and Zn and 140 mg kg-1 for Cu). Metal contents in biota evidence that Oak-tree bioaccumulates some metals, especially those with higher mobility in acidic conditions like Zn and Sb, with averages Bioaccumulation factor (BAF = plant concentration/soil concentration) of 0.48 and 0.85 respectively. Moss reaches high concentrations of Pb and Zn (3,000 mg kg-1 and 175 mg kg-1 in average respectively). Uptake pattern of Pb and Zn by plants leaves and mosses seems to be similar and can be characterized by logistic curves, with higher affinity of mosses to uptake metals from soils.
Evaluation of breastfeeding Web sites for patient education.
Dornan, Barbara A; Oermann, Marilyn H
2006-01-01
To evaluate the quality of Web sites on breastfeeding for patient education. Descriptive study of 30 Web sites on breastfeeding for patient education, evaluated based on the Health Information Technology Institute (HITI) criteria, readability, and eight content criteria from the American Academy of Pediatrics (AAP) policy statement on breastfeeding. The mean Flesch-Kincaid Grade Level for readability of the 30 sites was 9.2. Seven of the sites included all eight of the content criteria from the AAP, and three sites did not include any of the information recommended by the AAP content criteria. Nurses should be able to recommend best patient education materials for their patients. The five best Web sites for breastfeeding education are identified for patient teaching, and the HITI criteria are explained for nurses to learn how to evaluate Web sites for themselves and their patients.
15 CFR 971.207 - Antitrust information.
Code of Federal Regulations, 2010 CFR
2010-01-01
... ENVIRONMENTAL DATA SERVICE DEEP SEABED MINING REGULATIONS FOR COMMERCIAL RECOVERY PERMITS Applications Contents... share it has with respect to the mining or marketing of the metals proposed to be recovered under the...
15 CFR 971.203 - Commercial recovery plan.
Code of Federal Regulations, 2011 CFR
2011-01-01
... ENVIRONMENTAL DATA SERVICE DEEP SEABED MINING REGULATIONS FOR COMMERCIAL RECOVERY PERMITS Applications Contents... requirements for resource assessment and logical mining unit (§ 971.501); (6) A description of the methods and...
15 CFR 971.203 - Commercial recovery plan.
Code of Federal Regulations, 2010 CFR
2010-01-01
... ENVIRONMENTAL DATA SERVICE DEEP SEABED MINING REGULATIONS FOR COMMERCIAL RECOVERY PERMITS Applications Contents... requirements for resource assessment and logical mining unit (§ 971.501); (6) A description of the methods and...
The Readability of Information Literacy Content on Academic Library Web Sites
ERIC Educational Resources Information Center
Lim, Adriene
2010-01-01
This article reports on a study addressing the readability of content on academic libraries' Web sites, specifically content intended to improve users' information literacy skills. Results call for recognition of readability as an evaluative component of text in order to better meet the needs of diverse user populations. (Contains 8 tables.)
Separate but Equal? A Comparison of Content on Library Web Pages and Their Text Versions
ERIC Educational Resources Information Center
Hazard, Brenda L.
2008-01-01
This study examines the Web sites of the Association of Research Libraries member libraries to determine the presence of a separate text version of the default graphical homepage. The content of the text version and the homepage is compared. Of 121 Web sites examined, twenty libraries currently offer a text version. Ten sites maintain wholly…
e-Ana and e-Mia: A Content Analysis of Pro–Eating Disorder Web Sites
Schenk, Summer; Wilson, Jenny L.; Peebles, Rebecka
2010-01-01
Objectives. The Internet offers Web sites that describe, endorse, and support eating disorders. We examined the features of pro–eating disorder Web sites and the messages to which users may be exposed. Methods. We conducted a systematic content analysis of 180 active Web sites, noting site logistics, site accessories, “thinspiration” material (images and prose intended to inspire weight loss), tips and tricks, recovery, themes, and perceived harm. Results. Practically all (91%) of the Web sites were open to the public, and most (79%) had interactive features. A large majority (84%) offered pro-anorexia content, and 64% provided pro-bulimia content. Few sites focused on eating disorders as a lifestyle choice. Thinspiration material appeared on 85% of the sites, and 83% provided overt suggestions on how to engage in eating-disordered behaviors. Thirty-eight percent of the sites included recovery-oriented information or links. Common themes were success, control, perfection, and solidarity. Conclusions. Pro–eating disorder Web sites present graphic material to encourage, support, and motivate site users to continue their efforts with anorexia and bulimia. Continued monitoring will offer a valuable foundation to build a better understanding of the effects of these sites on their users. PMID:20558807
Produce and Consume Linked Data with Drupal!
NASA Astrophysics Data System (ADS)
Corlosquet, Stéphane; Delbru, Renaud; Clark, Tim; Polleres, Axel; Decker, Stefan
Currently a large number of Web sites are driven by Content Management Systems (CMS) which manage textual and multimedia content but also - inherently - carry valuable information about a site's structure and content model. Exposing this structured information to the Web of Data has so far required considerable expertise in RDF and OWL modelling and additional programming effort. In this paper we tackle one of the most popular CMS: Drupal. We enable site administrators to export their site content model and data to the Web of Data without requiring extensive knowledge on Semantic Web technologies. Our modules create RDFa annotations and - optionally - a SPARQL endpoint for any Drupal site out of the box. Likewise, we add the means to map the site data to existing ontologies on the Web with a search interface to find commonly used ontology terms. We also allow a Drupal site administrator to include existing RDF data from remote SPARQL endpoints on the Web in the site. When brought together, these features allow networked RDF Drupal sites that reuse and enrich Linked Data. We finally discuss the adoption of our modules and report on a use case in the biomedical field and the current status of its deployment.
Yager, Douglas B.; Stanton, Mark R.; Choate, LaDonna M.; Burchell,
2009-01-01
Mine planning efforts have historically overlooked the possible acid neutralizing capacity (ANC) that local igneous rocks can provide to help neutralize acidmine drainage. As a result, limestone has been traditionally hauled to mine sites for use in neutralizing acid drainage. Local igneous rocks, when used as part of mine life-cycle planning and acid mitigation strategy, may reduce the need to transport limestone to mine sites because these rocks can contain acid neutralizing minerals. Igneous hydrothermal events often introduce moderately altered mineral assemblages peripheral to more intensely altered rocks that host metal-bearing veins and ore bodies. These less altered rocks can contain ANC minerals (calcite-chlorite-epidote) and are referred to as a propylitic assemblage. In addition, the carbon contents of soils in areas of new mining or those areas undergoing restoration have been historically unknown. Soil organic carbon is an important constituent to characterize as a soil recovery benchmark that can be referred to during mine cycle planning and restoration.
This study addresses the mineralogy, ANC, and leachate chemistry of propylitic volcanic rocks that host polymetallic mineralization in the Animas River watershed near the historical Silverton, Colorado, mining area. Acid titration tests on volcanic rocks containing calcite (2 – 20 wt %) and chlorite (6 – 25 wt %), have ANC ranging from 4 – 146 kg/ton CaCO3 equivalence. Results from a 6-month duration, kinetic reaction vessel test containing layered pyritic mine waste and underlying ANC volcanic rock (saturated with deionized water) indicate that acid generating mine waste (pH 2.4) has not overwhelmed the ANC of propylitic volcanic rocks (pH 5.8). Sequential leachate laboratory experiments evaluated the concentration of metals liberated during leaching. Leachate concentrations of Cu-Zn-As-Pb for ANC volcanic rock are one-to-three orders of magnitude lower when compared to leached solution from mine waste used in the kinetic reaction vessel test. This finding suggests that mine waste and not ANC rock may generate the majority of leachable metals in a field scenario.
The organic carbon content of naturally reclaimed soils derived from weathering of propylitically-altered andesite was determined in catchments where ANC studies were initiated. Soils were found to have total carbon concentrations (TOC) that exceed global average soil TOC abundances by as much as 1.5 – 5 times. These data support an environmental management system involving use of ANC rocks as part of life-cycle mine planning to reduce post-mine closure acid mitigation measures. Carbon contents of undisturbed soils in mined catchments can possibly be used to validate post-reclamation success and help quantify carbon sequestration for CO2 emission offset trading as carbon markets mature.
McCauley, Jacob R; Bouldin, Jennifer L
2016-06-01
The Rush Mining District along the Buffalo River in Arkansas has a significant history of zinc and lead mining operations. The tails and spoils of these operations deposit heavy amounts of raw ore into streams. One element commonly found in the earth's crust that becomes a minor constituent of the deposition is cadmium. Periphyton samples from Rush Creek and Clabber Creek, two creeks within the Rush Mining District were measured for cadmium as well as two creeks with no history of mining, Spring Creek and Water Creek. Periphyton samples from Rush and Clabber Creek contained mean cadmium concentrations of 436.6 ± 67.3 and 93.38 ± 8.67 µg/kg, respectively. Spring Creek and Water Creek had a mean cadmium concentration of 40.49 ± 3.40 and 41.78 ± 3.99 µg/kg within periphyton. The results indicate increased metal concentrations in algal communities from mined areas. As periphyton is the base of the aquatic food chain, it acts as a conduit for movement of cadmium in the food web.
The ATLAS Public Web Pages: Online Management of HEP External Communication Content
NASA Astrophysics Data System (ADS)
Goldfarb, S.; Marcelloni, C.; Eli Phoboo, A.; Shaw, K.
2015-12-01
The ATLAS Education and Outreach Group is in the process of migrating its public online content to a professionally designed set of web pages built on the Drupal [1] content management system. Development of the front-end design passed through several key stages, including audience surveys, stakeholder interviews, usage analytics, and a series of fast design iterations, called sprints. Implementation of the web site involves application of the html design using Drupal templates, refined development iterations, and the overall population of the site with content. We present the design and development processes and share the lessons learned along the way, including the results of the data-driven discovery studies. We also demonstrate the advantages of selecting a back-end supported by content management, with a focus on workflow. Finally, we discuss usage of the new public web pages to implement outreach strategy through implementation of clearly presented themes, consistent audience targeting and messaging, and the enforcement of a well-defined visual identity.
Flood-Grady, Elizabeth; Paige, Samantha R; Karimipour, Nicki; Harris, Paul A; Cottler, Linda B; Krieger, Janice L
2017-12-01
There is a dearth of literature providing guidance on how to effectively communicate about clinical research (CR). Using the transactional model of communication, a content analysis of the investigator (n=62) and participant (n=18) Web sites of institutions funded through the National Institutes of Health Clinical and Translational Science Award (CTSA) was conducted to identify their strategies (e.g., messages) for communicating about CR participation. CTSAs targeted investigators with CR participation content across the main Web sites, although most CTSAs (n=55; 88.7%) also included CR participation content for participants. In total, 18 CTSAs (29%) hosted participant Web sites. Participant sites included 13 message types about CR participation (e.g., registry enrollment) and 5 additional channels (e.g., email, phone number) to communicate about CR. However, many CTSA participant Web sites excluded information explaining the CR process and offered CR content exclusively in English. CTSAs should identify their target audience and design strategies (e.g., messages, channels) accordingly.
Department of Agriculture, Food Safety and Inspection Service
... FSIS Forms Administrative Forms Standard Forms Skip Navigation Web Content Viewer (JSR 286) Actions ${title} Loading... Information ... resources and information on Siluriformes fish, including catfish Web Content Viewer (JSR 286) Actions ${title} Loading... Information ...
Development of Database for Accident Analysis in Indian Mines
NASA Astrophysics Data System (ADS)
Tripathy, Debi Prasad; Guru Raghavendra Reddy, K.
2016-10-01
Mining is a hazardous industry and high accident rates associated with underground mining is a cause of deep concern. Technological developments notwithstanding, rate of fatal accidents and reportable incidents have not shown corresponding levels of decline. This paper argues that adoption of appropriate safety standards by both mine management and the government may result in appreciable reduction in accident frequency. This can be achieved by using the technology in improving the working conditions, sensitising workers and managers about causes and prevention of accidents. Inputs required for a detailed analysis of an accident include information on location, time, type, cost of accident, victim, nature of injury, personal and environmental factors etc. Such information can be generated from data available in the standard coded accident report form. This paper presents a web based application for accident analysis in Indian mines during 2001-2013. An accident database (SafeStat) prototype based on Intranet of the TCP/IP agreement, as developed by the authors, is also discussed.
Fernandes, Christabelle E G; Malik, Ashish; Jineesh, V K; Fernandes, Sheryl O; Das, Anindita; Pandey, Sunita S; Kanolkar, Geeta; Sujith, P P; Velip, Dhillan M; Shaikh, Shagufta; Helekar, Samita; Gonsalves, Maria Judith; Nair, Shanta; LokaBharathi, P A
2015-08-01
The coastal waters of Goa and Ratnagiri lying on the West coast of India are influenced by terrestrial influx. However, Goa is influenced anthropogenically by iron-ore mining, while Ratnagiri is influenced by deposition of heavy minerals containing iron brought from the hinterlands. We hypothesize that there could be a shift in biological response along with changes in network of interactions between environmental and biological variables in these mining and non-mining impacted regions, lying 160 nmi apart. Biological and environmental parameters were analyzed during pre-monsoon season. Except silicates, the measured parameters were higher at Goa and related significantly, suggesting bacteria centric, detritus-driven region. At Ratnagiri, phytoplankton biomass related positively with silicate suggesting a region dominated by primary producers. This dominance perhaps got reflected as a higher tertiary yield. Thus, even though the regions are geographically proximate, the different biological response could be attributed to the differences in the web of interactions between the measured variables.
A Node Linkage Approach for Sequential Pattern Mining
Navarro, Osvaldo; Cumplido, René; Villaseñor-Pineda, Luis; Feregrino-Uribe, Claudia; Carrasco-Ochoa, Jesús Ariel
2014-01-01
Sequential Pattern Mining is a widely addressed problem in data mining, with applications such as analyzing Web usage, examining purchase behavior, and text mining, among others. Nevertheless, with the dramatic increase in data volume, the current approaches prove inefficient when dealing with large input datasets, a large number of different symbols and low minimum supports. In this paper, we propose a new sequential pattern mining algorithm, which follows a pattern-growth scheme to discover sequential patterns. Unlike most pattern growth algorithms, our approach does not build a data structure to represent the input dataset, but instead accesses the required sequences through pseudo-projection databases, achieving better runtime and reducing memory requirements. Our algorithm traverses the search space in a depth-first fashion and only preserves in memory a pattern node linkage and the pseudo-projections required for the branch being explored at the time. Experimental results show that our new approach, the Node Linkage Depth-First Traversal algorithm (NLDFT), has better performance and scalability in comparison with state of the art algorithms. PMID:24933123
Corner-cutting mining assembly
Bradley, J.A.
1981-07-01
This invention resulted from a contract with the United States Department of Energy and relates to a mining tool. More particularly, the invention relates to an assembly capable of drilling a hole having a square cross-sectional shape with radiused corners. In mining operations in which conventional auger-type drills are used to form a series of parallel, cylindrical holes in a coal seam, a large amount of coal remains in place in the seam because the shape of the holes leaves thick webs between the holes. A higher percentage of coal can be mined from a seam by a means capable of drilling holes having a substantially square cross section. It is an object of this invention to provide an improved mining apparatus by means of which the amount of coal recovered from a seam deposit can be increased. Another object of the invention is to provide a drilling assembly which cuts corners in a hole having a circular cross section. These objects and other advantages are attained by a preferred embodiment of the invention.
Mercury flow through an Asian rice-based food web.
Abeysinghe, Kasun S; Qiu, Guangle; Goodale, Eben; Anderson, Christopher W N; Bishop, Kevin; Evers, David C; Goodale, Morgan W; Hintelmann, Holger; Liu, Shengjie; Mammides, Christos; Quan, Rui-Chang; Wang, Jin; Wu, Pianpian; Xu, Xiao-Hang; Yang, Xiao-Dong; Feng, Xinbin
2017-10-01
Mercury (Hg) is a globally-distributed pollutant, toxic to humans and animals. Emissions are particularly high in Asia, and the source of exposure for humans there may also be different from other regions, including rice as well as fish consumption, particularly in contaminated areas. Yet the threats Asian wildlife face in rice-based ecosystems are as yet unclear. We sought to understand how Hg flows through rice-based food webs in historic mining and non-mining regions of Guizhou, China. We measured total Hg (THg) and methylmercury (MeHg) in soil, rice, 38 animal species (27 for MeHg) spanning multiple trophic levels, and examined the relationship between stable isotopes and Hg concentrations. Our results confirm biomagnification of THg/MeHg, with a high trophic magnification slope. Invertivorous songbirds had concentrations of THg in their feathers that were 15x and 3x the concentration reported to significantly impair reproduction, at mining and non-mining sites, respectively. High concentrations in specialist rice consumers and in granivorous birds, the later as high as in piscivorous birds, suggest rice is a primary source of exposure. Spiders had the highest THg concentrations among invertebrates and may represent a vector through which Hg is passed to vertebrates, especially songbirds. Our findings suggest there could be significant population level health effects and consequent biodiversity loss in sensitive ecosystems, like agricultural wetlands, across Asia, and invertivorous songbirds would be good subjects for further studies investigating this possibility. Copyright © 2017 Elsevier Ltd. All rights reserved.
Greb, S.F.; Popp, J.T.
1999-01-01
The Pond Creek seam is one of the leading producers of coal in the Eastern Kentucky Coal Field. The geologic factors that affect mining were investigated in several underground mines and categorized in terms of coal thickness, coal quality, and roof control. The limits of mining and thick coal are defined by splitting along the margin of the coal body. Within the coal body, local thickness variation occurs because of (1) leader coal benches filling narrow, elongated depressions, (2) rider coal benches coming near to or merging with the main bench, (3) overthrust coal benches being included along paleochannel margins, (4) cutouts occuring beneath paleochannels, and (5) very hard and unusual rock partings occuring along narrow, elongated trends. In the study area, the coal is mostly mined as a compliance product: sulfur contents are less than 1% and ash yields are less than 10%. Local increases in sulfur occur beneath sandstones, and are inferred to represent post-depositional migration of fluids through porous sands into the coal. Run-of-mine quality is also affected by several mine-roof conditions and trends of densely concentrated rock partings, which lead to increased in- and out-of-seam dilution and overall ash content of the mined coal. Roof control is largely a function of a heterolithic facies mosaic of coastal-estuarine origin, regional fracture trends, and unloading stress related to varying mine depth beneath the surface. Lateral variability of roof facies is the rule in most mines. The largest falls occur beneath modern valleys and parallel fractures, along paleochannel margins, within tidally affected 'stackrock,' and beneath rider coals. Shale spalling, kettlebottoms, and falls within other more isolated facies also occur. Many of the lithofacies, and falls related to bedding weaknesses within or between lithofacies, occur along northeast-southwest trends, which can be projected in advance of mining. Fracture-related falls occur independently of lithofacies trends along northwest-southeast trends, especially beneath modern valleys where overburden thickness decreases sharply. Differentiating roof falls related to these trends can aid in predicting roof quality in advance of mining.The Pond Creek-Lower Elkhorn seam has been an important exploration target because it typically has very low sulfur contents and ash yields. Geologic research in several large Pond Creek mines suggested variability in roof quality and coal thickness. Due to mine access, geologic problems encountered during mining are documented and described.
Environmental pleural plaques in residents of a Quebec chrysotile mining town
DOE Office of Scientific and Technical Information (OSTI.GOV)
Churg, A.; DePaoli, L.
1988-07-01
We report four cases of pleural plaques found at autopsy in individuals who resided in or near the chrysotile mining town of Thetford Mines, Quebec, and who had never been employed in the chrysotile mining and milling industry. Three of these patients were farmers, and one was a road construction worker. Lung asbestos content of these cases was compared with that of a group of nine persons living in the same vicinity who did not have pleural plaques. The plaque group was found to have an equal chrysotile content but about a fourfold elevation in median tremolite content, a statisticallymore » significant increase. Fiber sizes were the same in both groups. Also, one plaque case had an elevated level of relatively long titanium oxide fibers. These observations suggest that environmental pleural plaques in this region of Quebec are probably caused by exposure to tremolite derived from local soil and rock and that other types of mineral fibers such as titanium oxide may occasionally also be the cause of such lesions.« less
Salt Content Determination for Bentonite Mine Spoil: Saturation Extracts Versus 1:5 Extracts
Marguerite E. Voorhees; Daniel W. Uresk
2004-01-01
The reliability of estimating salt content in saturated extracts from 1:5 (1spoil:5water) extract levels for bentonite mine spoil was examined by regression analyses. Nine chemical variables were examined that included pH, EC, Ca++, Mg++, Na+, K+, HCO3-, SO4-, and Cl-. Ion concentrations from 1:5 extracts were estimated with high predictability for Ca++, Mg++, Na+, SO4...
Review of Extracting Information From the Social Web for Health Personalization
Karlsen, Randi; Bonander, Jason
2011-01-01
In recent years the Web has come into its own as a social platform where health consumers are actively creating and consuming Web content. Moreover, as the Web matures, consumers are gaining access to personalized applications adapted to their health needs and interests. The creation of personalized Web applications relies on extracted information about the users and the content to personalize. The Social Web itself provides many sources of information that can be used to extract information for personalization apart from traditional Web forms and questionnaires. This paper provides a review of different approaches for extracting information from the Social Web for health personalization. We reviewed research literature across different fields addressing the disclosure of health information in the Social Web, techniques to extract that information, and examples of personalized health applications. In addition, the paper includes a discussion of technical and socioethical challenges related to the extraction of information for health personalization. PMID:21278049
Christen, Matthias; Deutsch, Samuel; Christen, Beat
2015-08-21
Recent advances in synthetic biology have resulted in an increasing demand for the de novo synthesis of large-scale DNA constructs. Any process improvement that enables fast and cost-effective streamlining of digitized genetic information into fabricable DNA sequences holds great promise to study, mine, and engineer genomes. Here, we present Genome Calligrapher, a computer-aided design web tool intended for whole genome refactoring of bacterial chromosomes for de novo DNA synthesis. By applying a neutral recoding algorithm, Genome Calligrapher optimizes GC content and removes obstructive DNA features known to interfere with the synthesis of double-stranded DNA and the higher order assembly into large DNA constructs. Subsequent bioinformatics analysis revealed that synthesis constraints are prevalent among bacterial genomes. However, a low level of codon replacement is sufficient for refactoring bacterial genomes into easy-to-synthesize DNA sequences. To test the algorithm, 168 kb of synthetic DNA comprising approximately 20 percent of the synthetic essential genome of the cell-cycle bacterium Caulobacter crescentus was streamlined and then ordered from a commercial supplier of low-cost de novo DNA synthesis. The successful assembly into eight 20 kb segments indicates that Genome Calligrapher algorithm can be efficiently used to refactor difficult-to-synthesize DNA. Genome Calligrapher is broadly applicable to recode biosynthetic pathways, DNA sequences, and whole bacterial genomes, thus offering new opportunities to use synthetic biology tools to explore the functionality of microbial diversity. The Genome Calligrapher web tool can be accessed at https://christenlab.ethz.ch/GenomeCalligrapher .
Composition and trace element content of coal in Taiwan
Tsai, L.-Y.; Chen, C.-F.; Finkelman, R.B.
2005-01-01
To investigate the trace element contents of local coal, four coal samples were collected from operating mines in NW Taiwan. Detailed petrographic and chemical characterization analyses were then conducted. Analytical results indicate that (1) the samples were high volatile bituminous coal in rank with ash content ranging from 4.2 to 14.4% and with moisture content ranging from 2.7 to 4.6%; (2) the macerals were mostly composed of vitrinite with vitrinite reflectance less than 0.8%; (3) the sample of Wukeng mine has the highest Fe2O3 (29.5%), TI (54.8 ppm), Zn (140 ppm), and As (697 ppm) contents in ash and Hg (2.3 ppm) in the coal. If used properly, these coals should not present health hazards.
Web Prep: How to Prepare NAS Reports For Publication on the Web
NASA Technical Reports Server (NTRS)
Walatka, Pamela; Balakrishnan, Prithika; Clucas, Jean; McCabe, R. Kevin; Felchle, Gail; Brickell, Cristy
1996-01-01
This document contains specific advice and requirements for NASA Ames Code IN authors of NAS reports. Much of the information may be of interest to other authors writing for the Web. WebPrep has a graphic Table of Contents in the form of a WebToon, which simulates a discussion between a scientist and a Web publishing consultant. In the WebToon, Frequently Asked Questions about preparing reports for the Web are linked to relevant text in the body of this document. We also provide a text-only Table of Contents. The text for this document is divided into chapters: each chapter corresponds to one frame of the WebToons. The chapter topics are: converting text to HTML, converting 2D graphic images to gif, creating imagemaps and tables, converting movie and audio files to Web formats, supplying 3D interactive data, and (briefly) JAVA capabilities. The last chapter is specifically for NAS staff authors. The Glossary-Index lists web related words and links to topics covered in the main text.
One EPA Web Guidances and Checklists
These One EPA Web resources are available to editors with Web Guide access. Learn about content development, web council and EIC responsibilities, audiences and top tasks, website format and structure, and site review and approval.
Hansen, Henrik K; Yianatos, Juan B; Ottosen, Lisbeth M
2005-09-01
Mine tailing from the El Teniente-Codelco copper mine situated in VI Region of Chile was analysed in order to evaluate the mobility and speciation of copper in the solid material. Mine tailing was sampled after the rougher flotation circuits, and the copper content was measured to 1150 mg kg (-1) dry matter. This tailing was segmented into fractions of different size intervals: 0-38, 38-45, 45-53, 53-75, 75-106, 106-150, 150-212, and >212 microm, respectively. Copper content determination, sequential chemical extraction, and desorption experiments were carried out for each size interval in order to evaluate the speciation of copper. It was found that the particles of smallest size contained 50-60% weak acid leachable copper, whereas only 32% of the copper found in largest particles could be leached in weak acid. Copper oxides and carbonates were the dominating species in the smaller particles, and the larger particles contained considerable amounts of sulphides.
Church, Stan E.; Von Guerard, Paul; Finger, Susan E.
2007-01-01
This publication comprises a Volume Contents of chapters (listed below) and a CD-ROM of data (contents shown in column at right). The Animas River watershed in southwest Colorado is one of many watersheds in the western United States where historical mining has left a legacy of acid mine drainage and elevated concentrations of potentially toxic trace elements in surface streams. U.S. Geological Survey scientists have completed a major assessment of the environmental effects of historical mining in the Animas River watershed focusing on the area upstream of Silverton, Colo.?the Mineral Creek, Cement Creek, and upper Animas River basins. The study demonstrated how the watershed approach can be used to assess and rank mining-affected sites for possible cleanup. The study was conducted in collaboration with State and Federal land-management agencies and regional stakeholders groups. This book is available for purchase at Information Services, U.S. Geological Survey (1-888-ASK-USGS).
Methodologies for Crawler Based Web Surveys.
ERIC Educational Resources Information Center
Thelwall, Mike
2002-01-01
Describes Web survey methodologies used to study the content of the Web, and discusses search engines and the concept of crawling the Web. Highlights include Web page selection methodologies; obstacles to reliable automatic indexing of Web sites; publicly indexable pages; crawling parameters; and tests for file duplication. (Contains 62…
Mineralogical study of stream waters and efflorescent salts in Sierra Minera, SE Spain
NASA Astrophysics Data System (ADS)
Pérez-Sirvent, Carmen; Garcia-Lorenzo, Maria luz; Martinez-Sanchez, Maria Jose; Hernandez, Carmen; Hernandez-Cordoba, Manuel
2015-04-01
Trace elements contained in the residues from mining and metallurgical operations are often dispersed by wind and/or water after their disposal. These areas have severe erosion problems caused by water run-off in which soil and mine spoil texture, landscape topography and regional and microclimate play an important role. Water pollution by dissolved metals in mining areas has mainly been associated with the oxidation of sulphide-bearing minerals exposed to weathering conditions, resulting in low quality effluents of acidic pH and containing a high level of dissolved metals. The studied area, Sierra Minera, is close to the mining region of La Unión (Murcia, SE Spain). This area constituted an important mining centre for more than 2500 years, ceasing activity in 1991. The ore deposits of this zone have iron, lead and zinc as the main metal components. Studied area showed a lot of contaminations sources, formed by mining steriles, waste piles and foundry residues. As a consequence of the long period of mining activity, large volumes of wastes were generated during the mineral concentration and smelting processes. Historically, these wastes were dumped into watercourses, filling riverbeds and contaminating their surroundings. 40 sediment samples were collected from the area affected by mining exploitations, and at increasing distances from the contamination sources in 4 zones In addition, 36 surficial water samples were collected after a rain episode The Zn and Fe content was determined by flame atomic absorption spectrometry (FAAS). The Pb and Cd content was determined by electrothermal atomization atomic absorption spectrometry (ETAAS). The As content was measured by atomic fluorescence spectrometry using an automated continuous flow hydride generation spectrometer and Al content was determined by ICP-MS. Mineralogical composition of the samples was made by X Ray Diffraction (XRD) analysis using Cu-Kα radiation with a PW3040 Philips Diffractometer. Zone A: Water sample collected in A5 is strongly influenced by a tailing dump, and showed high trace element contents. In addition, is influenced by the sea water and then showed high bromide, chloride, sodium and magnesium content, together with a basic pH.The DRX results of evaporate water showed that halite, hexahydrite and gypsum are present: halite corroborates the sea influence and gypsum and hexahydrite the importance of soluble sulphates. A9 water showed acid pH and high trace elements content; is influenced by the tailing dump and also by waters from El Beal gully watercourse, transporting materials from Sierra Minera Waters affected by secondary contamination are influenced by mining wastes, the sea water and also are affected by agricultural activities (nitrate content). These waters have been mixed with carbonate materials, present in the zone increasing the pH. Some elements have precipitated, such as Cu and Pb, while Cd, Zn and As are soluble. The DRX analysis in the evaporate if A14 showed that halite and gypsum are present: halite confirms the seawater influence and gypsum the relationship between calcium and sulphates A2 and A6 waters are affected by tertiary contamination and showed basic pH, soluble carbonates and lower trace element content. Only Zn, Cd and Al are present. Zone B: All waters are strongly affected by mining activities and showed: acid pH, high trace element content and high content of soluble sulphates. The evaporate of B8 and B12 showed the presence of soluble sulphates: gypsum, halite, bianchite, paracoquimbite, halotrichite and siderotil in B8; gypsum, bianchite, paracoquimbite and coquimbite in B12; gypsum, hexahydrite, carnalite, bianchite, copiapite and sideroti in B10 and polihalite, gypsum, bianchite, coquimbite and paracoquimbite in B14. All the sampling points collected in Zone C are affected by primary contamination, because there are a lot of tailing dumps and sampling points are located close to them. C1 showed high trace element content because is a reception point of a lot of tailing dumps. Water samples from C3 to C8 also had acid pH and high trace element content, particularly As (remains soluble) and Zn and Cd (high mobility). In addition, they showed high soluble sulphates. C2 water showed neutral pH, soluble carbonate and low trace element content because is influenced by a stabilised tailing dump. However, the As remains soluble. Zone D: All waters collected in this zone showed acid pH and high trace element content, mainly Zn, Cd and As. Some differences were found from the high and the low part: samples located in the lower part (D2-D7) showed higher As content while Zn is higher in the high part (D8-D13) The DRX analysis in evaporates suggest that in D4 copiapite, coquimbite, gypsum, bianchite and ferrohexahydrite are formed and in D11 gypsum, bianchite, halotrichite and siderotil. D1 is affected by secondary contamination, which showed higher pH (still acid) and lower content in soluble salts and trace elements.
Quality of web-based information on cannabis addiction.
Khazaal, Yasser; Chatton, Anne; Cochand, Sophie; Zullino, Daniele
2008-01-01
This study evaluated the quality of Web-based information on cannabis use and addiction and investigated particular content quality indicators. Three keywords ("cannabis addiction," "cannabis dependence," and "cannabis abuse") were entered into two popular World Wide Web search engines. Websites were assessed with a standardized proforma designed to rate sites on the basis of accountability, presentation, interactivity, readability, and content quality. "Health on the Net" (HON) quality label, and DISCERN scale scores were used to verify their efficiency as quality indicators. Of the 94 Websites identified, 57 were included. Most were commercial sites. Based on outcome measures, the overall quality of the sites turned out to be poor. A global score (the sum of accountability, interactivity, content quality and esthetic criteria) appeared as a good content quality indicator. While cannabis education Websites for patients are widespread, their global quality is poor. There is a need for better evidence-based information about cannabis use and addiction on the Web.
[Legal aspects of Web 2.0 in the health field].
Beslay, Nathalie; Jeunehomme, Marie
2009-10-01
Web 2.0 sites are considered to be hosting providers and not publishers of user-generated content. The liability of hosting providers' liability is defined by the law enacted on June 21, 2004, on confidence in the digital economy. Hosting providers must promptly remove the information they host or make its access impossible once they are informed of its illegality. They are required to obtain and retain data to enable identification of any person who has contributed to content hosted by them. The liability of hosting providers has arisen in numerous disputes about user-produced content in various situations (discussion lists, blogs, etc.). The National Board of Physicians has developed specific ethical guidelines for web sites devoted to health issues and specifically for physician-authored content. The National Board of Physicians acknowledges that physicians can present themselves, their office, and their specific practice on their web site, notwithstanding any restrictions otherwise applicable to advertising.
NASA Astrophysics Data System (ADS)
Borne, K. D.
2009-12-01
The emergence of e-Science over the past decade as a paradigm for Internet-based science was an inevitable evolution of science that built upon the web protocols and access patterns that were prevalent at that time, including Web Services, XML-based information exchange, machine-to-machine communication, service registries, the Grid, and distributed data. We now see a major shift in web behavior patterns to social networks, user-provided content (e.g., tags and annotations), ubiquitous devices, user-centric experiences, and user-led activities. The inevitable accrual of these social networking patterns and protocols by scientists and science projects leads to U-Science as a new paradigm for online scientific research (i.e., ubiquitous, user-led, untethered, You-centered science). U-Science applications include components from semantic e-science (ontologies, taxonomies, folksonomies, tagging, annotations, and classification systems), which is much more than Web 2.0-based science (Wikis, blogs, and online environments like Second Life). Among the best examples of U-Science are Citizen Science projects, including Galaxy Zoo, Stardust@Home, Project Budburst, Volksdata, CoCoRaHS (the Community Collaborative Rain, Hail and Snow network), and projects utilizing Volunteer Geographic Information (VGI). There are also scientist-led projects for scientists that engage a wider community in building knowledge through user-provided content. Among the semantic-based U-Science projects for scientists are those that specifically enable user-based annotation of scientific results in databases. These include the Heliophysics Knowledgebase, BioDAS, WikiProteins, The Entity Describer, and eventually AstroDAS. Such collaborative tagging of scientific data addresses several petascale data challenges for scientists: how to find the most relevant data, how to reuse those data, how to integrate data from multiple sources, how to mine and discover new knowledge in large databases, how to represent and encode the new knowledge, and how to curate the discovered knowledge. This talk will address the emergence of U-Science as a type of Semantic e-Science, and will explore challenges, implementations, and results. Semantic e-Science and U-Science applications and concepts will be discussed within the context of one particular implementation (AstroDAS: Astronomy Distributed Annotation System) and its applicability to petascale science projects such as the LSST (Large Synoptic Survey Telescope), coming online within the next few years.
ERIC Educational Resources Information Center
Garcia-Barriocanal, Elena; Sicilia, Miguel-Angel; Sanchez-Alonso, Salvador; Lytras, Miltiadis
2011-01-01
Web 2.0 technologies can be considered a loosely defined set of Web application styles that foster a kind of media consumer more engaged, and usually active in creating and maintaining Internet contents. Thus, Web 2.0 applications have resulted in increased user participation and massive user-generated (or user-published) open multimedia content,…
ERIC Educational Resources Information Center
Harris, Christopher
2006-01-01
Blogs and the whole gamut of interactive, Web-based tools are redefining how we connect with people and content on the Internet. This is the widely discussed next step known as Web 2.0, a truly revolutionary concept in which pervasive interactivity, where multitudes of users online actively exchange or contribute content, will transform the very…
DOE Office of Scientific and Technical Information (OSTI.GOV)
Kamel Boulos, Maged; Sanfilippo, Antonio P.; Corley, Courtney D.
2010-03-17
This paper explores techno-social predictive analytics (TPA) and related methods for Web “data mining” where users’ posts and queries are garnered from Social Web (“Web 2.0”) tools such as blogs, microblogging and social networking sites to form coherent representations of real-time health events. The paper includes a brief introduction to commonly used Social Web tools such as mashups and aggregators, and maps their exponential growth as an open architecture of participation for the masses and an emerging way to gain insight about people’s collective health status of whole populations. Several health related tool examples are described and demonstrated as practicalmore » means through which health professionals might create clear location specific pictures of epidemiological data such as flu outbreaks.« less
ERIC Educational Resources Information Center
Gökçearslan, Sahin; Karademir, Tugra; Korucu, Agah Tugrul
2017-01-01
Technological Pedagogical Content Knowledge, one of the frameworks proposed in order to popularize the use of technology in a classroom environment, has been customized and has taken the form of Web Pedagogical Content Knowledge. The Relational Screening Model was used in this study. It aims to determine whether a profile of preservice teachers…
Stopping Web Plagiarists from Stealing Your Content
ERIC Educational Resources Information Center
Goldsborough, Reid
2004-01-01
This article gives tips on how to avoid having content stolen by plagiarists. Suggestions include: using a Web search service such as Google to search for unique strings of text at the individuals site to uncover other sites with the same content; buying a infringement-detection program; or hiring a public relations firm to do the work. There are…
Creating an Internal Content Management System
ERIC Educational Resources Information Center
Sennema, Greg
2004-01-01
In this article, the author talks about an internal content management system that they have created at Calvin College. It is a hybrid of CMS and intranet that organizes Web site content and a variety of internal tools to help librarians complete their daily tasks. Hobbes is a Web-based tool that uses Common Gateway Interface (CGI) scripts written…
Normalizing 13C values of animal tissue for lipid content is necessary to accurately interpret food web relationships from stable isotope analysis. This is because lipids are 13C-depleted relative to proteins and carbohydrates, and because lipid content varies among speci...
"UML Quiz": Automatic Conversion of Web-Based E-Learning Content in Mobile Applications
ERIC Educational Resources Information Center
von Franqué, Alexander; Tellioglu, Hilda
2014-01-01
Many educational institutions use Learning Management Systems to provide e-learning content to their students. This often includes quizzes that can help students to prepare for exams. However, the content is usually web-optimized and not very usable on mobile devices. In this work a native mobile application ("UML Quiz") that imports…
World Wide Web Homepages: An Examination of Content and Audience.
ERIC Educational Resources Information Center
Reynolds, Betty; And Others
This paper shows how the content of a World Wide Web page is selected and how an examination of the intended audience influences content. Examples from the New Mexico Tech (NMT) Library homepage show what sources are selected and what level of detail is appropriate for the intended audience. Six fundamental functions of libraries and information…
,
1996-01-01
Descriptions of mines, prospects, and mineral occurrences in the Alaska Resource Data File (ARDF) are published for individual U.S. Geological Survey 1:250,000 scale quadrangles in Alaska (see accompanying map) and are available for downloading from USGS World Wide Web site: http://www-rnrs-ak.wr.usgs.gov/ardf.These descriptions are divided into a number of fields which describe features of each mine, prospect, or mineral occurrence. These descriptions were complied from published literature and from unpublished reports and data from industry, the U.S. Bureau of Mines, and the U.S. Geological Survey and other sources. Compilation of this database is an ongoing process and each report is essentially a progress report. The authors of the individual quadrangle reports would appreciate any corrections or additional information that users may be able to contribute.
Hymenoptera Genome Database: integrating genome annotations in HymenopteraMine
Elsik, Christine G.; Tayal, Aditi; Diesh, Colin M.; Unni, Deepak R.; Emery, Marianne L.; Nguyen, Hung N.; Hagen, Darren E.
2016-01-01
We report an update of the Hymenoptera Genome Database (HGD) (http://HymenopteraGenome.org), a model organism database for insect species of the order Hymenoptera (ants, bees and wasps). HGD maintains genomic data for 9 bee species, 10 ant species and 1 wasp, including the versions of genome and annotation data sets published by the genome sequencing consortiums and those provided by NCBI. A new data-mining warehouse, HymenopteraMine, based on the InterMine data warehousing system, integrates the genome data with data from external sources and facilitates cross-species analyses based on orthology. New genome browsers and annotation tools based on JBrowse/WebApollo provide easy genome navigation, and viewing of high throughput sequence data sets and can be used for collaborative genome annotation. All of the genomes and annotation data sets are combined into a single BLAST server that allows users to select and combine sequence data sets to search. PMID:26578564
NASA Astrophysics Data System (ADS)
Abdaal, Ahmed; Jordan, Gyozo; Bartha, Andras; Fugedi, Ubul
2013-04-01
The Mine Waste Directive 2006/21/EC requires the risk-based inventory of all mine waste sites in Europe. The geochemical documentation concerning inert classification and ranking of the mine wastes requires detailed field study and laboratory testing and analyses of waste material to assess the Acid Mine Drainage potential and toxic element mobility. The procedure applied in this study used a multi-level decision support scheme including: 1) expert judgment, 2) data review, 3) representative field sampling and laboratory analysis of formations listed in the Inert Mining Waste List, and 4) requesting available laboratory analysis data from selected operating mines. Based on expert judgment, the listed formations were classified into three categories. A: inert B: probably inert, but has to be checked, C: probably not inert, has to be examined. This paper discusses the heavy metal contamination risk assessment (RA) in leached quarry-mine waste sites in Hungary. In total 34 mine waste sites (including tailing lagoons and heaps of both abandoned mines and active quarries) have been selected for scientific testing using the EU Pre-selection Protocol. Over 93 field samples have been collected from the mine sites including Ore (Andesite and Ryolite), Coal (Lignite, black and brown coals), Peat, Alginite, Bauxite, Clay and Limestone. Laboratory analyses of the total toxic element content (aqua regia extraction), the mobile toxic element content (deionized water leaching) and the analysis of different forms of sulfur (sulfuric acid potential) ) on the base of Hungarian GKM Decree No. 14/2008. (IV. 3) concerning mining waste management. A detailed geochemical study together with spatial analysis and GIS has been performed to derive a geochemically sound contamination RA of the mine waste sites. Key parameters such as heavy metal and sulphur content, in addition to the distance to the nearest surface and ground water bodies, or to sensitive receptors such as settlements and protected areas are calculated and statistically evaluated using STATGRAPHICS® in order to calibrate the RA methods. Results show that some of the waste rock materials assumed to be inert were found non/inert. Thus, regional RA needs more spatial and petrological examination with special care to rock and mineral deposit genetics.
ASCOT: a text mining-based web-service for efficient search and assisted creation of clinical trials
2012-01-01
Clinical trials are mandatory protocols describing medical research on humans and among the most valuable sources of medical practice evidence. Searching for trials relevant to some query is laborious due to the immense number of existing protocols. Apart from search, writing new trials includes composing detailed eligibility criteria, which might be time-consuming, especially for new researchers. In this paper we present ASCOT, an efficient search application customised for clinical trials. ASCOT uses text mining and data mining methods to enrich clinical trials with metadata, that in turn serve as effective tools to narrow down search. In addition, ASCOT integrates a component for recommending eligibility criteria based on a set of selected protocols. PMID:22595088
Korkontzelos, Ioannis; Mu, Tingting; Ananiadou, Sophia
2012-04-30
Clinical trials are mandatory protocols describing medical research on humans and among the most valuable sources of medical practice evidence. Searching for trials relevant to some query is laborious due to the immense number of existing protocols. Apart from search, writing new trials includes composing detailed eligibility criteria, which might be time-consuming, especially for new researchers. In this paper we present ASCOT, an efficient search application customised for clinical trials. ASCOT uses text mining and data mining methods to enrich clinical trials with metadata, that in turn serve as effective tools to narrow down search. In addition, ASCOT integrates a component for recommending eligibility criteria based on a set of selected protocols.
A Recommendation Algorithm for Automating Corollary Order Generation
Klann, Jeffrey; Schadow, Gunther; McCoy, JM
2009-01-01
Manual development and maintenance of decision support content is time-consuming and expensive. We explore recommendation algorithms, e-commerce data-mining tools that use collective order history to suggest purchases, to assist with this. In particular, previous work shows corollary order suggestions are amenable to automated data-mining techniques. Here, an item-based collaborative filtering algorithm augmented with association rule interestingness measures mined suggestions from 866,445 orders made in an inpatient hospital in 2007, generating 584 potential corollary orders. Our expert physician panel evaluated the top 92 and agreed 75.3% were clinically meaningful. Also, at least one felt 47.9% would be directly relevant in guideline development. This automated generation of a rough-cut of corollary orders confirms prior indications about automated tools in building decision support content. It is an important step toward computerized augmentation to decision support development, which could increase development efficiency and content quality while automatically capturing local standards. PMID:20351875
A recommendation algorithm for automating corollary order generation.
Klann, Jeffrey; Schadow, Gunther; McCoy, J M
2009-11-14
Manual development and maintenance of decision support content is time-consuming and expensive. We explore recommendation algorithms, e-commerce data-mining tools that use collective order history to suggest purchases, to assist with this. In particular, previous work shows corollary order suggestions are amenable to automated data-mining techniques. Here, an item-based collaborative filtering algorithm augmented with association rule interestingness measures mined suggestions from 866,445 orders made in an inpatient hospital in 2007, generating 584 potential corollary orders. Our expert physician panel evaluated the top 92 and agreed 75.3% were clinically meaningful. Also, at least one felt 47.9% would be directly relevant in guideline development. This automated generation of a rough-cut of corollary orders confirms prior indications about automated tools in building decision support content. It is an important step toward computerized augmentation to decision support development, which could increase development efficiency and content quality while automatically capturing local standards.
Perthes Disease: The Quality and Reliability of Information on the Internet.
Nassiri, Mujtaba; Bruce-Brand, Robert A; O'Neill, Francis; Chenouri, Shojaeddin; Curtin, Paul
2015-01-01
Research has shown that up to 89% of parents used the Internet to seek health information regarding their child's medical condition. Much of the information on the Internet is valuable; however, the quality of health information is variable and unregulated. The aim of this study was to evaluate the quality and content of information about Perthes disease on the Internet using recognized scoring systems, identification of quality markers, and describe a novel specific score. We searched the top 3 search engines (Google, Yahoo!, and Bing) for the following keywords: "Perthes disease." Forty-five unique Web sites were identified. The Web sites were then categorized by type and assessed using the DISCERN score, the Journal of the American Medical Association (JAMA) benchmark criteria, and a novel Perthes-specific Content score. The presence of the Health On the Net (HON) code, a reported quality assurance marker, was noted. Of the Web sites analyzed, the Majority were Governmental and Nonprofit Organizations (NPO) (37.8%), followed by commercial Web sites (22.2%). Only 6 of the Web sites were HONcode certified. The mean DISCERN score was 53.1 (SD=9.0). The Governmental and NPO Web sites had the highest overall DISCERN scores followed closely by Physician Web sites. The mean JAMA benchmark criteria score was 2.1 (SD=1.2). Nine Web sites had maximal scores and the Academic Web sites had the highest overall JAMA benchmark scores. DISCERN scores, JAMA benchmark scores, and Perthes-specific Content scores were all greater for Web sites that bore the HONcode seal. The quality of information available online regarding Perthes disease is of variable quality. Governmental and NPO Web sites predominate and also provide higher quality content. The HONcode seal is a reliable indicator of Web site quality. Physicians should recommend the HONcode seal to their patients as a reliable indicator of Web site quality or, better yet, refer patients to sites they have personally reviewed. Supplying parents with a guide to health information on the Internet will help exclude Web sites as sources of misinformation.
None Available
2018-02-06
To make the web work better for science, OSTI has developed state-of-the-art technologies and services including a deep web search capability. The deep web includes content in searchable databases available to web users but not accessible by popular search engines, such as Google. This video provides an introduction to the deep web search engine.
Web Content Management Systems: An Analysis of Forensic Investigatory Challenges.
Horsman, Graeme
2018-02-26
With an increase in the creation and maintenance of personal websites, web content management systems are now frequently utilized. Such systems offer a low cost and simple solution for those seeking to develop an online presence, and subsequently, a platform from which reported defamatory content, abuse, and copyright infringement has been witnessed. This article provides an introductory forensic analysis of the three current most popular web content management systems available, WordPress, Drupal, and Joomla! Test platforms have been created, and their site structures have been examined to provide guidance for forensic practitioners facing investigations of this type. Result's document available metadata for establishing site ownership, user interactions, and stored content following analysis of artifacts including Wordpress's wp_users, and wp_comments tables, Drupal's "watchdog" records, and Joomla!'s _users, and _content tables. Finally, investigatory limitations documenting the difficulties of investigating WCMS usage are noted, and analysis recommendations are offered. © 2018 American Academy of Forensic Sciences.
Text mining resources for the life sciences.
Przybyła, Piotr; Shardlow, Matthew; Aubin, Sophie; Bossy, Robert; Eckart de Castilho, Richard; Piperidis, Stelios; McNaught, John; Ananiadou, Sophia
2016-01-01
Text mining is a powerful technology for quickly distilling key information from vast quantities of biomedical literature. However, to harness this power the researcher must be well versed in the availability, suitability, adaptability, interoperability and comparative accuracy of current text mining resources. In this survey, we give an overview of the text mining resources that exist in the life sciences to help researchers, especially those employed in biocuration, to engage with text mining in their own work. We categorize the various resources under three sections: Content Discovery looks at where and how to find biomedical publications for text mining; Knowledge Encoding describes the formats used to represent the different levels of information associated with content that enable text mining, including those formats used to carry such information between processes; Tools and Services gives an overview of workflow management systems that can be used to rapidly configure and compare domain- and task-specific processes, via access to a wide range of pre-built tools. We also provide links to relevant repositories in each section to enable the reader to find resources relevant to their own area of interest. Throughout this work we give a special focus to resources that are interoperable-those that have the crucial ability to share information, enabling smooth integration and reusability. © The Author(s) 2016. Published by Oxford University Press.
Text mining resources for the life sciences
Shardlow, Matthew; Aubin, Sophie; Bossy, Robert; Eckart de Castilho, Richard; Piperidis, Stelios; McNaught, John; Ananiadou, Sophia
2016-01-01
Text mining is a powerful technology for quickly distilling key information from vast quantities of biomedical literature. However, to harness this power the researcher must be well versed in the availability, suitability, adaptability, interoperability and comparative accuracy of current text mining resources. In this survey, we give an overview of the text mining resources that exist in the life sciences to help researchers, especially those employed in biocuration, to engage with text mining in their own work. We categorize the various resources under three sections: Content Discovery looks at where and how to find biomedical publications for text mining; Knowledge Encoding describes the formats used to represent the different levels of information associated with content that enable text mining, including those formats used to carry such information between processes; Tools and Services gives an overview of workflow management systems that can be used to rapidly configure and compare domain- and task-specific processes, via access to a wide range of pre-built tools. We also provide links to relevant repositories in each section to enable the reader to find resources relevant to their own area of interest. Throughout this work we give a special focus to resources that are interoperable—those that have the crucial ability to share information, enabling smooth integration and reusability. PMID:27888231
CliniWeb: managing clinical information on the World Wide Web.
Hersh, W R; Brown, K E; Donohoe, L C; Campbell, E M; Horacek, A E
1996-01-01
The World Wide Web is a powerful new way to deliver on-line clinical information, but several problems limit its value to health care professionals: content is highly distributed and difficult to find, clinical information is not separated from non-clinical information, and the current Web technology is unable to support some advanced retrieval capabilities. A system called CliniWeb has been developed to address these problems. CliniWeb is an index to clinical information on the World Wide Web, providing a browsing and searching interface to clinical content at the level of the health care student or provider. Its database contains a list of clinical information resources on the Web that are indexed by terms from the Medical Subject Headings disease tree and retrieved with the assistance of SAPHIRE. Limitations of the processes used to build the database are discussed, together with directions for future research.
Characteristics of food industry web sites and "advergames" targeting children.
Culp, Jennifer; Bell, Robert A; Cassady, Diana
2010-01-01
To assess the content of food industry Web sites targeting children by describing strategies used to prolong their visits and foster brand loyalty; and to document health-promoting messages on these Web sites. A content analysis was conducted of Web sites advertised on 2 children's networks, Cartoon Network and Nickelodeon. A total of 290 Web pages and 247 unique games on 19 Internet sites were examined. Games, found on 81% of Web sites, were the most predominant promotion strategy used. All games had at least 1 brand identifier, with logos being most frequently used. On average Web sites contained 1 "healthful" message for every 45 exposures to brand identifiers. Food companies use Web sites to extend their television advertising to promote brand loyalty among children. These sites almost exclusively promoted food items high in sugar and fat. Health professionals need to monitor food industry marketing practices used in "new media." Published by Elsevier Inc.
Mining-related metals in terrestrial food webs of the upper Clark Fork River basin
DOE Office of Scientific and Technical Information (OSTI.GOV)
Pastorok, R.A.; LaTier, A.J.; Butcher, M.K.
1994-12-31
Fluvial deposits of tailings and other mining-related waste in selected riparian habitats of the Upper Clark Fork River basin (Montana) have resulted in metals enriched soils. The significance of metals exposure to selected wildlife species was evaluated by measuring tissue residues of metals (arsenic, cadmium, copper, lead, zinc) in key dietary species, including dominant grasses (tufted hair grass and redtop), willows, alfalfa, barley, invertebrates (grasshoppers, spiders, and beetles), and deer mice. Average metals concentrations in grasses, invertebrates, and deer mice collected from tailings-affected sites were elevated relative to reference to reference levels. Soil-tissue bioconcentration factors for grasses and invertebrates weremore » generally lower than expected based on the range of values in the literature, indicating the reduced bioavailability of metals from mining waste. In general, metals concentrations in willows, alfalfa, and barley were not elevated above reference levels. Using these data and plausible assumptions for other exposure parameters for white-tailed deer, red fox, and American kestrel, metals intake was estimated for soil and diet ingestion pathways. Comparisons of exposure estimates with toxicity reference values indicated that the elevated concentrations of metals in key food web species do not pose a significant risk to wildlife.« less
CrosstalkNet: A Visualization Tool for Differential Co-expression Networks and Communities.
Manem, Venkata; Adam, George Alexandru; Gruosso, Tina; Gigoux, Mathieu; Bertos, Nicholas; Park, Morag; Haibe-Kains, Benjamin
2018-04-15
Variations in physiological conditions can rewire molecular interactions between biological compartments, which can yield novel insights into gain or loss of interactions specific to perturbations of interest. Networks are a promising tool to elucidate intercellular interactions, yet exploration of these large-scale networks remains a challenge due to their high dimensionality. To retrieve and mine interactions, we developed CrosstalkNet, a user friendly, web-based network visualization tool that provides a statistical framework to infer condition-specific interactions coupled with a community detection algorithm for bipartite graphs to identify significantly dense subnetworks. As a case study, we used CrosstalkNet to mine a set of 54 and 22 gene-expression profiles from breast tumor and normal samples, respectively, with epithelial and stromal compartments extracted via laser microdissection. We show how CrosstalkNet can be used to explore large-scale co-expression networks and to obtain insights into the biological processes that govern cross-talk between different tumor compartments. Significance: This web application enables researchers to mine complex networks and to decipher novel biological processes in tumor epithelial-stroma cross-talk as well as in other studies of intercompartmental interactions. Cancer Res; 78(8); 2140-3. ©2018 AACR . ©2018 American Association for Cancer Research.
Textual and visual content-based anti-phishing: a Bayesian approach.
Zhang, Haijun; Liu, Gang; Chow, Tommy W S; Liu, Wenyin
2011-10-01
A novel framework using a Bayesian approach for content-based phishing web page detection is presented. Our model takes into account textual and visual contents to measure the similarity between the protected web page and suspicious web pages. A text classifier, an image classifier, and an algorithm fusing the results from classifiers are introduced. An outstanding feature of this paper is the exploration of a Bayesian model to estimate the matching threshold. This is required in the classifier for determining the class of the web page and identifying whether the web page is phishing or not. In the text classifier, the naive Bayes rule is used to calculate the probability that a web page is phishing. In the image classifier, the earth mover's distance is employed to measure the visual similarity, and our Bayesian model is designed to determine the threshold. In the data fusion algorithm, the Bayes theory is used to synthesize the classification results from textual and visual content. The effectiveness of our proposed approach was examined in a large-scale dataset collected from real phishing cases. Experimental results demonstrated that the text classifier and the image classifier we designed deliver promising results, the fusion algorithm outperforms either of the individual classifiers, and our model can be adapted to different phishing cases. © 2011 IEEE
X3DOM as Carrier of the Virtual Heritage
NASA Astrophysics Data System (ADS)
Jung, Y.; Behr, J.; Graf, H.
2011-09-01
Virtual Museums (VM) are a new model of communication that aims at creating a personalized, immersive, and interactive way to enhance our understanding of the world around us. The term "VM" is a short-cut that comprehends various types of digital creations. One of the carriers for the communication of the virtual heritage at future internet level as de-facto standard is browser front-ends presenting the content and assets of museums. A major driving technology for the documentation and presentation of heritage driven media is real-time 3D content, thus imposing new strategies for a web inclusion. 3D content must become a first class web media that can be created, modified, and shared in the same way as text, images, audio and video are handled on the web right now. A new integration model based on a DOM integration into the web browsers' architecture opens up new possibilities for declarative 3 D content on the web and paves the way for new application scenarios for the virtual heritage at future internet level. With special regards to the X3DOM project as enabling technology for declarative 3D in HTML, this paper describes application scenarios and analyses its technological requirements for an efficient presentation and manipulation of virtual heritage assets on the web.
A Tools-Based Approach to Teaching Data Mining Methods
ERIC Educational Resources Information Center
Jafar, Musa J.
2010-01-01
Data mining is an emerging field of study in Information Systems programs. Although the course content has been streamlined, the underlying technology is still in a state of flux. The purpose of this paper is to describe how we utilized Microsoft Excel's data mining add-ins as a front-end to Microsoft's Cloud Computing and SQL Server 2008 Business…
iHOPerator: user-scripting a personalized bioinformatics Web, starting with the iHOP website
Good, Benjamin M; Kawas, Edward A; Kuo, Byron Yu-Lin; Wilkinson, Mark D
2006-01-01
Background User-scripts are programs stored in Web browsers that can manipulate the content of websites prior to display in the browser. They provide a novel mechanism by which users can conveniently gain increased control over the content and the display of the information presented to them on the Web. As the Web is the primary medium by which scientists retrieve biological information, any improvements in the mechanisms that govern the utility or accessibility of this information may have profound effects. GreaseMonkey is a Mozilla Firefox extension that facilitates the development and deployment of user-scripts for the Firefox web-browser. We utilize this to enhance the content and the presentation of the iHOP (information Hyperlinked Over Proteins) website. Results The iHOPerator is a GreaseMonkey user-script that augments the gene-centred pages on iHOP by providing a compact, configurable visualization of the defining information for each gene and by enabling additional data, such as biochemical pathway diagrams, to be collected automatically from third party resources and displayed in the same browsing context. Conclusion This open-source script provides an extension to the iHOP website, demonstrating how user-scripts can personalize and enhance the Web browsing experience in a relevant biological setting. The novel, user-driven controls over the content and the display of Web resources made possible by user-scripts, such as the iHOPerator, herald the beginning of a transition from a resource-centric to a user-centric Web experience. We believe that this transition is a necessary step in the development of Web technology that will eventually result in profound improvements in the way life scientists interact with information. PMID:17173692
Minor and trace-elements in apiary products from a historical mining district (Les Malines, France).
Losfeld, Guillaume; Saunier, Jean-Baptiste; Grison, Claude
2014-03-01
The trace-elements (TE) contents of honey, royal-jelly and beeswax from a historical Zn-Pb mining district have been investigated to assess potential contamination. In spite of high levels of heavy metal (As, Cd, Tl, Pb) in wastes left after mining stopped, apiary products appear to be relatively free of TE contamination. For honey, the following average levels (±standard error) were observed: Zn 571±440μgkg(-1), Pb 26±20μgkg(-1), Tl 13±10μgkg(-1), Cd 7±6μgkg(-1) and As 3±4μg.kg(-1). These results bring additional data to the potential impact of brownfields left after mining on apiary products. They also bring new data to assess potential risks linked with honey consumption and discuss legal TE contents in honey and other food products from apiaries. Copyright © 2013 Elsevier Ltd. All rights reserved.
2009-11-01
relevance feedback algo- rithm. Four methods, εMap [1], MapA , P10A, and StatAP [2], were used in the track to measure the performance of Phase 2 runs...εMap and StatAP were applied to the runs us- ing the testing set of only ClueWeb09 Category-B, whereas MapA and P10A were applied to those using the...whole ClueWeb09 English set. Because our experiments were based on only ClueWeb09 Category-B, measuring our per- formance by MapA and P10A might not
Winker, M A; Flanagin, A; Chi-Lum, B; White, J; Andrews, K; Kennett, R L; DeAngelis, C D; Musacchio, R A
Access to medical information via the Internet has the potential to speed the transformation of the patient-physician relationship from that of physician authority ministering advice and treatment to that of shared decision making between patient and physician. However, barriers impeding this transformation include wide variations in quality of content on the Web, potential for commercial interests to influence online content, and uncertain preservation of personal privacy. To address these issues, the American Medical Association (AMA) has developed principles to guide development and posting of Web site content, govern acquisition and posting of online advertising and sponsorship, ensure site visitors' and patients' rights to privacy and confidentiality, and provide effective and secure means of e-commerce. While these guidelines were developed for the AMA Web sites and visitors to these sites, they also may be useful to other providers and users of medical information on the Web. These principles have been developed with the understanding that they will require frequent revision to keep pace with evolving technology and practices on the Internet. The AMA encourages review and feedback from readers, Web site visitors, policymakers, and all others interested in providing reliable quality information via the Web.
Organ donation on Web 2.0: content and audience analysis of organ donation videos on YouTube.
Tian, Yan
2010-04-01
This study examines the content of and audience response to organ donation videos on YouTube, a Web 2.0 platform, with framing theory. Positive frames were identified in both video content and audience comments. Analysis revealed a reciprocity relationship between media frames and audience frames. Videos covered content categories such as kidney, liver, organ donation registration process, and youth. Videos were favorably rated. No significant differences were found between videos produced by organizations and individuals in the United States and those produced in other countries. The findings provide insight into how new communication technologies are shaping health communication in ways that differ from traditional media. The implications of Web 2.0, characterized by user-generated content and interactivity, for health communication and health campaign practice are discussed.
The Semantic Web in Teacher Education
ERIC Educational Resources Information Center
Czerkawski, Betül Özkan
2014-01-01
The Semantic Web enables increased collaboration among computers and people by organizing unstructured data on the World Wide Web. Rather than a separate body, the Semantic Web is a functional extension of the current Web made possible by defining relationships among websites and other online content. When explicitly defined, these relationships…
DOE Office of Scientific and Technical Information (OSTI.GOV)
None Available
To make the web work better for science, OSTI has developed state-of-the-art technologies and services including a deep web search capability. The deep web includes content in searchable databases available to web users but not accessible by popular search engines, such as Google. This video provides an introduction to the deep web search engine.
Corporate Web Sites in Traditional Print Advertisements.
ERIC Educational Resources Information Center
Pardun, Carol J.; Lamb, Larry
1999-01-01
Describes the Web presence in print advertisements to determine how marketers are creating bridges between traditional advertising and the Internet. Content analysis showed Web addresses in print ads; categories of advertisers most likely to link print ads with Web sites; and whether the Web site attempts to develop a database of potential…
Kolker, A.; Panov, B.S.; Panov, Y.B.; Landa, E.R.; Conko, K.M.; Korchemagin, V.A.; Shendrik, T.; McCord, J.D.
2009-01-01
Mercury-rich coals in the Donets Basin (Donbas region) of Ukraine were sampled in active underground mines to assess the levels of potentially harmful elements and the potential for dispersion of metals through use of this coal. For 29 samples representing c11 to m3 Carboniferous coals, mercury contents range from 0.02 to 3.5 ppm (whole-coal dry basis). Mercury is well correlated with pyritic sulfur (0.01 to 3.2 wt.%), with an r2 of 0.614 (one outlier excluded). Sulfides in these samples show enrichment of minor constituents in late-stage pyrite formed as a result of interaction of coal with hydrothermal fluids. Mine water sampled at depth and at surface collection points does not show enrichment of trace metals at harmful levels, indicating pyrite stability at subsurface conditions. Four samples of coal exposed in the defunct open-cast Nikitovka mercury mines in Gorlovka have extreme mercury contents of 12.8 to 25.5 ppm. This coal was formerly produced as a byproduct of extracting sandstone-hosted cinnabar ore. Access to these workings is unrestricted and small amounts of extreme mercury-rich coal are collected for domestic use, posing a limited human health hazard. More widespread hazards are posed by the abandoned Nikitovka mercury processing plant, the extensive mercury mine tailings, and mercury enrichment of soils extending into residential areas of Gorlovka.
Study of Natural Radioactivity in Coal Samples of Baganuur Coal Mine, Mongolia
NASA Astrophysics Data System (ADS)
Altangerel, M.; Norov, N.; Altangerel, D.
2009-03-01
Coal and soil samples from Baganuur Coal Mine (BCM) of Mongolia have been investigated. The activities of 226Ra, 232Th and 40K have been measured by gamma-ray spectrometry using shielded HPGe detector. Contents of natural radionuclide elements (U, Th and K) have been determined. Also the activities and contents of radionuclide of ashes were determined which generated in Thermal Power Plant ♯3 of Ulaanbaatar from coal supplied from BCM.
Towards Web 3.0: taxonomies and ontologies for medical education -- a systematic review.
Blaum, Wolf E; Jarczweski, Anne; Balzer, Felix; Stötzner, Philip; Ahlers, Olaf
2013-01-01
Both for curricular development and mapping, as well as for orientation within the mounting supply of learning resources in medical education, the Semantic Web ("Web 3.0") poses a low-threshold, effective tool that enables identification of content related items across system boundaries. Replacement of the currently required manual with an automatically generated link, which is based on content and semantics, requires the use of a suitably structured vocabulary for a machine-readable description of object content. Aim of this study is to compile the existing taxonomies and ontologies used for the annotation of medical content and learning resources, to compare those using selected criteria, and to verify their suitability in the context described above. Based on a systematic literature search, existing taxonomies and ontologies for the description of medical learning resources were identified. Through web searches and/or direct contact with the respective editors, each of the structured vocabularies thus identified were examined in regards to topic, structure, language, scope, maintenance, and technology of the taxonomy/ontology. In addition, suitability for use in the Semantic Web was verified. Among 20 identified publications, 14 structured vocabularies were identified, which differed rather strongly in regards to language, scope, currency, and maintenance. None of the identified vocabularies fulfilled the necessary criteria for content description of medical curricula and learning resources in the German-speaking world. While moving towards Web 3.0, a significant problem lies in the selection and use of an appropriate German vocabulary for the machine-readable description of object content. Possible solutions include development, translation and/or combination of existing vocabularies, possibly including partial translations of English vocabularies.
Optimizing Crawler4j using MapReduce Programming Model
NASA Astrophysics Data System (ADS)
Siddesh, G. M.; Suresh, Kavya; Madhuri, K. Y.; Nijagal, Madhushree; Rakshitha, B. R.; Srinivasa, K. G.
2017-06-01
World wide web is a decentralized system that consists of a repository of information on the basis of web pages. These web pages act as a source of information or data in the present analytics world. Web crawlers are used for extracting useful information from web pages for different purposes. Firstly, it is used in web search engines where the web pages are indexed to form a corpus of information and allows the users to query on the web pages. Secondly, it is used for web archiving where the web pages are stored for later analysis phases. Thirdly, it can be used for web mining where the web pages are monitored for copyright purposes. The amount of information processed by the web crawler needs to be improved by using the capabilities of modern parallel processing technologies. In order to solve the problem of parallelism and the throughput of crawling this work proposes to optimize the Crawler4j using the Hadoop MapReduce programming model by parallelizing the processing of large input data. Crawler4j is a web crawler that retrieves useful information about the pages that it visits. The crawler Crawler4j coupled with data and computational parallelism of Hadoop MapReduce programming model improves the throughput and accuracy of web crawling. The experimental results demonstrate that the proposed solution achieves significant improvements with respect to performance and throughput. Hence the proposed approach intends to carve out a new methodology towards optimizing web crawling by achieving significant performance gain.
Cele, Emmanuel Nkosinathi; Maboeta, Mark
2016-11-01
The achievement of environmentally sound and economically feasible disposal strategies for biosolids is a major issue in the wastewater treatment industry around the world, including Swaziland. Currently, an iron ore mine site, which is located within a wildlife sanctuary, is being considered as a suitable place where controlled disposal of biosolids may be practiced. Therefore, this study was conducted to investigate the effects of urban biosolids on iron mine soils with regard to plant metal content and ecotoxicological effects on earthworms. This was done through chemical analysis of plants grown in biosolid-amended mine soil. Earthworm behaviour, reproduction and bioaccumulation tests were also conducted on biosolid-amended mine soil. According to the results obtained, the use of biosolids led to creation of soil conditions that were generally favourable to earthworms. However, plants were found to have accumulated Zn up to 346 mg kg -1 (in shoots) and 462 mg kg -1 (in roots). This was more than double the normal Zn content of plants. It was concluded that while biosolids can be beneficial to mine soils and earthworms, they can also lead to elevated metal content in plant tissues, which might be a concern to plant-dependant wildlife species. Nonetheless, it was not possible to satisfactorily estimate risks to forage quality since animal feeding tests with hyperaccumulator plants have not been reported. Quite possibly, there may be no cause for alarm since the uptake of metals from soil is greater in plants grown in pots in the greenhouse than from the same soil in the field since pot studies fail to mimic field conditions where the soil is heterogeneous and where the root system possesses a complex topology. It was thought that further field trials might assist in arriving at more satisfactory conclusions.
NASA Astrophysics Data System (ADS)
Jamison, Mark A.; Hauge, Janice A.
It is commonplace for sellers of goods and services to enhance the value of their products by paying extra for premium delivery service. For example, package delivery services such as Federal Express and the US Postal Service offer shippers a variety of delivery speeds and insurance programs. Web content providers such as Yahoo! and MSN Live Earth can purchase web-enhancing services from companies such as Akamai to speed the delivery of their web content to customers.1
76 FR 35968 - Maintenance of Incombustible Content of Rock Dust in Underground Coal Mines
Federal Register 2010, 2011, 2012, 2013, 2014
2011-06-21
...) Pursuant to the Regulatory Flexibility Act (RFA) of 1980, as amended by SBREFA, MSHA has analyzed the.... 60 Pages. [Missing Appendices I through L] Department of Labor, Mine Safety and Health Administration...
Make Mine a Metasearcher, Please!
ERIC Educational Resources Information Center
Repman, Judi; Carlson, Randal D.
2000-01-01
Describes metasearch tools and explains their value in helping library media centers improve students' Web searches. Discusses Boolean queries and the emphasis on speed at the expense of comprehensiveness; and compares four metasearch tools, including the number of search engines consulted, user control, and databases included. (LRW)
DataFerrett is a data extraction software and a data mining tool that accesses data stored in TheDataWeb through the Internet. It can be installed as an application on your desktop or use a java applet with an Internet browser. Census Bureau and Bureau of Labor Statistics release...
The Umbra Simulation and Integration Framework Applied to Emergency Response Training
NASA Technical Reports Server (NTRS)
Hamilton, Paul Lawrence; Britain, Robert
2010-01-01
The Mine Emergency Response Interactive Training Simulation (MERITS) is intended to prepare personnel to manage an emergency in an underground coal mine. The creation of an effective training environment required realistic emergent behavior in response to simulation events and trainee interventions, exploratory modification of miner behavior rules, realistic physics, and incorporation of legacy code. It also required the ability to add rich media to the simulation without conflicting with normal desktop security settings. Our Umbra Simulation and Integration Framework facilitated agent-based modeling of miners and rescuers and made it possible to work with subject matter experts to quickly adjust behavior through script editing, rather than through lengthy programming and recompilation. Integration of Umbra code with the WebKit browser engine allowed the use of JavaScript-enabled local web pages for media support. This project greatly extended the capabilities of Umbra in support of training simulations and has implications for simulations that combine human behavior, physics, and rich media.
A web server for mining Comparative Genomic Hybridization (CGH) data
NASA Astrophysics Data System (ADS)
Liu, Jun; Ranka, Sanjay; Kahveci, Tamer
2007-11-01
Advances in cytogenetics and molecular biology has established that chromosomal alterations are critical in the pathogenesis of human cancer. Recurrent chromosomal alterations provide cytological and molecular markers for the diagnosis and prognosis of disease. They also facilitate the identification of genes that are important in carcinogenesis, which in the future may help in the development of targeted therapy. A large amount of publicly available cancer genetic data is now available and it is growing. There is a need for public domain tools that allow users to analyze their data and visualize the results. This chapter describes a web based software tool that will allow researchers to analyze and visualize Comparative Genomic Hybridization (CGH) datasets. It employs novel data mining methodologies for clustering and classification of CGH datasets as well as algorithms for identifying important markers (small set of genomic intervals with aberrations) that are potentially cancer signatures. The developed software will help in understanding the relationships between genomic aberrations and cancer types.
Sun, Jiayang; Bogie, Kath M; Teagno, Joe; Sun, Yu-Hsiang (Sam); Carter, Rebecca R; Cui, Licong; Zhang, Guo-Qiang
2014-01-01
Ovarian cancer (OvCa) is the most lethal gynecologic disease in the United States, with an overall 5-year survival rate of 44.5%, about half of the 89.2% for all breast cancer patients. To identify factors that possibly contribute to the long-term survivorship of women with OvCa, we conducted a comprehensive online Ovarian Cancer Survivorship Survey from 2009 to 2013. This paper presents the design and implementation of our survey, introduces its resulting data source, the OVA-CRADLE™ (Clinical Research Analytics and Data Lifecycle Environment), and illustrates a sample application of the survey and data by an analysis of prediagnosis symptoms, using text mining and statistics. The OVA-CRADLE™ is an application of our patented Physio-MIMI technology, facilitating Web-based access, online query and exploration of data. The prediagnostic symptoms and association of early-stage OvCa diagnosis with endometriosis provide potentially important indicators for future studies in this field. PMID:25861211
NCBI GEO: mining millions of expression profiles--database and tools.
Barrett, Tanya; Suzek, Tugba O; Troup, Dennis B; Wilhite, Stephen E; Ngau, Wing-Chi; Ledoux, Pierre; Rudnev, Dmitry; Lash, Alex E; Fujibuchi, Wataru; Edgar, Ron
2005-01-01
The Gene Expression Omnibus (GEO) at the National Center for Biotechnology Information (NCBI) is the largest fully public repository for high-throughput molecular abundance data, primarily gene expression data. The database has a flexible and open design that allows the submission, storage and retrieval of many data types. These data include microarray-based experiments measuring the abundance of mRNA, genomic DNA and protein molecules, as well as non-array-based technologies such as serial analysis of gene expression (SAGE) and mass spectrometry proteomic technology. GEO currently holds over 30,000 submissions representing approximately half a billion individual molecular abundance measurements, for over 100 organisms. Here, we describe recent database developments that facilitate effective mining and visualization of these data. Features are provided to examine data from both experiment- and gene-centric perspectives using user-friendly Web-based interfaces accessible to those without computational or microarray-related analytical expertise. The GEO database is publicly accessible through the World Wide Web at http://www.ncbi.nlm.nih.gov/geo.
Wang, Jia-Hong; Zhao, Ling-Feng; Lin, Pei; Su, Xiao-Rong; Chen, Shi-Jun; Huang, Li-Qiang; Wang, Hua-Feng; Zhang, Hai; Hu, Zhen-Fu; Yao, Kai-Tai; Huang, Zhong-Xi
2014-09-01
Identifying biological functions and molecular networks in a gene list and how the genes may relate to various topics is of considerable value to biomedical researchers. Here, we present a web-based text-mining server, GenCLiP 2.0, which can analyze human genes with enriched keywords and molecular interactions. Compared with other similar tools, GenCLiP 2.0 offers two unique features: (i) analysis of gene functions with free terms (i.e. any terms in the literature) generated by literature mining or provided by the user and (ii) accurate identification and integration of comprehensive molecular interactions from Medline abstracts, to construct molecular networks and subnetworks related to the free terms. http://ci.smu.edu.cn. Supplementary data are available at Bioinformatics online. © The Author 2014. Published by Oxford University Press. All rights reserved. For Permissions, please e-mail: journals.permissions@oup.com.
String Mining in Bioinformatics
NASA Astrophysics Data System (ADS)
Abouelhoda, Mohamed; Ghanem, Moustafa
Sequence analysis is a major area in bioinformatics encompassing the methods and techniques for studying the biological sequences, DNA, RNA, and proteins, on the linear structure level. The focus of this area is generally on the identification of intra- and inter-molecular similarities. Identifying intra-molecular similarities boils down to detecting repeated segments within a given sequence, while identifying inter-molecular similarities amounts to spotting common segments among two or multiple sequences. From a data mining point of view, sequence analysis is nothing but string- or pattern mining specific to biological strings. For a long time, this point of view, however, has not been explicitly embraced neither in the data mining nor in the sequence analysis text books, which may be attributed to the co-evolution of the two apparently independent fields. In other words, although the word "data-mining" is almost missing in the sequence analysis literature, its basic concepts have been implicitly applied. Interestingly, recent research in biological sequence analysis introduced efficient solutions to many problems in data mining, such as querying and analyzing time series [49,53], extracting information from web pages [20], fighting spam mails [50], detecting plagiarism [22], and spotting duplications in software systems [14].
How to Serve Content to PDA Users on-the-Go
ERIC Educational Resources Information Center
Cuddy, Colleen
2006-01-01
Fans of mobile devices are everywhere, and they are using their PDAs, smart phones, and mobile phones to access Web-based content. Chances are that they are trying to access your library's Web site or find library-based content for their devices. In this article, the author presents some tips on how to serve those who wants to grab some fast info…
ERIC Educational Resources Information Center
Katirci, Hakan
2016-01-01
Considering various themes, this study aims to examine the content of web sites of universities that provide sports management education in higher education level in Turkey and in England. Within this framework, the websites of the higher education institutions that provide sports management education are analyzed by using the content analysis…
Shaath, M Kareem; Yeranosian, Michael G; Ippolito, Joseph A; Adams, Mark R; Sirkin, Michael S; Reilly, Mark C
2018-05-02
Orthopaedic trauma fellowship applicants use online-based resources when researching information on potential U.S. fellowship programs. The 2 primary sources for identifying programs are the Orthopaedic Trauma Association (OTA) database and the San Francisco Match (SF Match) database. Previous studies in other orthopaedic subspecialty areas have demonstrated considerable discrepancies among fellowship programs. The purpose of this study was to analyze content and availability of information on orthopaedic trauma surgery fellowship web sites. The online databases of the OTA and SF Match were reviewed to determine the availability of embedded program links or external links for the included programs. Thereafter, a Google search was performed for each program individually by typing the program's name, followed by the term "orthopaedic trauma fellowship." All identified fellowship web sites were analyzed for accessibility and content. Web sites were evaluated for comprehensiveness in mentioning key components of the orthopaedic trauma surgery curriculum. By consensus, we refined the final list of variables utilizing the methodology of previous studies on the topic. We identified 54 OTA-accredited fellowship programs, offering 87 positions. The majority (94%) of programs had web sites accessible through a Google search. Of the 51 web sites found, all (100%) described their program. Most commonly, hospital affiliation (88%), operative experiences (76%), and rotation overview (65%) were listed, and, least commonly, interview dates (6%), selection criteria (16%), on-call requirements (20%), and fellow evaluation criteria (20%) were listed. Programs with ≥2 fellows provided more information with regard to education content (p = 0.0001) and recruitment content (p = 0.013). Programs with Accreditation Council for Graduate Medical Education (ACGME) accreditation status also provided greater information with regard to education content (odds ratio, 4.0; p = 0.0001). Otherwise, no differences were seen by region, residency affiliation, medical school affiliation, or hospital affiliation. The SF Match and OTA databases provide few direct links to fellowship web sites. Individual program web sites do not effectively and completely convey information about the programs. The Internet is an underused resource for fellow recruitment. The lack of information on these sites allows for future opportunity to optimize this resource.
Hoelzer, Simon; Schweiger, Ralf K; Rieger, Joerg; Meyer, Michael
2006-01-01
The organizational structures of web contents and electronic information resources must adapt to the demands of a growing volume of information and user requirements. Otherwise the information society will be threatened by disinformation. The biomedical sciences are especially vulnerable in this regard, since they are strongly oriented toward text-based knowledge sources. Here sustainable improvement can only be achieved by using a comprehensive, integrated approach that not only includes data management but also specifically incorporates the editorial processes, including structuring information sources and publication. The technical resources needed to effectively master these tasks are already available in the form of the data standards and tools of the Semantic Web. They include Rich Site Summaries (RSS), which have become an established means of distributing and syndicating conventional news messages and blogs. They can also provide access to the contents of the previously mentioned information sources, which are conventionally classified as 'deep web' content.
Harvesting Intelligence in Multimedia Social Tagging Systems
NASA Astrophysics Data System (ADS)
Giannakidou, Eirini; Kaklidou, Foteini; Chatzilari, Elisavet; Kompatsiaris, Ioannis; Vakali, Athena
As more people adopt tagging practices, social tagging systems tend to form rich knowledge repositories that enable the extraction of patterns reflecting the way content semantics is perceived by the web users. This is of particular importance, especially in the case of multimedia content, since the availability of such content in the web is very high and its efficient retrieval using textual annotations or content-based automatically extracted metadata still remains a challenge. It is argued that complementing multimedia analysis techniques with knowledge drawn from web social annotations may facilitate multimedia content management. This chapter focuses on analyzing tagging patterns and combining them with content feature extraction methods, generating, thus, intelligence from multimedia social tagging systems. Emphasis is placed on using all available "tracks" of knowledge, that is tag co-occurrence together with semantic relations among tags and low-level features of the content. Towards this direction, a survey on the theoretical background and the adopted practices for analysis of multimedia social content are presented. A case study from Flickr illustrates the efficiency of the proposed approach.
Pistelli, Laura; D'Angiolillo, Francesca; Morelli, Elisabetta; Basso, Barbara; Rosellini, Irene; Posarelli, Mauro; Barbafieri, Meri
2017-03-01
The release of large amounts of toxic metals in the neighboring sites of abandoned mine areas represents an important environmental risk for the ecosystem, because it adversely affects soil, water, and plant growth. The aim of the present study was to investigate the metal(loid) (As, Cr, Cu, Ni, Pb, and Zn) contents of native Mediterranean plants grown on the ex-mining area of Elba island (Italy), with the prospective of its recovery by further phytoremediation technology. Soil samples were collected and characterized for metal(loid) content in total and potentially available (EDTA-extractable) fractions. Arsenic was particularly high, being 338 and 2.1 mg kg -1 as total and available fractions, respectively. Predominant native species, namely Dittrichia viscosa L. Greuter, Cistus salviifolius L., Lavandula stoechas L., and Bituminaria bituminosa L., were analyzed for metal content in the different plant organs. D. viscosa exhibited the highest metal(loid) content in the leaves and the singular behavior of translocating arsenic to the leaves (transfer factor about 2.06 and mean bioconcentration factor about 12.48). To assess the healthy status of D. viscosa plants, the leaves were investigated further. The activities of the main antioxidant enzymes and the levels of secondary metabolites linked to oxidative stress in plants from the ex-mining area were not significantly different from those of control plants, except for a lower content of carotenoids, indicating that native plants were adapted to grow in these polluted soils. These results indicate that D. viscosa can be suitable for the revegetation of highly metal-contaminated areas.
NEIMiner: nanomaterial environmental impact data miner.
Tang, Kaizhi; Liu, Xiong; Harper, Stacey L; Steevens, Jeffery A; Xu, Roger
2013-01-01
As more engineered nanomaterials (eNM) are developed for a wide range of applications, it is crucial to minimize any unintended environmental impacts resulting from the application of eNM. To realize this vision, industry and policymakers must base risk management decisions on sound scientific information about the environmental fate of eNM, their availability to receptor organisms (eg, uptake), and any resultant biological effects (eg, toxicity). To address this critical need, we developed a model-driven, data mining system called NEIMiner, to study nanomaterial environmental impact (NEI). NEIMiner consists of four components: NEI modeling framework, data integration, data management and access, and model building. The NEI modeling framework defines the scope of NEI modeling and the strategy of integrating NEI models to form a layered, comprehensive predictability. The data integration layer brings together heterogeneous data sources related to NEI via automatic web services and web scraping technologies. The data management and access layer reuses and extends a popular content management system (CMS), Drupal, and consists of modules that model the complex data structure for NEI-related bibliography and characterization data. The model building layer provides an advanced analysis capability for NEI data. Together, these components provide significant value to the process of aggregating and analyzing large-scale distributed NEI data. A prototype of the NEIMiner system is available at http://neiminer.i-a-i.com/.
NASA Astrophysics Data System (ADS)
Esperschütz, J.; Zimmermann, C.; Dümig, A.; Welzl, G.; Buegger, F.; Elmer, M.; Munch, J. C.; Schloter, M.
2013-07-01
In initial ecosystems, concentrations of all macro- and micronutrients can be considered as extremely low. Plant litter therefore strongly influences the development of a degrader's food web and is an important source for C and N input into soil in such ecosystems. In the present study, a 13C litter decomposition field experiment was performed for 30 weeks in initial soils from a post-mining area near the city of Cottbus (Germany). Two of this region's dominant but contrasting pioneering plant species (Lotus corniculatus L. and Calamagrostis epigejos L.) were chosen to investigate the effects of litter quality on the litter decomposing microbial food web in initially nutrient-poor substrates. The results clearly indicate the importance of litter quality, as indicated by its N content, its bioavailability for the degradation process and the development of microbial communities in the detritusphere and soil. The degradation of the L. corniculatus litter, which had a low C / N ratio, was fast and showed pronounced changes in the microbial community structure 1-4 weeks after litter addition. The degradation of the C. epigejos litter material was slow and microbial community changes mainly occurred between 4 and 30 weeks after litter addition to the soil. However, for both litter materials a clear indication of the importance of fungi for the degradation process was observed both in terms of fungal abundance and activity (13C incorporation activity)
NEIMiner: nanomaterial environmental impact data miner
Tang, Kaizhi; Liu, Xiong; Harper, Stacey L; Steevens, Jeffery A; Xu, Roger
2013-01-01
As more engineered nanomaterials (eNM) are developed for a wide range of applications, it is crucial to minimize any unintended environmental impacts resulting from the application of eNM. To realize this vision, industry and policymakers must base risk management decisions on sound scientific information about the environmental fate of eNM, their availability to receptor organisms (eg, uptake), and any resultant biological effects (eg, toxicity). To address this critical need, we developed a model-driven, data mining system called NEIMiner, to study nanomaterial environmental impact (NEI). NEIMiner consists of four components: NEI modeling framework, data integration, data management and access, and model building. The NEI modeling framework defines the scope of NEI modeling and the strategy of integrating NEI models to form a layered, comprehensive predictability. The data integration layer brings together heterogeneous data sources related to NEI via automatic web services and web scraping technologies. The data management and access layer reuses and extends a popular content management system (CMS), Drupal, and consists of modules that model the complex data structure for NEI-related bibliography and characterization data. The model building layer provides an advanced analysis capability for NEI data. Together, these components provide significant value to the process of aggregating and analyzing large-scale distributed NEI data. A prototype of the NEIMiner system is available at http://neiminer.i-a-i.com/. PMID:24098076
Staccini, P; Douali, N
2014-08-15
To provide a review of the current excellent research published in the field of Consumer Health Informatics. We searched MEDLINE® and WEB OF SCIENCE® databases for papers published in 2013 in relation with Consumer Health Informatics. The authors identified 16 candidate best papers, which were then reviewed by four reviewers. Five out of the 16 candidate papers were selected as best papers. One paper presents the key features of a system to automate the collection of web-based social media content for subsequent semantic annotation. This paper emphasizes the importance of mining social media to collect novel data from which new findings in drug abuse research were uncovered. The second paper presents a practical method to predict how a community structure would impact the spreading of information within the community. The third paper presents a method for improving the quality of online health communities. The fourth presents a new social network to allow the monitoring of the evolution of individuals' health status and diagnostic deficiencies, difficulties or barriers in rehabilitation. The last paper reports on teenage patients' perception on privacy and social media. Selected papers not only show the value of using social media in the medical field but how to use these media to detect emergent diseases or risks, inform patients, promote disease prevention, and follow patients' opinion on healthcare resources.
Social Media and Patient Health Outcomes
2014-01-01
Summary Objectives To provide a review of the current excellent research published in the field of Consumer Health Informatics. Method We searched MEDLINE® and WEB OF SCIENCE® databases for papers published in 2013 in relation with Consumer Health Informatics. The authors identified 16 candidate best papers, which were then reviewed by four reviewers. Results Five out of the 16 candidate papers were selected as best papers. One paper presents the key features of a system to automate the collection of web-based social media content for subsequent semantic annotation. This paper emphasizes the importance of mining social media to collect novel data from which new findings in drug abuse research were uncovered. The second paper presents a practical method to predict how a community structure would impact the spreading of information within the community. The third paper presents a method for improving the quality of online health communities. The fourth presents a new social network to allow the monitoring of the evolution of individuals’ health status and diagnostic deficiencies, difficulties or barriers in rehabilitation. The last paper reports on teenage patients’ perception on privacy and social media. Conclusion Selected papers not only show the value of using social media in the medical field but how to use these media to detect emergent diseases or risks, inform patients, promote disease prevention, and follow patients’ opinion on healthcare resources. PMID:25123742
Web Site Design Benchmarking within Industry Groups.
ERIC Educational Resources Information Center
Kim, Sung-Eon; Shaw, Thomas; Schneider, Helmut
2003-01-01
Discussion of electronic commerce focuses on Web site evaluation criteria and applies them to different industry groups in Korea. Defines six categories of Web site evaluation criteria: business function, corporate credibility, contents reliability, Web site attractiveness, systematic structure, and navigation; and discusses differences between…
Assessment of respirable dust and its free silica contents in different Indian coalmines.
Mukherjee, Ashit K; Bhattacharya, Sanat K; Saiyed, Habibullah N
2005-04-01
Assessment of respirable dust, personal exposures of miners and free silica contents in dust were undertaken to find out the associated risk of coal workers' pneumoconiosis in 9 coal mines of Eastern India during 1988-91. Mine Research Establishment (MRE), 113A Gravimetric Dust Sampler (GDS) and personal samplers (AFC 123), Cassella, London, approved by Director General of Mines Safety (DGMS) were used respectively for monitoring of mine air dust and personal exposures of miners. Fourier Transform Infra-red (FTIR) Spectroscopy determined free silica in respirable dusts. Thermal Conditions like Wet Bulb Globe Temperature (WBGT) index, humidity and wind velocity were also recorded during monitoring. The dust levels in the face return air of both, Board & Pillar (B&P) and Long Wall (LW) mining were found above the permissible level recommended by DGMS, Govt. of India. The drilling, blasting and loading are the major dusty operations in B&P method. Exposures of driller and loader were varied between, 0.81-9.48 mg/m3 and 0.05-9.84 mg/m3 respectively in B&P mining, whereas exposures of DOSCO loader, Shearer operator and Power Support Face Worker were varied between 2.65-9.11 mg/m3, 0.22-10.00 mg/m3 and 0.12-9.32 mg/m3 respectively in LW mining. In open cast mining, compressor and driller operators are the major exposed groups. The percentage silica in respirable dusts found below 5% in all most all the workers except among query loaders and drillers of open cast mines.