Tags Extarction from Spatial Documents in Search Engines
NASA Astrophysics Data System (ADS)
Borhaninejad, S.; Hakimpour, F.; Hamzei, E.
2015-12-01
Nowadays the selective access to information on the Web is provided by search engines, but in the cases which the data includes spatial information the search task becomes more complex and search engines require special capabilities. The purpose of this study is to extract the information which lies in spatial documents. To that end, we implement and evaluate information extraction from GML documents and a retrieval method in an integrated approach. Our proposed system consists of three components: crawler, database and user interface. In crawler component, GML documents are discovered and their text is parsed for information extraction; storage. The database component is responsible for indexing of information which is collected by crawlers. Finally the user interface component provides the interaction between system and user. We have implemented this system as a pilot system on an Application Server as a simulation of Web. Our system as a spatial search engine provided searching capability throughout the GML documents and thus an important step to improve the efficiency of search engines has been taken.
Electronic Document Management Using Inverted Files System
NASA Astrophysics Data System (ADS)
Suhartono, Derwin; Setiawan, Erwin; Irwanto, Djon
2014-03-01
The amount of documents increases so fast. Those documents exist not only in a paper based but also in an electronic based. It can be seen from the data sample taken by the SpringerLink publisher in 2010, which showed an increase in the number of digital document collections from 2003 to mid of 2010. Then, how to manage them well becomes an important need. This paper describes a new method in managing documents called as inverted files system. Related with the electronic based document, the inverted files system will closely used in term of its usage to document so that it can be searched over the Internet using the Search Engine. It can improve document search mechanism and document save mechanism.
A Survey in Indexing and Searching XML Documents.
ERIC Educational Resources Information Center
Luk, Robert W. P.; Leong, H. V.; Dillon, Tharam S.; Chan, Alvin T. S.; Croft, W. Bruce; Allan, James
2002-01-01
Discussion of XML focuses on indexing techniques for XML documents, grouping them into flat-file, semistructured, and structured indexing paradigms. Highlights include searching techniques, including full text search and multistage search; search result presentations; database and information retrieval system integration; XML query languages; and…
Agent-based method for distributed clustering of textual information
Potok, Thomas E [Oak Ridge, TN; Reed, Joel W [Knoxville, TN; Elmore, Mark T [Oak Ridge, TN; Treadwell, Jim N [Louisville, TN
2010-09-28
A computer method and system for storing, retrieving and displaying information has a multiplexing agent (20) that calculates a new document vector (25) for a new document (21) to be added to the system and transmits the new document vector (25) to master cluster agents (22) and cluster agents (23) for evaluation. These agents (22, 23) perform the evaluation and return values upstream to the multiplexing agent (20) based on the similarity of the document to documents stored under their control. The multiplexing agent (20) then sends the document (21) and the document vector (25) to the master cluster agent (22), which then forwards it to a cluster agent (23) or creates a new cluster agent (23) to manage the document (21). The system also searches for stored documents according to a search query having at least one term and identifying the documents found in the search, and displays the documents in a clustering display (80) of similarity so as to indicate similarity of the documents to each other.
NASA Astrophysics Data System (ADS)
Stalcup, Bruce W.; Dennis, Phillip W.; Dydyk, Robert B.
1999-10-01
Litton PRC and Litton Data Systems Division are developing a system, the Imaged Document Optical Correlation and Conversion System (IDOCCS), to provide a total solution to the problem of managing and retrieving textual and graphic information from imaged document archives. At the heart of IDOCCS, optical correlation technology provides the search and retrieval of information from imaged documents. IDOCCS can be used to rapidly search for key words or phrases within the imaged document archives. In addition, IDOCCS can automatically compare an input document with the archived database to determine if it is a duplicate, thereby reducing the overall resources required to maintain and access the document database. Embedded graphics on imaged pages can also be exploited; e.g., imaged documents containing an agency's seal or logo can be singled out. In this paper, we present a description of IDOCCS as well as preliminary performance results and theoretical projections.
A Synchronous Search for Documents
An algorithm is described of a synchronous search in a complex system of selective retrieval of documents, with an allowance for exclusion of...stored on a magnetic tape. The number of topics served by the synchronous search goes into thousands; a search within 500-600 topics is performed without additional access to the tape.
What Can Pictures Tell Us About Web Pages? Improving Document Search Using Images.
Rodriguez-Vaamonde, Sergio; Torresani, Lorenzo; Fitzgibbon, Andrew W
2015-06-01
Traditional Web search engines do not use the images in the HTML pages to find relevant documents for a given query. Instead, they typically operate by computing a measure of agreement between the keywords provided by the user and only the text portion of each page. In this paper we study whether the content of the pictures appearing in a Web page can be used to enrich the semantic description of an HTML document and consequently boost the performance of a keyword-based search engine. We present a Web-scalable system that exploits a pure text-based search engine to find an initial set of candidate documents for a given query. Then, the candidate set is reranked using visual information extracted from the images contained in the pages. The resulting system retains the computational efficiency of traditional text-based search engines with only a small additional storage cost needed to encode the visual information. We test our approach on one of the TREC Million Query Track benchmarks where we show that the exploitation of visual content yields improvement in accuracies for two distinct text-based search engines, including the system with the best reported performance on this benchmark. We further validate our approach by collecting document relevance judgements on our search results using Amazon Mechanical Turk. The results of this experiment confirm the improvement in accuracy produced by our image-based reranker over a pure text-based system.
System for gathering and summarizing internet information
Potok, Thomas E.; Elmore, Mark Thomas; Reed, Joel Wesley; Treadwell, Jim N.; Samatova, Nagiza Faridovna
2006-07-04
A computer method of gathering and summarizing large amounts of information comprises collecting information from a plurality of information sources (14, 51) according to respective maps (52) of the information sources (14), converting the collected information from a storage format to XML-language documents (26, 53) and storing the XML-language documents in a storage medium, searching for documents (55) according to a search query (13) having at least one term and identifying the documents (26) found in the search, and displaying the documents as nodes (33) of a tree structure (32) having links (34) and nodes (33) so as to indicate similarity of the documents to each other.
A suffix arrays based approach to semantic search in P2P systems
NASA Astrophysics Data System (ADS)
Shi, Qingwei; Zhao, Zheng; Bao, Hu
2007-09-01
Building a semantic search system on top of peer-to-peer (P2P) networks is becoming an attractive and promising alternative scheme for the reason of scalability, Data freshness and search cost. In this paper, we present a Suffix Arrays based algorithm for Semantic Search (SASS) in P2P systems, which generates a distributed Semantic Overlay Network (SONs) construction for full-text search in P2P networks. For each node through the P2P network, SASS distributes document indices based on a set of suffix arrays, by which clusters are created depending on words or phrases shared between documents, therefore, the search cost for a given query is decreased by only scanning semantically related documents. In contrast to recently announced SONs scheme designed by using metadata or predefined-class, SASS is an unsupervised approach for decentralized generation of SONs. SASS is also an incremental, linear time algorithm, which efficiently handle the problem of nodes update in P2P networks. Our simulation results demonstrate that SASS yields high search efficiency in dynamic environments.
BOSS: context-enhanced search for biomedical objects
2012-01-01
Background There exist many academic search solutions and most of them can be put on either ends of spectrum: general-purpose search and domain-specific "deep" search systems. The general-purpose search systems, such as PubMed, offer flexible query interface, but churn out a list of matching documents that users have to go through the results in order to find the answers to their queries. On the other hand, the "deep" search systems, such as PPI Finder and iHOP, return the precompiled results in a structured way. Their results, however, are often found only within some predefined contexts. In order to alleviate these problems, we introduce a new search engine, BOSS, Biomedical Object Search System. Methods Unlike the conventional search systems, BOSS indexes segments, rather than documents. A segment refers to a Maximal Coherent Semantic Unit (MCSU) such as phrase, clause or sentence that is semantically coherent in the given context (e.g., biomedical objects or their relations). For a user query, BOSS finds all matching segments, identifies the objects appearing in those segments, and aggregates the segments for each object. Finally, it returns the ranked list of the objects along with their matching segments. Results The working prototype of BOSS is available at http://boss.korea.ac.kr. The current version of BOSS has indexed abstracts of more than 20 million articles published during last 16 years from 1996 to 2011 across all science disciplines. Conclusion BOSS fills the gap between either ends of the spectrum by allowing users to pose context-free queries and by returning a structured set of results. Furthermore, BOSS exhibits the characteristic of good scalability, just as with conventional document search engines, because it is designed to use a standard document-indexing model with minimal modifications. Considering the features, BOSS notches up the technological level of traditional solutions for search on biomedical information. PMID:22595092
Do People Experience Cognitive Biases while Searching for Information?
Lau, Annie Y.S.; Coiera, Enrico W.
2007-01-01
Objective To test whether individuals experience cognitive biases whilst searching using information retrieval systems. Biases investigated are anchoring, order, exposure and reinforcement. Design A retrospective analysis and a prospective experiment were conducted to investigate whether cognitive biases affect the way that documentary evidence is interpreted while searching online. The retrospective analysis was conducted on the search and decision behaviors of 75 clinicians (44 doctors, 31 nurses), answering questions for 8 clinical scenarios within 80 minutes in a controlled setting. The prospective study was conducted on 227 undergraduate students, who used the same search engine to answer two of six randomly assigned consumer health questions. Measurements Frequencies of correct answers pre- and post- search, and confidence in answers were collected. The impact of reading a document on the final decision was measured by the population likelihood ratio (LR) of the frequency of reading the document and the frequency of obtaining a correct answer. Documents with a LR > 1 were most likely to be associated with a correct answer, and those with a LR < 1 were most likely to be associated with an incorrect answer to a question. Agreement between a subject and the evidence they read was estimated by a concurrence rate, which measured the frequency that subjects’ answers agreed with the likelihood ratios of a group of documents, normalized for document order, time exposure or reinforcement through repeated access. Serial position curves were plotted for the relationship between subjects’ pre-search confidence, document order, the number of times and length of time a document was accessed, and concurrence with post-search answers. Chi-square analyses tested for the presence of biases, and the Kolmogorov-Smirnov test checked for equality of distribution of evidence in the comparison populations. Results A person’s prior belief (anchoring) has a significant impact on their post-search answer (retrospective: P < 0.001; prospective: P < 0.001). Documents accessed at different positions in a search session (order effect [retrospective: P = 0.76; prospective: P = 0.026]), and documents processed for different lengths of time (exposure effect [retrospective: P = 0.27; prospective: P = 0.0081]) also influenced decision post-search more than expected in the prospective experiment but not in the retrospective analysis. Reinforcement through repeated exposure to a document did not yield statistical differences in decision outcome post-search (retrospective: P = 0.31; prospective: P = 0.81). Conclusion People may experience anchoring, exposure and order biases while searching for information, and these biases may influence the quality of decision making during and after the use of information retrieval systems. PMID:17600097
NASA Technical Reports Server (NTRS)
Driscoll, James N.
1994-01-01
The high-speed data search system developed for KSC incorporates existing and emerging information retrieval technology to help a user intelligently and rapidly locate information found in large textual databases. This technology includes: natural language input; statistical ranking of retrieved information; an artificial intelligence concept called semantics, where 'surface level' knowledge found in text is used to improve the ranking of retrieved information; and relevance feedback, where user judgements about viewed information are used to automatically modify the search for further information. Semantics and relevance feedback are features of the system which are not available commercially. The system further demonstrates focus on paragraphs of information to decide relevance; and it can be used (without modification) to intelligently search all kinds of document collections, such as collections of legal documents medical documents, news stories, patents, and so forth. The purpose of this paper is to demonstrate the usefulness of statistical ranking, our semantic improvement, and relevance feedback.
Nearest Neighbor Searching in Binary Search Trees: Simulation of a Multiprocessor System.
ERIC Educational Resources Information Center
Stewart, Mark; Willett, Peter
1987-01-01
Describes the simulation of a nearest neighbor searching algorithm for document retrieval using a pool of microprocessors. Three techniques are described which allow parallel searching of a binary search tree as well as a PASCAL-based system, PASSIM, which can simulate these techniques. Fifty-six references are provided. (Author/LRW)
Misra, Dharitri; Chen, Siyuan; Thoma, George R
2009-01-01
One of the most expensive aspects of archiving digital documents is the manual acquisition of context-sensitive metadata useful for the subsequent discovery of, and access to, the archived items. For certain types of textual documents, such as journal articles, pamphlets, official government records, etc., where the metadata is contained within the body of the documents, a cost effective method is to identify and extract the metadata in an automated way, applying machine learning and string pattern search techniques.At the U. S. National Library of Medicine (NLM) we have developed an automated metadata extraction (AME) system that employs layout classification and recognition models with a metadata pattern search model for a text corpus with structured or semi-structured information. A combination of Support Vector Machine and Hidden Markov Model is used to create the layout recognition models from a training set of the corpus, following which a rule-based metadata search model is used to extract the embedded metadata by analyzing the string patterns within and surrounding each field in the recognized layouts.In this paper, we describe the design of our AME system, with focus on the metadata search model. We present the extraction results for a historic collection from the Food and Drug Administration, and outline how the system may be adapted for similar collections. Finally, we discuss some ongoing enhancements to our AME system.
Personalization of Rule-based Web Services.
Choi, Okkyung; Han, Sang Yong
2008-04-04
Nowadays Web users have clearly expressed their wishes to receive personalized services directly. Personalization is the way to tailor services directly to the immediate requirements of the user. However, the current Web Services System does not provide any features supporting this such as consideration of personalization of services and intelligent matchmaking. In this research a flexible, personalized Rule-based Web Services System to address these problems and to enable efficient search, discovery and construction across general Web documents and Semantic Web documents in a Web Services System is proposed. This system utilizes matchmaking among service requesters', service providers' and users' preferences using a Rule-based Search Method, and subsequently ranks search results. A prototype of efficient Web Services search and construction for the suggested system is developed based on the current work.
Automated document analysis system
NASA Astrophysics Data System (ADS)
Black, Jeffrey D.; Dietzel, Robert; Hartnett, David
2002-08-01
A software application has been developed to aid law enforcement and government intelligence gathering organizations in the translation and analysis of foreign language documents with potential intelligence content. The Automated Document Analysis System (ADAS) provides the capability to search (data or text mine) documents in English and the most commonly encountered foreign languages, including Arabic. Hardcopy documents are scanned by a high-speed scanner and are optical character recognized (OCR). Documents obtained in an electronic format bypass the OCR and are copied directly to a working directory. For translation and analysis, the script and the language of the documents are first determined. If the document is not in English, the document is machine translated to English. The documents are searched for keywords and key features in either the native language or translated English. The user can quickly review the document to determine if it has any intelligence content and whether detailed, verbatim human translation is required. The documents and document content are cataloged for potential future analysis. The system allows non-linguists to evaluate foreign language documents and allows for the quick analysis of a large quantity of documents. All document processing can be performed manually or automatically on a single document or a batch of documents.
FAPA: Faculty Appointment Policy Archive, 1998. [CD-ROM.
ERIC Educational Resources Information Center
Trower, C. Ann
This CD-ROM presents 220 documents collected in Harvard University's Faculty Appointment Policy Archive (FAPA), the ZyFIND search and retrieval system, and instructions for their use. The FAPA system and ZyFIND allow browsing through documents, inserting bookmarks in documents, attaching notes to documents without modifying them, and selecting…
Update on CERN Search based on SharePoint 2013
NASA Astrophysics Data System (ADS)
Alvarez, E.; Fernandez, S.; Lossent, A.; Posada, I.; Silva, B.; Wagner, A.
2017-10-01
CERN’s enterprise Search solution “CERN Search” provides a central search solution for users and CERN service providers. A total of about 20 million public and protected documents from a wide range of document collections is indexed, including Indico, TWiki, Drupal, SharePoint, JACOW, E-group archives, EDMS, and CERN Web pages. In spring 2015, CERN Search was migrated to a new infrastructure based on SharePoint 2013. In the context of this upgrade, the document pre-processing and indexing process was redesigned and generalised. The new data feeding framework allows to profit from new functionality and it facilitates the long term maintenance of the system.
Computer-Assisted Search Of Large Textual Data Bases
NASA Technical Reports Server (NTRS)
Driscoll, James R.
1995-01-01
"QA" denotes high-speed computer system for searching diverse collections of documents including (but not limited to) technical reference manuals, legal documents, medical documents, news releases, and patents. Incorporates previously available and emerging information-retrieval technology to help user intelligently and rapidly locate information found in large textual data bases. Technology includes provision for inquiries in natural language; statistical ranking of retrieved information; artificial-intelligence implementation of semantics, in which "surface level" knowledge found in text used to improve ranking of retrieved information; and relevance feedback, in which user's judgements of relevance of some retrieved documents used automatically to modify search for further information.
Misra, Dharitri; Chen, Siyuan; Thoma, George R.
2010-01-01
One of the most expensive aspects of archiving digital documents is the manual acquisition of context-sensitive metadata useful for the subsequent discovery of, and access to, the archived items. For certain types of textual documents, such as journal articles, pamphlets, official government records, etc., where the metadata is contained within the body of the documents, a cost effective method is to identify and extract the metadata in an automated way, applying machine learning and string pattern search techniques. At the U. S. National Library of Medicine (NLM) we have developed an automated metadata extraction (AME) system that employs layout classification and recognition models with a metadata pattern search model for a text corpus with structured or semi-structured information. A combination of Support Vector Machine and Hidden Markov Model is used to create the layout recognition models from a training set of the corpus, following which a rule-based metadata search model is used to extract the embedded metadata by analyzing the string patterns within and surrounding each field in the recognized layouts. In this paper, we describe the design of our AME system, with focus on the metadata search model. We present the extraction results for a historic collection from the Food and Drug Administration, and outline how the system may be adapted for similar collections. Finally, we discuss some ongoing enhancements to our AME system. PMID:21179386
Automated Patent Categorization and Guided Patent Search using IPC as Inspired by MeSH and PubMed.
Eisinger, Daniel; Tsatsaronis, George; Bundschus, Markus; Wieneke, Ulrich; Schroeder, Michael
2013-04-15
Document search on PubMed, the pre-eminent database for biomedical literature, relies on the annotation of its documents with relevant terms from the Medical Subject Headings ontology (MeSH) for improving recall through query expansion. Patent documents are another important information source, though they are considerably less accessible. One option to expand patent search beyond pure keywords is the inclusion of classification information: Since every patent is assigned at least one class code, it should be possible for these assignments to be automatically used in a similar way as the MeSH annotations in PubMed. In order to develop a system for this task, it is necessary to have a good understanding of the properties of both classification systems. This report describes our comparative analysis of MeSH and the main patent classification system, the International Patent Classification (IPC). We investigate the hierarchical structures as well as the properties of the terms/classes respectively, and we compare the assignment of IPC codes to patents with the annotation of PubMed documents with MeSH terms.Our analysis shows a strong structural similarity of the hierarchies, but significant differences of terms and annotations. The low number of IPC class assignments and the lack of occurrences of class labels in patent texts imply that current patent search is severely limited. To overcome these limits, we evaluate a method for the automated assignment of additional classes to patent documents, and we propose a system for guided patent search based on the use of class co-occurrence information and external resources.
Automated Patent Categorization and Guided Patent Search using IPC as Inspired by MeSH and PubMed
2013-01-01
Document search on PubMed, the pre-eminent database for biomedical literature, relies on the annotation of its documents with relevant terms from the Medical Subject Headings ontology (MeSH) for improving recall through query expansion. Patent documents are another important information source, though they are considerably less accessible. One option to expand patent search beyond pure keywords is the inclusion of classification information: Since every patent is assigned at least one class code, it should be possible for these assignments to be automatically used in a similar way as the MeSH annotations in PubMed. In order to develop a system for this task, it is necessary to have a good understanding of the properties of both classification systems. This report describes our comparative analysis of MeSH and the main patent classification system, the International Patent Classification (IPC). We investigate the hierarchical structures as well as the properties of the terms/classes respectively, and we compare the assignment of IPC codes to patents with the annotation of PubMed documents with MeSH terms. Our analysis shows a strong structural similarity of the hierarchies, but significant differences of terms and annotations. The low number of IPC class assignments and the lack of occurrences of class labels in patent texts imply that current patent search is severely limited. To overcome these limits, we evaluate a method for the automated assignment of additional classes to patent documents, and we propose a system for guided patent search based on the use of class co-occurrence information and external resources. PMID:23734562
The present status and problems in document retrieval system : document input type retrieval system
NASA Astrophysics Data System (ADS)
Inagaki, Hirohito
The office-automation (OA) made many changes. Many documents were begun to maintained in an electronic filing system. Therefore, it is needed to establish efficient document retrieval system to extract useful information. Current document retrieval systems are using simple word-matching, syntactic-matching, semantic-matching to obtain high retrieval efficiency. On the other hand, the document retrieval systems using special hardware devices, such as ISSP, were developed for aiming high speed retrieval. Since these systems can accept a single sentence or keywords as input, it is difficult to explain searcher's request. We demonstrated document input type retrieval system, which can directly accept document as an input, and can search similar documents from document data-base.
What Friends Are For: Collaborative Intelligence Analysis and Search
2014-06-01
14. SUBJECT TERMS Intelligence Community, information retrieval, recommender systems , search engines, social networks, user profiling, Lucene...improvements over existing search systems . The improvements are shown to be robust to high levels of human error and low similarity between users ...precision NOLH nearly orthogonal Latin hypercubes P@ precision at documents RS recommender systems TREC Text REtrieval Conference USM user
Implementation of the common phrase index method on the phrase query for information retrieval
NASA Astrophysics Data System (ADS)
Fatmawati, Triyah; Zaman, Badrus; Werdiningsih, Indah
2017-08-01
As the development of technology, the process of finding information on the news text is easy, because the text of the news is not only distributed in print media, such as newspapers, but also in electronic media that can be accessed using the search engine. In the process of finding relevant documents on the search engine, a phrase often used as a query. The number of words that make up the phrase query and their position obviously affect the relevance of the document produced. As a result, the accuracy of the information obtained will be affected. Based on the outlined problem, the purpose of this research was to analyze the implementation of the common phrase index method on information retrieval. This research will be conducted in English news text and implemented on a prototype to determine the relevance level of the documents produced. The system is built with the stages of pre-processing, indexing, term weighting calculation, and cosine similarity calculation. Then the system will display the document search results in a sequence, based on the cosine similarity. Furthermore, system testing will be conducted using 100 documents and 20 queries. That result is then used for the evaluation stage. First, determine the relevant documents using kappa statistic calculation. Second, determine the system success rate using precision, recall, and F-measure calculation. In this research, the result of kappa statistic calculation was 0.71, so that the relevant documents are eligible for the system evaluation. Then the calculation of precision, recall, and F-measure produces precision of 0.37, recall of 0.50, and F-measure of 0.43. From this result can be said that the success rate of the system to produce relevant documents is low.
Clinician search behaviors may be influenced by search engine design.
Lau, Annie Y S; Coiera, Enrico; Zrimec, Tatjana; Compton, Paul
2010-06-30
Searching the Web for documents using information retrieval systems plays an important part in clinicians' practice of evidence-based medicine. While much research focuses on the design of methods to retrieve documents, there has been little examination of the way different search engine capabilities influence clinician search behaviors. Previous studies have shown that use of task-based search engines allows for faster searches with no loss of decision accuracy compared with resource-based engines. We hypothesized that changes in search behaviors may explain these differences. In all, 75 clinicians (44 doctors and 31 clinical nurse consultants) were randomized to use either a resource-based or a task-based version of a clinical information retrieval system to answer questions about 8 clinical scenarios in a controlled setting in a university computer laboratory. Clinicians using the resource-based system could select 1 of 6 resources, such as PubMed; clinicians using the task-based system could select 1 of 6 clinical tasks, such as diagnosis. Clinicians in both systems could reformulate search queries. System logs unobtrusively capturing clinicians' interactions with the systems were coded and analyzed for clinicians' search actions and query reformulation strategies. The most frequent search action of clinicians using the resource-based system was to explore a new resource with the same query, that is, these clinicians exhibited a "breadth-first" search behaviour. Of 1398 search actions, clinicians using the resource-based system conducted 401 (28.7%, 95% confidence interval [CI] 26.37-31.11) in this way. In contrast, the majority of clinicians using the task-based system exhibited a "depth-first" search behavior in which they reformulated query keywords while keeping to the same task profiles. Of 585 search actions conducted by clinicians using the task-based system, 379 (64.8%, 95% CI 60.83-68.55) were conducted in this way. This study provides evidence that different search engine designs are associated with different user search behaviors.
Balancing Efficiency and Effectiveness for Fusion-Based Search Engines in the "Big Data" Environment
ERIC Educational Resources Information Center
Li, Jieyu; Huang, Chunlan; Wang, Xiuhong; Wu, Shengli
2016-01-01
Introduction: In the big data age, we have to deal with a tremendous amount of information, which can be collected from various types of sources. For information search systems such as Web search engines or online digital libraries, the collection of documents becomes larger and larger. For some queries, an information search system needs to…
Supporting the education evidence portal via text mining
Ananiadou, Sophia; Thompson, Paul; Thomas, James; Mu, Tingting; Oliver, Sandy; Rickinson, Mark; Sasaki, Yutaka; Weissenbacher, Davy; McNaught, John
2010-01-01
The UK Education Evidence Portal (eep) provides a single, searchable, point of access to the contents of the websites of 33 organizations relating to education, with the aim of revolutionizing work practices for the education community. Use of the portal alleviates the need to spend time searching multiple resources to find relevant information. However, the combined content of the websites of interest is still very large (over 500 000 documents and growing). This means that searches using the portal can produce very large numbers of hits. As users often have limited time, they would benefit from enhanced methods of performing searches and viewing results, allowing them to drill down to information of interest more efficiently, without having to sift through potentially long lists of irrelevant documents. The Joint Information Systems Committee (JISC)-funded ASSIST project has produced a prototype web interface to demonstrate the applicability of integrating a number of text-mining tools and methods into the eep, to facilitate an enhanced searching, browsing and document-viewing experience. New features include automatic classification of documents according to a taxonomy, automatic clustering of search results according to similar document content, and automatic identification and highlighting of key terms within documents. PMID:20643679
Spotting words in handwritten Arabic documents
NASA Astrophysics Data System (ADS)
Srihari, Sargur; Srinivasan, Harish; Babu, Pavithra; Bhole, Chetan
2006-01-01
The design and performance of a system for spotting handwritten Arabic words in scanned document images is presented. Three main components of the system are a word segmenter, a shape based matcher for words and a search interface. The user types in a query in English within a search window, the system finds the equivalent Arabic word, e.g., by dictionary look-up, locates word images in an indexed (segmented) set of documents. A two-step approach is employed in performing the search: (1) prototype selection: the query is used to obtain a set of handwritten samples of that word from a known set of writers (these are the prototypes), and (2) word matching: the prototypes are used to spot each occurrence of those words in the indexed document database. A ranking is performed on the entire set of test word images-- where the ranking criterion is a similarity score between each prototype word and the candidate words based on global word shape features. A database of 20,000 word images contained in 100 scanned handwritten Arabic documents written by 10 different writers was used to study retrieval performance. Using five writers for providing prototypes and the other five for testing, using manually segmented documents, 55% precision is obtained at 50% recall. Performance increases as more writers are used for training.
Development of a full-text information retrieval system
DOE Office of Scientific and Technical Information (OSTI.GOV)
Keizo Oyama; AKira Miyazawa, Atsuhiro Takasu; Kouji Shibano
The authors have executed a project to realize a full-text information retrieval system. The system is designed to deal with a document database comprising full text of a large number of documents such as academic papers. The document structures are utilized in searching and extracting appropriate information. The concept of structure handling and the configuration of the system are described in this paper.
Repetition and Diversification in Multi-Session Task Oriented Search
ERIC Educational Resources Information Center
Tyler, Sarah K.
2013-01-01
As the number of documents and the availability of information online grows, so to can the difficulty in sifting through documents to find what we're searching for. Traditional Information Retrieval (IR) systems consider the query as the representation of the user's needs, and as such are limited to the user's ability to describe the information…
Information Storage and Retrieval. Reports on Analysis, Search, and Iterative Retrieval.
ERIC Educational Resources Information Center
Salton, Gerard
As the fourteenth report in a series describing research in automatic information storage and retrieval, this document covers work carried out on the SMART project for approximately one year (summer 1967 to summer 1968). The document is divided into four main parts: (1) SMART systems design, (2) analysis and search experiments, (3) user feedback…
An exponentiation method for XML element retrieval.
Wichaiwong, Tanakorn
2014-01-01
XML document is now widely used for modelling and storing structured documents. The structure is very rich and carries important information about contents and their relationships, for example, e-Commerce. XML data-centric collections require query terms allowing users to specify constraints on the document structure; mapping structure queries and assigning the weight are significant for the set of possibly relevant documents with respect to structural conditions. In this paper, we present an extension to the MEXIR search system that supports the combination of structural and content queries in the form of content-and-structure queries, which we call the Exponentiation function. It has been shown the structural information improve the effectiveness of the search system up to 52.60% over the baseline BM25 at MAP.
Search Search Home SH Reference Manual E19 Documentation Program Management Training/Drills Other Dataweb National Water Information System Database SH Reference Manual, E-19 Docs, Program Management
ERIC Educational Resources Information Center
Girill, T. R.
1991-01-01
This article continues the description of DFT (Document, Find, Theseus), an online documentation system that provides computer-managed on-demand printing of software manuals as well as the interactive retrieval of reference passages. Document boundaries in the hypertext database are discussed, search vocabulary complexities are described, and text…
The JPL Library information retrieval system
NASA Technical Reports Server (NTRS)
Walsh, J.
1975-01-01
The development, capabilities, and products of the computer-based retrieval system of the Jet Propulsion Laboratory Library are described. The system handles books and documents, produces a book catalog, and provides a machine search capability. Programs and documentation are available to the public through NASA's computer software dissemination program.
Essie: A Concept-based Search Engine for Structured Biomedical Text
Ide, Nicholas C.; Loane, Russell F.; Demner-Fushman, Dina
2007-01-01
This article describes the algorithms implemented in the Essie search engine that is currently serving several Web sites at the National Library of Medicine. Essie is a phrase-based search engine with term and concept query expansion and probabilistic relevancy ranking. Essie’s design is motivated by an observation that query terms are often conceptually related to terms in a document, without actually occurring in the document text. Essie’s performance was evaluated using data and standard evaluation methods from the 2003 and 2006 Text REtrieval Conference (TREC) Genomics track. Essie was the best-performing search engine in the 2003 TREC Genomics track and achieved results comparable to those of the highest-ranking systems on the 2006 TREC Genomics track task. Essie shows that a judicious combination of exploiting document structure, phrase searching, and concept based query expansion is a useful approach for information retrieval in the biomedical domain. PMID:17329729
Front-End/Gateway Software: Availability and Usefulness.
ERIC Educational Resources Information Center
Kesselman, Martin
1985-01-01
Reviews features of front-end software packages (interface between user and online system)--database selection, search strategy development, saving and downloading, hardware and software requirements, training and documentation, online systems and database accession, and costs--and discusses gateway services (user searches through intermediary…
A comparison of Boolean-based retrieval to the WAIS system for retrieval of aeronautical information
NASA Technical Reports Server (NTRS)
Marchionini, Gary; Barlow, Diane
1994-01-01
An evaluation of an information retrieval system using a Boolean-based retrieval engine and inverted file architecture and WAIS, which uses a vector-based engine, was conducted. Four research questions in aeronautical engineering were used to retrieve sets of citations from the NASA Aerospace Database which was mounted on a WAIS server and available through Dialog File 108 which served as the Boolean-based system (BBS). High recall and high precision searches were done in the BBS and terse and verbose queries were used in the WAIS condition. Precision values for the WAIS searches were consistently above the precision values for high recall BBS searches and consistently below the precision values for high precision BBS searches. Terse WAIS queries gave somewhat better precision performance than verbose WAIS queries. In every case, a small number of relevant documents retrieved by one system were not retrieved by the other, indicating the incomplete nature of the results from either retrieval system. Relevant documents in the WAIS searches were found to be randomly distributed in the retrieved sets rather than distributed by ranks. Advantages and limitations of both types of systems are discussed.
Automated Text Markup for Information Retrieval from an Electronic Textbook of Infectious Disease
Berrios, Daniel C.; Kehler, Andrew; Kim, David K.; Yu, Victor L.; Fagan, Lawrence M.
1998-01-01
The information needs of practicing clinicians frequently require textbook or journal searches. Making these sources available in electronic form improves the speed of these searches, but precision (i.e., the fraction of relevant to total documents retrieved) remains low. Improving the traditional keyword search by transforming search terms into canonical concepts does not improve search precision greatly. Kim et al. have designed and built a prototype system (MYCIN II) for computer-based information retrieval from a forthcoming electronic textbook of infectious disease. The system requires manual indexing by experts in the form of complex text markup. However, this mark-up process is time consuming (about 3 person-hours to generate, review, and transcribe the index for each of 218 chapters). We have designed and implemented a system to semiautomate the markup process. The system, information extraction for semiautomated indexing of documents (ISAID), uses query models and existing information-extraction tools to provide support for any user, including the author of the source material, to mark up tertiary information sources quickly and accurately.
NASA Astrophysics Data System (ADS)
Jürgens, Björn; Herrero-Solana, Victor
2017-04-01
Patents are an essential information source used to monitor, track, and analyze nanotechnology. When it comes to search nanotechnology-related patents, a keyword search is often incomplete and struggles to cover such an interdisciplinary discipline. Patent classification schemes can reveal far better results since they are assigned by experts who classify the patent documents according to their technology. In this paper, we present the most important classifications to search nanotechnology patents and analyze how nanotechnology is covered in the main patent classification systems used in search systems nowadays: the International Patent Classification (IPC), the United States Patent Classification (USPC), and the Cooperative Patent Classification (CPC). We conclude that nanotechnology has a significantly better patent coverage in the CPC since considerable more nanotechnology documents were retrieved than by using other classifications, and thus, recommend its use for all professionals involved in nanotechnology patent searches.
An Exponentiation Method for XML Element Retrieval
2014-01-01
XML document is now widely used for modelling and storing structured documents. The structure is very rich and carries important information about contents and their relationships, for example, e-Commerce. XML data-centric collections require query terms allowing users to specify constraints on the document structure; mapping structure queries and assigning the weight are significant for the set of possibly relevant documents with respect to structural conditions. In this paper, we present an extension to the MEXIR search system that supports the combination of structural and content queries in the form of content-and-structure queries, which we call the Exponentiation function. It has been shown the structural information improve the effectiveness of the search system up to 52.60% over the baseline BM25 at MAP. PMID:24696643
Natural language information retrieval in digital libraries
DOE Office of Scientific and Technical Information (OSTI.GOV)
Strzalkowski, T.; Perez-Carballo, J.; Marinescu, M.
In this paper we report on some recent developments in joint NYU and GE natural language information retrieval system. The main characteristic of this system is the use of advanced natural language processing to enhance the effectiveness of term-based document retrieval. The system is designed around a traditional statistical backbone consisting of the indexer module, which builds inverted index files from pre-processed documents, and a retrieval engine which searches and ranks the documents in response to user queries. Natural language processing is used to (1) preprocess the documents in order to extract content-carrying terms, (2) discover inter-term dependencies and buildmore » a conceptual hierarchy specific to the database domain, and (3) process user`s natural language requests into effective search queries. This system has been used in NIST-sponsored Text Retrieval Conferences (TREC), where we worked with approximately 3.3 GBytes of text articles including material from the Wall Street Journal, the Associated Press newswire, the Federal Register, Ziff Communications`s Computer Library, Department of Energy abstracts, U.S. Patents and the San Jose Mercury News, totaling more than 500 million words of English. The system have been designed to facilitate its scalability to deal with ever increasing amounts of data. In particular, a randomized index-splitting mechanism has been installed which allows the system to create a number of smaller indexes that can be independently and efficiently searched.« less
National Centers for Environmental Prediction
Organization Search Enter text Search Navigation Bar End Cap Search EMC Go Branches Global Climate and Weather / VISION | About EMC EMC > GLOBAL BRANCH > GFS > HOME Home Implementations Documentation References Products Model Guidance Performance Developers VLab GLOBAL FORECAST SYSTEM Global Data
Dynamic reduction of dimensions of a document vector in a document search and retrieval system
Jiao, Yu; Potok, Thomas E.
2011-05-03
The method and system of the invention involves processing each new document (20) coming into the system into a document vector (16), and creating a document vector with reduced dimensionality (17) for comparison with the data model (15) without recomputing the data model (15). These operations are carried out by a first computer (11) while a second computer (12) updates the data model (18), which can be comprised of an initial large group of documents (19) and is premised on the computing an initial data model (13, 14, 15) to provide a reference point for determining document vectors from documents processed from the data stream (20).
A tutorial on information retrieval: basic terms and concepts
Zhou, Wei; Smalheiser, Neil R; Yu, Clement
2006-01-01
This informal tutorial is intended for investigators and students who would like to understand the workings of information retrieval systems, including the most frequently used search engines: PubMed and Google. Having a basic knowledge of the terms and concepts of information retrieval should improve the efficiency and productivity of searches. As well, this knowledge is needed in order to follow current research efforts in biomedical information retrieval and text mining that are developing new systems not only for finding documents on a given topic, but extracting and integrating knowledge across documents. PMID:16722601
Issues and solutions for storage, retrieval, and searching of MPEG-7 documents
NASA Astrophysics Data System (ADS)
Chang, Yuan-Chi; Lo, Ming-Ling; Smith, John R.
2000-10-01
The ongoing MPEG-7 standardization activity aims at creating a standard for describing multimedia content in order to facilitate the interpretation of the associated information content. Attempting to address a broad range of applications, MPEG-7 has defined a flexible framework consisting of Descriptors, Description Schemes, and Description Definition Language. Descriptors and Description Schemes describe features, structure and semantics of multimedia objects. They are written in the Description Definition Language (DDL). In the most recent revision, DDL applies XML (Extensible Markup Language) Schema with MPEG-7 extensions. DDL has constructs that support inclusion, inheritance, reference, enumeration, choice, sequence, and abstract type of Description Schemes and Descriptors. In order to enable multimedia systems to use MPEG-7, a number of important problems in storing, retrieving and searching MPEG-7 documents need to be solved. This paper reports on initial finding on issues and solutions of storing and accessing MPEG-7 documents. In particular, we discuss the benefits of using a virtual document management framework based on XML Access Server (XAS) in order to bridge the MPEG-7 multimedia applications and database systems. The need arises partly because MPEG-7 descriptions need customized storage schema, indexing and search engines. We also discuss issues arising in managing dependence and cross-description scheme search.
2007-12-01
Boyle, “Important issues in hypertext documentation usability,” In Proceedings of the 9th Annual international Conference on Systems Documentation...Tufte’s principles of information design to creating effective Web sites.” In Proceedings of the 15th Annual international Conference on Computer...usability,” In Proceedings of the 9th Annual international Conference on Systems Documentation (Chicago, Illinois, 1991). SIGDOC . ACM, New York, NY
Imaged Document Optical Correlation and Conversion System (IDOCCS)
NASA Astrophysics Data System (ADS)
Stalcup, Bruce W.; Dennis, Phillip W.; Dydyk, Robert B.
1999-03-01
Today, the paper document is fast becoming a thing of the past. With the rapid development of fast, inexpensive computing and storage devices, many government and private organizations are archiving their documents in electronic form (e.g., personnel records, medical records, patents, etc.). In addition, many organizations are converting their paper archives to electronic images, which are stored in a computer database. Because of this, there is a need to efficiently organize this data into comprehensive and accessible information resources. The Imaged Document Optical Correlation and Conversion System (IDOCCS) provides a total solution to the problem of managing and retrieving textual and graphic information from imaged document archives. At the heart of IDOCCS, optical correlation technology provides the search and retrieval capability of document images. The IDOCCS can be used to rapidly search for key words or phrases within the imaged document archives and can even determine the types of languages contained within a document. In addition, IDOCCS can automatically compare an input document with the archived database to determine if it is a duplicate, thereby reducing the overall resources required to maintain and access the document database. Embedded graphics on imaged pages can also be exploited, e.g., imaged documents containing an agency's seal or logo, or documents with a particular individual's signature block, can be singled out. With this dual capability, IDOCCS outperforms systems that rely on optical character recognition as a basis for indexing and storing only the textual content of documents for later retrieval.
Search and Graph Database Technologies for Biomedical Semantic Indexing: Experimental Analysis.
Segura Bedmar, Isabel; Martínez, Paloma; Carruana Martín, Adrián
2017-12-01
Biomedical semantic indexing is a very useful support tool for human curators in their efforts for indexing and cataloging the biomedical literature. The aim of this study was to describe a system to automatically assign Medical Subject Headings (MeSH) to biomedical articles from MEDLINE. Our approach relies on the assumption that similar documents should be classified by similar MeSH terms. Although previous work has already exploited the document similarity by using a k-nearest neighbors algorithm, we represent documents as document vectors by search engine indexing and then compute the similarity between documents using cosine similarity. Once the most similar documents for a given input document are retrieved, we rank their MeSH terms to choose the most suitable set for the input document. To do this, we define a scoring function that takes into account the frequency of the term into the set of retrieved documents and the similarity between the input document and each retrieved document. In addition, we implement guidelines proposed by human curators to annotate MEDLINE articles; in particular, the heuristic that says if 3 MeSH terms are proposed to classify an article and they share the same ancestor, they should be replaced by this ancestor. The representation of the MeSH thesaurus as a graph database allows us to employ graph search algorithms to quickly and easily capture hierarchical relationships such as the lowest common ancestor between terms. Our experiments show promising results with an F1 of 69% on the test dataset. To the best of our knowledge, this is the first work that combines search and graph database technologies for the task of biomedical semantic indexing. Due to its horizontal scalability, ElasticSearch becomes a real solution to index large collections of documents (such as the bibliographic database MEDLINE). Moreover, the use of graph search algorithms for accessing MeSH information could provide a support tool for cataloging MEDLINE abstracts in real time. ©Isabel Segura Bedmar, Paloma Martínez, Adrián Carruana Martín. Originally published in JMIR Medical Informatics (http://medinform.jmir.org), 01.12.2017.
NASA STI Program Seminar: Electronic documents
NASA Technical Reports Server (NTRS)
1994-01-01
The theme of this NASA Scientific and Technical Information Program Seminar was electronic documents. Topics covered included Electronic Documents Management at the CASI, the Impact of Electronic Publishing on User Expectations and Searching Image Record Management, Secondary Publisher Considerations for Electronic Journal Literature, and the Technical Manual Publishing On Demand System (TMPODS).
RDBMS Based Lexical Resource for Indian Heritage: The Case of Mahābhārata
NASA Astrophysics Data System (ADS)
Mani, Diwakar
The paper describes a lexical resource in the form of a relational database based indexing system for Sanskrit documents - Mahābhārata (MBh) as an example. The system is available online on http://sanskrit.jnu.ac.in/mb with input and output in Devanāgarī Unicode, using technologies such as RDBMS and Java Servlet. The system works as an interactive and multi-dimensional indexing system with search facility for MBh and has potentials for use as a generic system for all Sanskrit texts of similar structure. Currently, the system allows three types of searching facilities- 'Direct Search', 'Alphabetical Search' and 'Search by Classes'. The input triggers an indexing process by which a temporary index is created for the search string, and then clicking on any indexed word displays the details for that word and also a facility to search that word in some other online lexical resources.
Integrating query of relational and textual data in clinical databases: a case study.
Fisk, John M; Mutalik, Pradeep; Levin, Forrest W; Erdos, Joseph; Taylor, Caroline; Nadkarni, Prakash
2003-01-01
The authors designed and implemented a clinical data mart composed of an integrated information retrieval (IR) and relational database management system (RDBMS). Using commodity software, which supports interactive, attribute-centric text and relational searches, the mart houses 2.8 million documents that span a five-year period and supports basic IR features such as Boolean searches, stemming, and proximity and fuzzy searching. Results are relevance-ranked using either "total documents per patient" or "report type weighting." Non-curated medical text has a significant degree of malformation with respect to spelling and punctuation, which creates difficulties for text indexing and searching. Presently, the IR facilities of RDBMS packages lack the features necessary to handle such malformed text adequately. A robust IR+RDBMS system can be developed, but it requires integrating RDBMSs with third-party IR software. RDBMS vendors need to make their IR offerings more accessible to non-programmers.
Adding Hierarchical Objects to Relational Database General-Purpose XML-Based Information Managements
NASA Technical Reports Server (NTRS)
Lin, Shu-Chun; Knight, Chris; La, Tracy; Maluf, David; Bell, David; Tran, Khai Peter; Gawdiak, Yuri
2006-01-01
NETMARK is a flexible, high-throughput software system for managing, storing, and rapid searching of unstructured and semi-structured documents. NETMARK transforms such documents from their original highly complex, constantly changing, heterogeneous data formats into well-structured, common data formats in using Hypertext Markup Language (HTML) and/or Extensible Markup Language (XML). The software implements an object-relational database system that combines the best practices of the relational model utilizing Structured Query Language (SQL) with those of the object-oriented, semantic database model for creating complex data. In particular, NETMARK takes advantage of the Oracle 8i object-relational database model using physical-address data types for very efficient keyword searches of records across both context and content. NETMARK also supports multiple international standards such as WEBDAV for drag-and-drop file management and SOAP for integrated information management using Web services. The document-organization and -searching capabilities afforded by NETMARK are likely to make this software attractive for use in disciplines as diverse as science, auditing, and law enforcement.
The Electronic Documentation Project in the NASA mission control center environment
NASA Technical Reports Server (NTRS)
Wang, Lui; Leigh, Albert
1994-01-01
NASA's space programs like many other technical programs of its magnitude is supported by a large volume of technical documents. These documents are not only diverse but also abundant. Management, maintenance, and retrieval of these documents is a challenging problem by itself; but, relating and cross-referencing this wealth of information when it is all on a medium of paper is an even greater challenge. The Electronic Documentation Project (EDP) is to provide an electronic system capable of developing, distributing and controlling changes for crew/ground controller procedures and related documents. There are two primary motives for the solution. The first motive is to reduce the cost of maintaining the current paper based method of operations by replacing paper documents with electronic information storage and retrieval. And, the other is to improve the efficiency and provide enhanced flexibility in document usage. Initially, the current paper based system will be faithfully reproduced in an electronic format to be used in the document viewing system. In addition, this metaphor will have hypertext extensions. Hypertext features support basic functions such as full text searches, key word searches, data retrieval, and traversal between nodes of information as well as speeding up the data access rate. They enable related but separate documents to have relationships, and allow the user to explore information naturally through non-linear link traversals. The basic operational requirements of the document viewing system are to: provide an electronic corollary to the current method of paper based document usage; supplement and ultimately replace paper-based documents; maintain focused toward control center operations such as Flight Data File, Flight Rules and Console Handbook viewing; and be available NASA wide.
Abdulla, Ahmed AbdoAziz Ahmed; Lin, Hongfei; Xu, Bo; Banbhrani, Santosh Kumar
2016-07-25
Biomedical literature retrieval is becoming increasingly complex, and there is a fundamental need for advanced information retrieval systems. Information Retrieval (IR) programs scour unstructured materials such as text documents in large reserves of data that are usually stored on computers. IR is related to the representation, storage, and organization of information items, as well as to access. In IR one of the main problems is to determine which documents are relevant and which are not to the user's needs. Under the current regime, users cannot precisely construct queries in an accurate way to retrieve particular pieces of data from large reserves of data. Basic information retrieval systems are producing low-quality search results. In our proposed system for this paper we present a new technique to refine Information Retrieval searches to better represent the user's information need in order to enhance the performance of information retrieval by using different query expansion techniques and apply a linear combinations between them, where the combinations was linearly between two expansion results at one time. Query expansions expand the search query, for example, by finding synonyms and reweighting original terms. They provide significantly more focused, particularized search results than do basic search queries. The retrieval performance is measured by some variants of MAP (Mean Average Precision) and according to our experimental results, the combination of best results of query expansion is enhanced the retrieved documents and outperforms our baseline by 21.06 %, even it outperforms a previous study by 7.12 %. We propose several query expansion techniques and their combinations (linearly) to make user queries more cognizable to search engines and to produce higher-quality search results.
Analyzing Document Retrievability in Patent Retrieval Settings
NASA Astrophysics Data System (ADS)
Bashir, Shariq; Rauber, Andreas
Most information retrieval settings, such as web search, are typically precision-oriented, i.e. they focus on retrieving a small number of highly relevant documents. However, in specific domains, such as patent retrieval or law, recall becomes more relevant than precision: in these cases the goal is to find all relevant documents, requiring algorithms to be tuned more towards recall at the cost of precision. This raises important questions with respect to retrievability and search engine bias: depending on how the similarity between a query and documents is measured, certain documents may be more or less retrievable in certain systems, up to some documents not being retrievable at all within common threshold settings. Biases may be oriented towards popularity of documents (increasing weight of references), towards length of documents, favour the use of rare or common words; rely on structural information such as metadata or headings, etc. Existing accessibility measurement techniques are limited as they measure retrievability with respect to all possible queries. In this paper, we improve accessibility measurement by considering sets of relevant and irrelevant queries for each document. This simulates how recall oriented users create their queries when searching for relevant information. We evaluate retrievability scores using a corpus of patents from US Patent and Trademark Office.
An Optical Disk-Based Information Retrieval System.
ERIC Educational Resources Information Center
Bender, Avi
1988-01-01
Discusses a pilot project by the Nuclear Regulatory Commission to apply optical disk technology to the storage and retrieval of documents related to its high level waste management program. Components and features of the microcomputer-based system which provides full-text and image access to documents are described. A sample search is included.…
Documentation systems for educators seeking academic promotion in U.S. medical schools.
Simpson, Deborah; Hafler, Janet; Brown, Diane; Wilkerson, LuAnn
2004-08-01
To explore the state and use of teaching portfolios in promotion and tenure in U.S. medical schools. A two-phase qualitative study using a Web-based search procedure and telephone interviews was conducted. The first phase assessed the penetration of teaching portfolio-like systems in U.S. medical schools using a keyword search of medical school Web sites. The second phase examined the current use of teaching portfolios in 16 U.S. medical schools that reported their use in a survey in 1992. The individual designated as having primary responsibility for faculty appointments/promotions was contacted to participate in a 30-60 minute interview. The Phase 1 search of U.S. medical schools' Web sites revealed that 76 medical schools have Web-based access to information on documenting educational activities for promotion. A total of 16 of 17 medical schools responded to Phase 2. All 16 continued to use a portfolio-like system in 2003. Two documentation categories, honors/awards and philosophy/personal statement regarding education, were included by six more of these schools than used these categories in 1992. Dissemination of work to colleagues is now a key inclusion at 15 of the Phase 2 schools. The most common type of evidence used to document education was learner and/or peer ratings with infrequent use of outcome measures and internal/external review. The number of medical schools whose promotion packets include portfolio-like documentation associated with a faculty member's excellence in education has increased by more than 400% in just over ten years. Among early-responder schools the types of documentation categories have increased, but students' ratings of teaching remain the primary evidence used to document the quality or outcomes of the educational efforts reported.
Ontology-based reusable clinical document template production system.
Nam, Sejin; Lee, Sungin; Kim, James G Boram; Kim, Hong-Gee
2012-01-01
Clinical documents embody professional clinical knowledge. This paper shows an effective clinical document template (CDT) production system that uses a clinical description entity (CDE) model, a CDE ontology, and a knowledge management system called STEP that manages ontology-based clinical description entities. The ontology represents CDEs and their inter-relations, and the STEP system stores and manages CDE ontology-based information regarding CDTs. The system also provides Web Services interfaces for search and reasoning over clinical entities. The system was populated with entities and relations extracted from 35 CDTs that were used in admission, discharge, and progress reports, as well as those used in nursing and operation functions. A clinical document template editor is shown that uses STEP.
Duftschmid, Georg; Rinner, Christoph; Kohler, Michael; Huebner-Bloder, Gudrun; Saboor, Samrend; Ammenwerth, Elske
2013-12-01
While contributing to an improved continuity of care, Shared Electronic Health Record (EHR) systems may also lead to information overload of healthcare providers. Document-oriented architectures, such as the commonly employed IHE XDS profile, which only support information retrieval at the level of documents, are particularly susceptible for this problem. The objective of the EHR-ARCHE project was to develop a methodology and a prototype to efficiently satisfy healthcare providers' information needs when accessing a patient's Shared EHR during a treatment situation. We especially aimed to investigate whether this objective can be reached by integrating EHR Archetypes into an IHE XDS environment. Using methodical triangulation, we first analysed the information needs of healthcare providers, focusing on the treatment of diabetes patients as an exemplary application domain. We then designed ISO/EN 13606 Archetypes covering the identified information needs. To support a content-based search for fine-grained information items within EHR documents, we extended the IHE XDS environment with two additional actors. Finally, we conducted a formative and summative evaluation of our approach within a controlled study. We identified 446 frequently needed diabetes-specific information items, representing typical information needs of healthcare providers. We then created 128 Archetypes and 120 EHR documents for two fictive patients. All seven diabetes experts, who evaluated our approach, preferred the content-based search to a conventional XDS search. Success rates of finding relevant information was higher for the content-based search (100% versus 80%) and the latter was also more time-efficient (8-14min versus 20min or more). Our results show that for an efficient satisfaction of health care providers' information needs, a content-based search that rests upon the integration of Archetypes into an IHE XDS-based Shared EHR system is superior to a conventional metadata-based XDS search. Copyright © 2013 The Authors. Published by Elsevier Ireland Ltd.. All rights reserved.
Duftschmid, Georg; Rinner, Christoph; Kohler, Michael; Huebner-Bloder, Gudrun; Saboor, Samrend; Ammenwerth, Elske
2013-01-01
Purpose While contributing to an improved continuity of care, Shared Electronic Health Record (EHR) systems may also lead to information overload of healthcare providers. Document-oriented architectures, such as the commonly employed IHE XDS profile, which only support information retrieval at the level of documents, are particularly susceptible for this problem. The objective of the EHR-ARCHE project was to develop a methodology and a prototype to efficiently satisfy healthcare providers’ information needs when accessing a patient's Shared EHR during a treatment situation. We especially aimed to investigate whether this objective can be reached by integrating EHR Archetypes into an IHE XDS environment. Methods Using methodical triangulation, we first analysed the information needs of healthcare providers, focusing on the treatment of diabetes patients as an exemplary application domain. We then designed ISO/EN 13606 Archetypes covering the identified information needs. To support a content-based search for fine-grained information items within EHR documents, we extended the IHE XDS environment with two additional actors. Finally, we conducted a formative and summative evaluation of our approach within a controlled study. Results We identified 446 frequently needed diabetes-specific information items, representing typical information needs of healthcare providers. We then created 128 Archetypes and 120 EHR documents for two fictive patients. All seven diabetes experts, who evaluated our approach, preferred the content-based search to a conventional XDS search. Success rates of finding relevant information was higher for the content-based search (100% versus 80%) and the latter was also more time-efficient (8–14 min versus 20 min or more). Conclusions Our results show that for an efficient satisfaction of health care providers’ information needs, a content-based search that rests upon the integration of Archetypes into an IHE XDS-based Shared EHR system is superior to a conventional metadata-based XDS search. PMID:23999002
PubMed Interact: an Interactive Search Application for MEDLINE/PubMed
Muin, Michael; Fontelo, Paul; Ackerman, Michael
2006-01-01
Online search and retrieval systems are important resources for medical literature research. Progressive Web 2.0 technologies provide opportunities to improve search strategies and user experience. Using PHP, Document Object Model (DOM) manipulation and Asynchronous JavaScript and XML (Ajax), PubMed Interact allows greater functionality so users can refine search parameters with ease and interact with the search results to retrieve and display relevant information and related articles. PMID:17238658
Statistical Techniques for Efficient Indexing and Retrieval of Document Images
ERIC Educational Resources Information Center
Bhardwaj, Anurag
2010-01-01
We have developed statistical techniques to improve the performance of document image search systems where the intermediate step of OCR based transcription is not used. Previous research in this area has largely focused on challenges pertaining to generation of small lexicons for processing handwritten documents and enhancement of poor quality…
Online Catalog Documentation Task Force: Onscreen Documentation Subgroup. Final Report.
ERIC Educational Resources Information Center
Campbell, John; And Others
This report describes the work of the Onscreen Documentation Subgroup in revising screens for GALIN, the University of Georgia (UGA) Libraries new online system. The responsibilities of the Subgroup were to revise screens from the previous online catalog and to create basic help screens, primarily for the command searching mode. Responsibilities…
Skip Navigation Links www.nws.noaa.gov NOAA logo - Click to go to the NOAA home page National Weather Service NWS logo - Click to go to the NWS home page Climate Forecast System Home News Organization Search : Go Search Go CFS Home CFS version 2 News Documentation Downloads Reanalysis CFSv2 at CPC CFS
Imaged document information location and extraction using an optical correlator
NASA Astrophysics Data System (ADS)
Stalcup, Bruce W.; Dennis, Phillip W.; Dydyk, Robert B.
1999-12-01
Today, the paper document is fast becoming a thing of the past. With the rapid development of fast, inexpensive computing and storage devices, many government and private organizations are archiving their documents in electronic form (e.g., personnel records, medical records, patents, etc.). Many of these organizations are converting their paper archives to electronic images, which are then stored in a computer database. Because of this, there is a need to efficiently organize this data into comprehensive and accessible information resources and provide for rapid access to the information contained within these imaged documents. To meet this need, Litton PRC and Litton Data Systems Division are developing a system, the Imaged Document Optical Correlation and Conversion System (IDOCCS), to provide a total solution to the problem of managing and retrieving textual and graphic information from imaged document archives. At the heart of IDOCCS, optical correlation technology provide a means for the search and retrieval of information from imaged documents. IDOCCS can be used to rapidly search for key words or phrases within the imaged document archives and has the potential to determine the types of languages contained within a document. In addition, IDOCCS can automatically compare an input document with the archived database to determine if it is a duplicate, thereby reducing the overall resources required to maintain and access the document database. Embedded graphics on imaged pages can also be exploited, e.g., imaged documents containing an agency's seal or logo can be singled out. In this paper, we present a description of IDOCCS as well as preliminary performance results and theoretical projections.
Combining the Bourne-Shell, sed and awk in the UNIX Environment for Language Analysis.
ERIC Educational Resources Information Center
Schmitt, Lothar M.; Christianson, Kiel T.
This document describes how to construct tools for language analysis in research and teaching using the Bourne-shell, sed, and awk, three search tools, in the UNIX operating system. Applications include: searches for words, phrases, grammatical patterns, and phonemic patterns in text; statistical analysis of text in regard to such searches,…
Information extraction for enhanced access to disease outbreak reports.
Grishman, Ralph; Huttunen, Silja; Yangarber, Roman
2002-08-01
Document search is generally based on individual terms in the document. However, for collections within limited domains it is possible to provide more powerful access tools. This paper describes a system designed for collections of reports of infectious disease outbreaks. The system, Proteus-BIO, automatically creates a table of outbreaks, with each table entry linked to the document describing that outbreak; this makes it possible to use database operations such as selection and sorting to find relevant documents. Proteus-BIO consists of a Web crawler which gathers relevant documents; an information extraction engine which converts the individual outbreak events to a tabular database; and a database browser which provides access to the events and, through them, to the documents. The information extraction engine uses sets of patterns and word classes to extract the information about each event. Preparing these patterns and word classes has been a time-consuming manual operation in the past, but automated discovery tools now make this task significantly easier. A small study comparing the effectiveness of the tabular index with conventional Web search tools demonstrated that users can find substantially more documents in a given time period with Proteus-BIO.
Applying Hypertext Structures to Software Documentation.
ERIC Educational Resources Information Center
French, James C.; And Others
1997-01-01
Describes a prototype system for software documentation management called SLEUTH (Software Literacy Enhancing Usefulness to Humans) being developed at the University of Virginia. Highlights include information retrieval techniques, hypertext links that are installed automatically, a WAIS (Wide Area Information Server) search engine, user…
ERIC Educational Resources Information Center
Girill, T. R.; And Others
1991-01-01
Describes enhancements made to a hypertext information retrieval system at the National Energy Research Supercomputer Center (NERSC) called DFT (Document, Find, and Theseus). The enrichment of DFT's entry vocabulary is described, DFT and other hypertext systems are compared, and problems that occur due to the need for frequent updates are…
NASA Astrophysics Data System (ADS)
Srinivasa, K. G.; Shree Devi, B. N.
2017-10-01
String searching in documents has become a tedious task with the evolution of Big Data. Generation of large data sets demand for a high performance search algorithm in areas such as text mining, information retrieval and many others. The popularity of GPU's for general purpose computing has been increasing for various applications. Therefore it is of great interest to exploit the thread feature of a GPU to provide a high performance search algorithm. This paper proposes an optimized new approach to N-gram model for string search in a number of lengthy documents and its GPU implementation. The algorithm exploits GPGPUs for searching strings in many documents employing character level N-gram matching with parallel Score Table approach and search using CUDA API. The new approach of Score table used for frequency storage of N-grams in a document, makes the search independent of the document's length and allows faster access to the frequency values, thus decreasing the search complexity. The extensive thread feature in a GPU has been exploited to enable parallel pre-processing of trigrams in a document for Score Table creation and parallel search in huge number of documents, thus speeding up the whole search process even for a large pattern size. Experiments were carried out for many documents of varied length and search strings from the standard Lorem Ipsum text on NVIDIA's GeForce GT 540M GPU with 96 cores. Results prove that the parallel approach for Score Table creation and searching gives a good speed up than the same approach executed serially.
Pipelining Architecture of Indexing Using Agglomerative Clustering
NASA Astrophysics Data System (ADS)
Goyal, Deepika; Goyal, Deepti; Gupta, Parul
2010-11-01
The World Wide Web is an interlinked collection of billions of documents. Ironically the huge size of this collection has become an obstacle for information retrieval. To access the information from Internet, search engine is used. Search engine retrieve the pages from indexer. This paper introduce a novel pipelining technique for structuring the core index-building system that substantially reduces the index construction time and also clustering algorithm that aims at partitioning the set of documents into ordered clusters so that the documents within the same cluster are similar and are being assigned the closer document identifiers. After assigning to the clusters it creates the hierarchy of index so that searching is efficient. It will make the super cluster then mega cluster by itself. The pipeline architecture will create the index in such a way that it will be efficient in space and time saving manner. It will direct the search from higher level to lower level of index or higher level of clusters to lower level of cluster so that the user gets the possible match result in time saving manner. As one cluster is making by taking only two clusters so it search is limited to two clusters for lower level of index and so on. So it is efficient in time saving manner.
78 FR 68832 - National Advisory Committee on Institutional Quality and Integrity (NACIQI)
Federal Register 2010, 2011, 2012, 2013, 2014
2013-11-15
... process for institutions of higher education under Title IV, of the HEA, together with recommendations for... System at: www.gpo.gov/fdsys . At this site you can view this document, as well as all other documents of... documents of the Department published in the Federal Register by using the article search feature at: www...
New Tools to Document and Manage Data/Metadata: Example NGEE Arctic and UrbIS
NASA Astrophysics Data System (ADS)
Crow, M. C.; Devarakonda, R.; Hook, L.; Killeffer, T.; Krassovski, M.; Boden, T.; King, A. W.; Wullschleger, S. D.
2016-12-01
Tools used for documenting, archiving, cataloging, and searching data are critical pieces of informatics. This discussion describes tools being used in two different projects at Oak Ridge National Laboratory (ORNL), but at different stages of the data lifecycle. The Metadata Entry and Data Search Tool is being used for the documentation, archival, and data discovery stages for the Next Generation Ecosystem Experiment - Arctic (NGEE Arctic) project while the Urban Information Systems (UrbIS) Data Catalog is being used to support indexing, cataloging, and searching. The NGEE Arctic Online Metadata Entry Tool [1] provides a method by which researchers can upload their data and provide original metadata with each upload. The tool is built upon a Java SPRING framework to parse user input into, and from, XML output. Many aspects of the tool require use of a relational database including encrypted user-login, auto-fill functionality for predefined sites and plots, and file reference storage and sorting. The UrbIS Data Catalog is a data discovery tool supported by the Mercury cataloging framework [2] which aims to compile urban environmental data from around the world into one location, and be searchable via a user-friendly interface. Each data record conveniently displays its title, source, and date range, and features: (1) a button for a quick view of the metadata, (2) a direct link to the data and, for some data sets, (3) a button for visualizing the data. The search box incorporates autocomplete capabilities for search terms and sorted keyword filters are available on the side of the page, including a map for searching by area. References: [1] Devarakonda, Ranjeet, et al. "Use of a metadata documentation and search tool for large data volumes: The NGEE arctic example." Big Data (Big Data), 2015 IEEE International Conference on. IEEE, 2015. [2] Devarakonda, R., Palanisamy, G., Wilson, B. E., & Green, J. M. (2010). Mercury: reusable metadata management, data discovery and access system. Earth Science Informatics, 3(1-2), 87-94.
JournalMap: Geo-semantic searching for relevant knowledge
USDA-ARS?s Scientific Manuscript database
Ecologists struggling to understand rapidly changing environments and evolving ecosystem threats need quick access to relevant research and documentation of natural systems. The advent of semantic and aggregation searching (e.g., Google Scholar, Web of Science) has made it easier to find useful lite...
Integrating Query of Relational and Textual Data in Clinical Databases: A Case Study
Fisk, John M.; Mutalik, Pradeep; Levin, Forrest W.; Erdos, Joseph; Taylor, Caroline; Nadkarni, Prakash
2003-01-01
Objectives: The authors designed and implemented a clinical data mart composed of an integrated information retrieval (IR) and relational database management system (RDBMS). Design: Using commodity software, which supports interactive, attribute-centric text and relational searches, the mart houses 2.8 million documents that span a five-year period and supports basic IR features such as Boolean searches, stemming, and proximity and fuzzy searching. Measurements: Results are relevance-ranked using either “total documents per patient” or “report type weighting.” Results: Non-curated medical text has a significant degree of malformation with respect to spelling and punctuation, which creates difficulties for text indexing and searching. Presently, the IR facilities of RDBMS packages lack the features necessary to handle such malformed text adequately. Conclusion: A robust IR+RDBMS system can be developed, but it requires integrating RDBMSs with third-party IR software. RDBMS vendors need to make their IR offerings more accessible to non-programmers. PMID:12509355
2011-09-01
search engines to find information. Most commercial search engines (Google, Yahoo, Bing, etc.) provide their indexing and search services...at no cost. The DoD can achieve large gains at a small cost by making public documents available to search engines . This can be achieved through the...were organized on the website dodreports.com. The results of this research revealed improvement gains of 8-20% for finding reports through commercial search engines during the first six months of
The JPL Library Information Retrieval System
ERIC Educational Resources Information Center
Walsh, Josephine
1975-01-01
The development, capabilities, and products of the computer-based retrieval system of the Jet Propulsion Laboratory Library are described. The system handles books and documents, produces a book catalog, and provides a machine search capability. (Author)
Relevance of Web Documents:Ghosts Consensus Method.
ERIC Educational Resources Information Center
Gorbunov, Andrey L.
2002-01-01
Discusses how to improve the quality of Internet search systems and introduces the Ghosts Consensus Method which is free from the drawbacks of digital democracy algorithms and is based on linear programming tasks. Highlights include vector space models; determining relevant documents; and enriching query terms. (LRW)
Saadawi, Gilan M; Harrison, James H
2006-10-01
Clinical laboratory procedure manuals are typically maintained as word processor files and are inefficient to store and search, require substantial effort for review and updating, and integrate poorly with other laboratory information. Electronic document management systems could improve procedure management and utility. As a first step toward building such systems, we have developed a prototype electronic format for laboratory procedures using Extensible Markup Language (XML). Representative laboratory procedures were analyzed to identify document structure and data elements. This information was used to create a markup vocabulary, CLP-ML, expressed as an XML Document Type Definition (DTD). To determine whether this markup provided advantages over generic markup, we compared procedures structured with CLP-ML or with the vocabulary of the Health Level Seven, Inc. (HL7) Clinical Document Architecture (CDA) narrative block. CLP-ML includes 124 XML tags and supports a variety of procedure types across different laboratory sections. When compared with a general-purpose markup vocabulary (CDA narrative block), CLP-ML documents were easier to edit and read, less complex structurally, and simpler to traverse for searching and retrieval. In combination with appropriate software, CLP-ML is designed to support electronic authoring, reviewing, distributing, and searching of clinical laboratory procedures from a central repository, decreasing procedure maintenance effort and increasing the utility of procedure information. A standard electronic procedure format could also allow laboratories and vendors to share procedures and procedure layouts, minimizing duplicative word processor editing. Our results suggest that laboratory-specific markup such as CLP-ML will provide greater benefit for such systems than generic markup.
ERIC Educational Resources Information Center
Herrera-Viedma, Enrique; Peis, Eduardo
2003-01-01
Presents a fuzzy evaluation method of SGML documents based on computing with words. Topics include filtering the amount of information available on the Web to assist users in their search processes; document type definitions; linguistic modeling; user-system interaction; and use with XML and other markup languages. (Author/LRW)
The CIS Database: Occupational Health and Safety Information Online.
ERIC Educational Resources Information Center
Siegel, Herbert; Scurr, Erica
1985-01-01
Describes document acquisition, selection, indexing, and abstracting and discusses online searching of the CIS database, an online system produced by the International Occupational Safety and Health Information Centre. This database comprehensively covers information in the field of occupational health and safety. Sample searches and search…
Miwa, Makoto; Ohta, Tomoko; Rak, Rafal; Rowley, Andrew; Kell, Douglas B.; Pyysalo, Sampo; Ananiadou, Sophia
2013-01-01
Motivation: To create, verify and maintain pathway models, curators must discover and assess knowledge distributed over the vast body of biological literature. Methods supporting these tasks must understand both the pathway model representations and the natural language in the literature. These methods should identify and order documents by relevance to any given pathway reaction. No existing system has addressed all aspects of this challenge. Method: We present novel methods for associating pathway model reactions with relevant publications. Our approach extracts the reactions directly from the models and then turns them into queries for three text mining-based MEDLINE literature search systems. These queries are executed, and the resulting documents are combined and ranked according to their relevance to the reactions of interest. We manually annotate document-reaction pairs with the relevance of the document to the reaction and use this annotation to study several ranking methods, using various heuristic and machine-learning approaches. Results: Our evaluation shows that the annotated document-reaction pairs can be used to create a rule-based document ranking system, and that machine learning can be used to rank documents by their relevance to pathway reactions. We find that a Support Vector Machine-based system outperforms several baselines and matches the performance of the rule-based system. The success of the query extraction and ranking methods are used to update our existing pathway search system, PathText. Availability: An online demonstration of PathText 2 and the annotated corpus are available for research purposes at http://www.nactem.ac.uk/pathtext2/. Contact: makoto.miwa@manchester.ac.uk Supplementary information: Supplementary data are available at Bioinformatics online. PMID:23813008
Keywords image retrieval in historical handwritten Arabic documents
NASA Astrophysics Data System (ADS)
Saabni, Raid; El-Sana, Jihad
2013-01-01
A system is presented for spotting and searching keywords in handwritten Arabic documents. A slightly modified dynamic time warping algorithm is used to measure similarities between words. Two sets of features are generated from the outer contour of the words/word-parts. The first set is based on the angles between nodes on the contour and the second set is based on the shape context features taken from the outer contour. To recognize a given word, the segmentation-free approach is partially adopted, i.e., continuous word parts are used as the basic alphabet, instead of individual characters or complete words. Additional strokes, such as dots and detached short segments, are classified and used in a postprocessing step to determine the final comparison decision. The search for a keyword is performed by the search for its word parts given in the correct order. The performance of the presented system was very encouraging in terms of efficiency and match rates. To evaluate the presented system its performance is compared to three different systems. Unfortunately, there are no publicly available standard datasets with ground truth for testing Arabic key word searching systems. Therefore, a private set of images partially taken from Juma'a Al-Majid Center in Dubai for evaluation is used, while using a slightly modified version of the IFN/ENIT database for training.
78 FR 29159 - Electric Power Research Institute; Seismic Evaluation Guidance
Federal Register 2010, 2011, 2012, 2013, 2014
2013-05-17
..., conducted field investigations, and used more recent methods than were previously available. In performing... available, by searching on http://www.regulations.gov under Docket ID NRC-2013-0038. Federal Rulemaking Web... Agencywide Documents Access and Management System (ADAMS): You may access publicly-available documents online...
Date Search Document Number Search Document Type Search Notice: Inactive UCCs (UCCs that are more than one year past their date of lapse) are purged from the database on an annual schedule. Documents are
ETDEWEB versus the World-Wide-Web: a specific database/web comparison
DOE Office of Scientific and Technical Information (OSTI.GOV)
Cutler, Debbie
2010-06-28
A study was performed comparing user search results from the specialized scientific database on energy-related information, ETDEWEB, with search results from the internet search engines Google and Google Scholar. The primary objective of the study was to determine if ETDEWEB (the Energy Technology Data Exchange – World Energy Base) continues to bring the user search results that are not being found by Google and Google Scholar. As a multilateral information exchange initiative, ETDE’s member countries and partners contribute cost- and task-sharing resources to build the largest database of energy-related information in the world. As of early 2010, the ETDEWEB databasemore » has 4.3 million citations to world-wide energy literature. One of ETDEWEB’s strengths is its focused scientific content and direct access to full text for its grey literature (over 300,000 documents in PDF available for viewing from the ETDE site and over a million additional links to where the documents can be found at research organizations and major publishers globally). Google and Google Scholar are well-known for the wide breadth of the information they search, with Google bringing in news, factual and opinion-related information, and Google Scholar also emphasizing scientific content across many disciplines. The analysis compared the results of 15 energy-related queries performed on all three systems using identical words/phrases. A variety of subjects was chosen, although the topics were mostly in renewable energy areas due to broad international interest. Over 40,000 search result records from the three sources were evaluated. The study concluded that ETDEWEB is a significant resource to energy experts for discovering relevant energy information. For the 15 topics in this study, ETDEWEB was shown to bring the user unique results not shown by Google or Google Scholar 86.7% of the time. Much was learned from the study beyond just metric comparisons. Observations about the strengths of each system and factors impacting the search results are also shared along with background information and summary tables of the results. If a user knows a very specific title of a document, all three systems are helpful in finding the user a source for the document. But if the user is looking to discover relevant documents on a specific topic, each of the three systems will bring back a considerable volume of data, but quite different in focus. Google is certainly a highly-used and valuable tool to find significant ‘non-specialist’ information, and Google Scholar does help the user focus on scientific disciplines. But if a user’s interest is scientific and energy-specific, ETDEWEB continues to hold a strong position in the energy research, technology and development (RTD) information field and adds considerable value in knowledge discovery. (auth)« less
Sifting Through Chaos: Extracting Information from Unstructured Legal Opinions.
Oliveira, Bruno Miguel; Guimarães, Rui Vasconcellos; Antunes, Luís; Rodrigues, Pedro Pereira
2018-01-01
Abiding to the law is, in some cases, a delicate balance between the rights of different players. Re-using health records is such a case. While the law grants reuse rights to public administration documents, in which health records produced in public health institutions are included, it also grants privacy to personal records. To safeguard a correct usage of data, public hospitals in Portugal employ jurists that are responsible for allowing or withholding access rights to health records. To help decision making, these jurists can consult the legal opinions issued by the national committee on public administration documents usage. While these legal opinions are of undeniable value, due to their doctrine contribution, they are only available in a format best suited from printing, forcing individual consultation of each document, with no option, whatsoever of clustered search, filtering or indexing, which are standard operations nowadays in a document management system. When having to decide on tens of data requests a day, it becomes unfeasible to consult the hundreds of legal opinions already available. With the objective to create a modern document management system, we devised an open, platform agnostic system that extracts and compiles the legal opinions, ex-tracts its contents and produces metadata, allowing for a fast searching and filtering of said legal opinions.
14 CFR 11.33 - How can I track FAA's rulemaking activities?
Code of Federal Regulations, 2010 CFR
2010-01-01
... rulemaking document proceeding. Each rulemaking document FAA issues in a particular rulemaking proceeding, as... search the Federal Docket Management System (FDMS) for information on most rulemaking proceedings. You....regulations.gov. If you can't find the material in the electronic docket, contact the person listed under FOR...
14 CFR 11.33 - How can I track FAA's rulemaking activities?
Code of Federal Regulations, 2014 CFR
2014-01-01
... rulemaking document proceeding. Each rulemaking document FAA issues in a particular rulemaking proceeding, as... search the Federal Docket Management System (FDMS) for information on most rulemaking proceedings. You....regulations.gov. If you can't find the material in the electronic docket, contact the person listed under FOR...
14 CFR 11.33 - How can I track FAA's rulemaking activities?
Code of Federal Regulations, 2012 CFR
2012-01-01
... rulemaking document proceeding. Each rulemaking document FAA issues in a particular rulemaking proceeding, as... search the Federal Docket Management System (FDMS) for information on most rulemaking proceedings. You....regulations.gov. If you can't find the material in the electronic docket, contact the person listed under FOR...
14 CFR 11.33 - How can I track FAA's rulemaking activities?
Code of Federal Regulations, 2011 CFR
2011-01-01
... rulemaking document proceeding. Each rulemaking document FAA issues in a particular rulemaking proceeding, as... search the Federal Docket Management System (FDMS) for information on most rulemaking proceedings. You....regulations.gov. If you can't find the material in the electronic docket, contact the person listed under FOR...
Information Retrieval Using UMLS-based Structured Queries
Fagan, Lawrence M.; Berrios, Daniel C.; Chan, Albert; Cucina, Russell; Datta, Anupam; Shah, Maulik; Surendran, Sujith
2001-01-01
During the last three years, we have developed and described components of ELBook, a semantically based information-retrieval system [1-4]. Using these components, domain experts can specify a query model, indexers can use the query model to index documents, and end-users can search these documents for instances of indexed queries.
Brehmer, Matthew; Ingram, Stephen; Stray, Jonathan; Munzner, Tamara
2014-12-01
For an investigative journalist, a large collection of documents obtained from a Freedom of Information Act request or a leak is both a blessing and a curse: such material may contain multiple newsworthy stories, but it can be difficult and time consuming to find relevant documents. Standard text search is useful, but even if the search target is known it may not be possible to formulate an effective query. In addition, summarization is an important non-search task. We present Overview, an application for the systematic analysis of large document collections based on document clustering, visualization, and tagging. This work contributes to the small set of design studies which evaluate a visualization system "in the wild", and we report on six case studies where Overview was voluntarily used by self-initiated journalists to produce published stories. We find that the frequently-used language of "exploring" a document collection is both too vague and too narrow to capture how journalists actually used our application. Our iterative process, including multiple rounds of deployment and observations of real world usage, led to a much more specific characterization of tasks. We analyze and justify the visual encoding and interaction techniques used in Overview's design with respect to our final task abstractions, and propose generalizable lessons for visualization design methodology.
Software Tools Streamline Project Management
NASA Technical Reports Server (NTRS)
2009-01-01
Three innovative software inventions from Ames Research Center (NETMARK, Program Management Tool, and Query-Based Document Management) are finding their way into NASA missions as well as industry applications. The first, NETMARK, is a program that enables integrated searching of data stored in a variety of databases and documents, meaning that users no longer have to look in several places for related information. NETMARK allows users to search and query information across all of these sources in one step. This cross-cutting capability in information analysis has exponentially reduced the amount of time needed to mine data from days or weeks to mere seconds. NETMARK has been used widely throughout NASA, enabling this automatic integration of information across many documents and databases. NASA projects that use NETMARK include the internal reporting system and project performance dashboard, Erasmus, NASA s enterprise management tool, which enhances organizational collaboration and information sharing through document routing and review; the Integrated Financial Management Program; International Space Station Knowledge Management; Mishap and Anomaly Information Reporting System; and management of the Mars Exploration Rovers. Approximately $1 billion worth of NASA s projects are currently managed using Program Management Tool (PMT), which is based on NETMARK. PMT is a comprehensive, Web-enabled application tool used to assist program and project managers within NASA enterprises in monitoring, disseminating, and tracking the progress of program and project milestones and other relevant resources. The PMT consists of an integrated knowledge repository built upon advanced enterprise-wide database integration techniques and the latest Web-enabled technologies. The current system is in a pilot operational mode allowing users to automatically manage, track, define, update, and view customizable milestone objectives and goals. The third software invention, Query-Based Document Management (QBDM) is a tool that enables content or context searches, either simple or hierarchical, across a variety of databases. The system enables users to specify notification subscriptions where they associate "contexts of interest" and "events of interest" to one or more documents or collection(s) of documents. Based on these subscriptions, users receive notification when the events of interest occur within the contexts of interest for associated document or collection(s) of documents. Users can also associate at least one notification time as part of the notification subscription, with at least one option for the time period of notifications.
ERIC Educational Resources Information Center
Jacob, Willis H.; Carter, Robert, III
This document features review questions, crossword puzzles, and word search puzzles on human anatomy. Topics include: (1) Anatomical Terminology; (2) The Skeletal System and Joints; (3) The Muscular System; (4) The Nervous System; (5) The Eye and Ear; (6) The Circulatory System and Blood; (7) The Respiratory System; (8) The Urinary System; (9) The…
78 FR 52996 - 60-Day Notice of Proposed Information Collection: Voluntary Disclosures.
Federal Register 2010, 2011, 2012, 2013, 2014
2013-08-27
... System (FDMS) to comment on this notice by going to www.regulations.gov . You may search for the document by entering ``Public Notice '' in the search bar. If necessary, use the ``narrow by agency'' filter... collection, including the validity of the methodology and assumptions used. Enhance the quality, utility, and...
78 FR 19695 - National Advisory Committee on Institutional Quality and Integrity (NACIQI)
Federal Register 2010, 2011, 2012, 2013, 2014
2013-04-02
... of higher education under Title IV, of the HEA, together with recommendations for improvement in such... available via the Federal Digital System at: www.gpo.gov/fdsys . At this site you can view this document, as... the article search feature at: www.federalregister.gov . Specifically, through the advanced search...
77 FR 64493 - National Advisory Committee on Institutional Quality and Integrity (NACIQI)
Federal Register 2010, 2011, 2012, 2013, 2014
2012-10-22
... process for institutions of higher education under Title IV, of the HEA, together with recommendations for... available via the Federal Digital System at: www.gpo.gov/fdsys . At this site you can view this document, as... the article search feature at: www.federalregister.gov . Specifically, through the advanced search...
Beyond Search Technique and Strategy: Helping Students to Be Informed.
ERIC Educational Resources Information Center
Aversa, Elizabeth
1991-01-01
Suggests ways that the school library media specialist can help students become more successful as users of online information. Four types of barriers to becoming informed are described: (1) indicative access barriers (indexing systems and relevance); (2) physical access to documents identified in a search; (3) linguistic barriers; and (4)…
Ad-Hoc Queries over Document Collections - A Case Study
NASA Astrophysics Data System (ADS)
Löser, Alexander; Lutter, Steffen; Düssel, Patrick; Markl, Volker
We discuss the novel problem of supporting analytical business intelligence queries over web-based textual content, e.g., BI-style reports based on 100.000's of documents from an ad-hoc web search result. Neither conventional search engines nor conventional Business Intelligence and ETL tools address this problem, which lies at the intersection of their capabilities. "Google Squared" or our system GOOLAP.info, are examples of these kinds of systems. They execute information extraction methods over one or several document collections at query time and integrate extracted records into a common view or tabular structure. Frequent extraction and object resolution failures cause incomplete records which could not be joined into a record answering the query. Our focus is the identification of join-reordering heuristics maximizing the size of complete records answering a structured query. With respect to given costs for document extraction we propose two novel join-operations: The multi-way CJ-operator joins records from multiple relationships extracted from a single document. The two-way join-operator DJ ensures data density by removing incomplete records from results. In a preliminary case study we observe that our join-reordering heuristics positively impact result size, record density and lower execution costs.
A Method for Search Engine Selection using Thesaurus for Selective Meta-Search Engine
NASA Astrophysics Data System (ADS)
Goto, Shoji; Ozono, Tadachika; Shintani, Toramatsu
In this paper, we propose a new method for selecting search engines on WWW for selective meta-search engine. In selective meta-search engine, a method is needed that would enable selecting appropriate search engines for users' queries. Most existing methods use statistical data such as document frequency. These methods may select inappropriate search engines if a query contains polysemous words. In this paper, we describe an search engine selection method based on thesaurus. In our method, a thesaurus is constructed from documents in a search engine and is used as a source description of the search engine. The form of a particular thesaurus depends on the documents used for its construction. Our method enables search engine selection by considering relationship between terms and overcomes the problems caused by polysemous words. Further, our method does not have a centralized broker maintaining data, such as document frequency for all search engines. As a result, it is easy to add a new search engine, and meta-search engines become more scalable with our method compared to other existing methods.
First Report of Using Portable Unmanned Aircraft Systems (Drones) for Search and Rescue.
Van Tilburg, Christopher
2017-06-01
Unmanned aircraft systems (UAS), colloquially called drones, are used commonly for military, government, and civilian purposes, including both commercial and consumer applications. During a search and rescue mission in Oregon, a UAS was used to confirm a fatality in a slot canyon; this eliminated the need for a dangerous rappel at night by rescue personnel. A second search mission in Oregon used several UAS to clear terrain. This allowed search of areas that were not accessible or were difficult to clear by ground personnel. UAS with cameras may be useful for searching, observing, and documenting missions. It is possible that UAS might be useful for delivering equipment in difficult areas and in communication. Copyright © 2017. Published by Elsevier Inc.
A User-Centered View of Document Delivery and Interlibrary Loan.
ERIC Educational Resources Information Center
Martin, Harry S., III; Kendrick, Curtis L.
1994-01-01
Discusses reasons why libraries are being forced to seek new forms of information storage. A hypothetical scenario of user-initiated document delivery alternatives integrated with a search process used by a professor using the Harvard OnLine Library Information System is presented. Extensive notes elaborate on the process and the technology…
DOE Research and Development Accomplishments Help
be used to search, locate, access, and electronically download full-text research and development (R Browse Downloading, Viewing, and/or Searching Full-text Documents/Pages Searching the Database Search Features Search allows you to search the OCRed full-text document and bibliographic information, the
ERIC Educational Resources Information Center
Borko, Harold
1985-01-01
Defines artificial intelligence (AI) and expert systems; describes library applications utilizing AI to automate creation of document representations, request formulations, and design and modify search strategies for information retrieval systems; discusses expert system development for information services; and reviews impact of these…
ERIC Educational Resources Information Center
Potocki, Anna; Ros, Christine; Vibert, Nicolas; Rouet, Jean-François
2017-01-01
This study examines children's strategies when scanning a document to answer a specific question. More specifically, we wanted to know whether they make use of organizers (i.e., headings) when searching and whether strategic search is related to their knowledge of reading strategies. Twenty-six French fifth graders were asked to search single-page…
US photovoltaic patents, 1951--1987
NASA Astrophysics Data System (ADS)
1988-09-01
This document contains 2195 U.S. patents on terrestrial photovoltaic (PV) power applications, including systems, components, and materials as well as manufacturing and support functions. The patent entries in this document were issued from 1951 through 1987; no patents were found in 1950. The entries were located by searching USPA, the data base of the U.S. Patent Office. The final search retrieved all patents under the class Batteries, Thermoelectric and Photoelectric, and the subclasses Photoelectric, Testing, and Applications. The search also located patents that contained the words photovoltaic(s) or solar cell(s) and their derivatives. A manual search of the patents in the Solar Energy Research Institute (SERI) patent file augmented the data base search. After the initial list was compiled, most of the patents on the following subjects were excluded: space photovoltaic technology, use of the photovoltaic effect for detectors, and subjects only peripherally concerned with photovoltaics. Some patents on these three subjects were included when it appeared that those inventions might be of use in terrwstrial PV power technologies.
Jones, B E; South, B R; Shao, Y; Lu, C C; Leng, J; Sauer, B C; Gundlapalli, A V; Samore, M H; Zeng, Q
2018-01-01
Identifying pneumonia using diagnosis codes alone may be insufficient for research on clinical decision making. Natural language processing (NLP) may enable the inclusion of cases missed by diagnosis codes. This article (1) develops a NLP tool that identifies the clinical assertion of pneumonia from physician emergency department (ED) notes, and (2) compares classification methods using diagnosis codes versus NLP against a gold standard of manual chart review to identify patients initially treated for pneumonia. Among a national population of ED visits occurring between 2006 and 2012 across the Veterans Affairs health system, we extracted 811 physician documents containing search terms for pneumonia for training, and 100 random documents for validation. Two reviewers annotated span- and document-level classifications of the clinical assertion of pneumonia. An NLP tool using a support vector machine was trained on the enriched documents. We extracted diagnosis codes assigned in the ED and upon hospital discharge and calculated performance characteristics for diagnosis codes, NLP, and NLP plus diagnosis codes against manual review in training and validation sets. Among the training documents, 51% contained clinical assertions of pneumonia; in the validation set, 9% were classified with pneumonia, of which 100% contained pneumonia search terms. After enriching with search terms, the NLP system alone demonstrated a recall/sensitivity of 0.72 (training) and 0.55 (validation), and a precision/positive predictive value (PPV) of 0.89 (training) and 0.71 (validation). ED-assigned diagnostic codes demonstrated lower recall/sensitivity (0.48 and 0.44) but higher precision/PPV (0.95 in training, 1.0 in validation); the NLP system identified more "possible-treated" cases than diagnostic coding. An approach combining NLP and ED-assigned diagnostic coding classification achieved the best performance (sensitivity 0.89 and PPV 0.80). System-wide application of NLP to clinical text can increase capture of initial diagnostic hypotheses, an important inclusion when studying diagnosis and clinical decision-making under uncertainty. Schattauer GmbH Stuttgart.
Microcomputer-Based Acquisitions.
ERIC Educational Resources Information Center
Desmarais, Norman
1986-01-01
This discussion of three automated acquisitions systems--Bib-Base/Acq, The Book Trak Ordering System, and Card Datalog Acquisitions Module--covers searching and updating, editing, acquisitions functions and statistics, purchase orders and order file, budgeting and accounts maintenance, defining parameters, documentation, security, printing, and…
A strategy for electronic dissemination of NASA Langley technical publications
NASA Technical Reports Server (NTRS)
Roper, Donna G.; Mccaskill, Mary K.; Holland, Scott D.; Walsh, Joanne L.; Nelson, Michael L.; Adkins, Susan L.; Ambur, Manjula Y.; Campbell, Bryan A.
1994-01-01
To demonstrate NASA Langley Research Center's relevance and to transfer technology to external customers in a timely and efficient manner, Langley has formed a working group to study and recommend a course of action for the electronic dissemination of technical reports (EDTR). The working group identified electronic report requirements (e.g., accessibility, file format, search requirements) of customers in U.S. industry through numerous site visits and personal contacts. Internal surveys were also used to determine commonalities in document preparation methods. From these surveys, a set of requirements for an electronic dissemination system was developed. Two candidate systems were identified and evaluated against the set of requirements: the Full-Text Electronic Documents System (FEDS), which is a full-text retrieval system based on the commercial document management package Interleaf, and the Langley Technical Report Server (LTRS), which is a Langley-developed system based on the publicly available World Wide Web (WWW) software system. Factors that led to the selection of LTRS as the vehicle for electronic dissemination included searching and viewing capability, current system operability, and client software availability for multiple platforms at no cost to industry. This report includes the survey results, evaluations, a description of the LTRS architecture, recommended policy statement, and suggestions for future implementations.
Search Interface Design Using Faceted Indexing for Web Resources.
ERIC Educational Resources Information Center
Devadason, Francis; Intaraksa, Neelawat; Patamawongjariya, Pornprapa; Desai, Kavita
2001-01-01
Describes an experimental system designed to organize and provide access to Web documents using a faceted pre-coordinate indexing system based on the Deep Structure Indexing System (DSIS) derived from POPSI (Postulate based Permuted Subject Indexing) of Bhattacharyya, and the facet analysis and chain indexing system of Ranganathan. (AEF)
Use of speech-to-text technology for documentation by healthcare providers.
Ajami, Sima
2016-01-01
Medical records are a critical component of a patient's treatment. However, documentation of patient-related information is considered a secondary activity in the provision of healthcare services, often leading to incomplete medical records and patient data of low quality. Advances in information technology (IT) in the health system and registration of information in electronic health records (EHR) using speechto- text conversion software have facilitated service delivery. This narrative review is a literature search with the help of libraries, books, conference proceedings, databases of Science Direct, PubMed, Proquest, Springer, SID (Scientific Information Database), and search engines such as Yahoo, and Google. I used the following keywords and their combinations: speech recognition, automatic report documentation, voice to text software, healthcare, information, and voice recognition. Due to lack of knowledge of other languages, I searched all texts in English or Persian with no time limits. Of a total of 70, only 42 articles were selected. Speech-to-text conversion technology offers opportunities to improve the documentation process of medical records, reduce cost and time of recording information, enhance the quality of documentation, improve the quality of services provided to patients, and support healthcare providers in legal matters. Healthcare providers should recognize the impact of this technology on service delivery.
Exploring Contextual Models in Chemical Patent Search
NASA Astrophysics Data System (ADS)
Urbain, Jay; Frieder, Ophir
We explore the development of probabilistic retrieval models for integrating term statistics with entity search using multiple levels of document context to improve the performance of chemical patent search. A distributed indexing model was developed to enable efficient named entity search and aggregation of term statistics at multiple levels of patent structure including individual words, sentences, claims, descriptions, abstracts, and titles. The system can be scaled to an arbitrary number of compute instances in a cloud computing environment to support concurrent indexing and query processing operations on large patent collections.
A hypertext system that learns from user feedback
NASA Technical Reports Server (NTRS)
Mathe, Nathalie
1994-01-01
Retrieving specific information from large amounts of documentation is not an easy task. It could be facilitated if information relevant in the current problem solving context could be automatically supplied to the user. As a first step towards this goal, we have developed an intelligent hypertext system called CID (Computer Integrated Documentation). Besides providing an hypertext interface for browsing large documents, the CID system automatically acquires and reuses the context in which previous searches were appropriate. This mechanism utilizes on-line user information requirements and relevance feedback either to reinforce current indexing in case of success or to generate new knowledge in case of failure. Thus, the user continually augments and refines the intelligence of the retrieval system. This allows the CID system to provide helpful responses, based on previous usage of the documentation, and to improve its performance over time. We successfully tested the CID system with users of the Space Station Freedom requirements documents. We are currently extending CID to other application domains (Space Shuttle operations documents, airplane maintenance manuals, and on-line training). We are also exploring the potential commercialization of this technique.
ERIC Educational Resources Information Center
Lynch, Michael F.; Willett, Peter
1987-01-01
Discusses research into chemical information and document retrieval systems at the University of Sheffield. Highlights include the use of cluster analysis methods for document retrieval and drug design, representation and searching of files of generic chemical structures, and the application of parallel computer hardware to information retrieval.…
Code of Federal Regulations, 2013 CFR
2013-01-01
... direct costs associated with any response it has prepared. (5) If fees for document search are authorized... searching for documents and other direct costs of a search, even if a search fails to locate records or if records located are determined to be exempt from disclosure. Searches should be conducted in the most...
Code of Federal Regulations, 2011 CFR
2011-01-01
... direct costs associated with any response it has prepared. (5) If fees for document search are authorized... searching for documents and other direct costs of a search, even if a search fails to locate records or if records located are determined to be exempt from disclosure. Searches should be conducted in the most...
Code of Federal Regulations, 2012 CFR
2012-01-01
... direct costs associated with any response it has prepared. (5) If fees for document search are authorized... searching for documents and other direct costs of a search, even if a search fails to locate records or if records located are determined to be exempt from disclosure. Searches should be conducted in the most...
Code of Federal Regulations, 2014 CFR
2014-01-01
... direct costs associated with any response it has prepared. (5) If fees for document search are authorized... searching for documents and other direct costs of a search, even if a search fails to locate records or if records located are determined to be exempt from disclosure. Searches should be conducted in the most...
Search and selection methodology of systematic reviews in orthodontics (2000-2004).
Flores-Mir, Carlos; Major, Michael P; Major, Paul W
2006-08-01
More systematic reviews related to orthodontic topics are published each year, although little has been done to evaluate their search and selection methodologies. Systematic reviews related to orthodontics published between January 1, 2000, and December 31, 2004, were searched for their use of multiple electronic databases and secondary searches. The search and selection methods of identified systematic reviews were evaluated against the Cochrane Handbook's guidelines. Sixteen orthodontic systematic reviews were identified in this period. The percentage of reviews documenting and using each criterion of article searching has changed over the last 5 years, with no recognizable directional trend. On average, most systematic reviews documented their electronic search terms (88%) and inclusion-exclusion criteria (100%), and used secondary searching (75%). Many still failed to search more than MEDLINE (56%), failed to document the database names and search dates (37%), failed to document the search strategy (62%), did not use several reviewers for selecting studies (75%), and did not include all languages (81%). The methodology of systematic reviews in orthodontics is still limited, with key methodological components frequently absent or not appropriately described.
Tautomerism in chemical information management systems
NASA Astrophysics Data System (ADS)
Warr, Wendy A.
2010-06-01
Tautomerism has an impact on many of the processes in chemical information management systems including novelty checking during registration into chemical structure databases; storage of structures; exact and substructure searching in chemical structure databases; and depiction of structures retrieved by a search. The approaches taken by 27 different software vendors and database producers are compared. It is hoped that this comparison will act as a discussion document that could ultimately improve databases and software for researchers in the future.
EquiX-A Search and Query Language for XML.
ERIC Educational Resources Information Center
Cohen, Sara; Kanza, Yaron; Kogan, Yakov; Sagiv, Yehoshua; Nutt, Werner; Serebrenik, Alexander
2002-01-01
Describes EquiX, a search language for XML that combines querying with searching to query the data and the meta-data content of Web pages. Topics include search engines; a data model for XML documents; search query syntax; search query semantics; an algorithm for evaluating a query on a document; and indexing EquiX queries. (LRW)
Developing topic-specific search filters for PubMed with click-through data.
Li, J; Lu, Z
2013-01-01
Search filters have been developed and demonstrated for better information access to the immense and ever-growing body of publications in the biomedical domain. However, to date the number of filters remains quite limited because the current filter development methods require significant human efforts in manual document review and filter term selection. In this regard, we aim to investigate automatic methods for generating search filters. We present an automated method to develop topic-specific filters on the basis of users' search logs in PubMed. Specifically, for a given topic, we first detect its relevant user queries and then include their corresponding clicked articles to serve as the topic-relevant document set accordingly. Next, we statistically identify informative terms that best represent the topic-relevant document set using a background set composed of topic irrelevant articles. Lastly, the selected representative terms are combined with Boolean operators and evaluated on benchmark datasets to derive the final filter with the best performance. We applied our method to develop filters for four clinical topics: nephrology, diabetes, pregnancy, and depression. For the nephrology filter, our method obtained performance comparable to the state of the art (sensitivity of 91.3%, specificity of 98.7%, precision of 94.6%, and accuracy of 97.2%). Similarly, high-performing results (over 90% in all measures) were obtained for the other three search filters. Based on PubMed click-through data, we successfully developed a high-performance method for generating topic-specific search filters that is significantly more efficient than existing manual methods. All data sets (topic-relevant and irrelevant document sets) used in this study and a demonstration system are publicly available at http://www.ncbi.nlm.nih.gov/CBBresearch/Lu/downloads/CQ_filter/
Developing Topic-Specific Search Filters for PubMed with Click-Through Data
Li, Jiao; Lu, Zhiyong
2013-01-01
Summary Objectives Search filters have been developed and demonstrated for better information access to the immense and ever-growing body of publications in the biomedical domain. However, to date the number of filters remains quite limited because the current filter development methods require significant human efforts in manual document review and filter term selection. In this regard, we aim to investigate automatic methods for generating search filters. Methods We present an automated method to develop topic-specific filters on the basis of users’ search logs in PubMed. Specifically, for a given topic, we first detect its relevant user queries and then include their corresponding clicked articles to serve as the topic-relevant document set accordingly. Next, we statistically identify informative terms that best represent the topic-relevant document set using a background set composed of topic irrelevant articles. Lastly, the selected representative terms are combined with Boolean operators and evaluated on benchmark datasets to derive the final filter with the best performance. Results We applied our method to develop filters for four clinical topics: nephrology, diabetes, pregnancy, and depression. For the nephrology filter, our method obtained performance comparable to the state of the art (sensitivity of 91.3%, specificity of 98.7%, precision of 94.6%, and accuracy of 97.2%). Similarly, high-performing results (over 90% in all measures) were obtained for the other three search filters. Conclusion Based on PubMed click-through data, we successfully developed a high-performance method for generating topic-specific search filters that is significantly more efficient than existing manual methods. All data sets (topic-relevant and irrelevant document sets) used in this study and a demonstration system are publicly available at http://www.ncbi.nlm.nih.gov/CBBresearch/Lu/downloads/CQ_filter/ PMID:23666447
Methods and means used in programming intelligent searches of technical documents
NASA Technical Reports Server (NTRS)
Gross, David L.
1993-01-01
In order to meet the data research requirements of the Safety, Reliability & Quality Assurance activities at Kennedy Space Center (KSC), a new computer search method for technical data documents was developed. By their very nature, technical documents are partially encrypted because of the author's use of acronyms, abbreviations, and shortcut notations. This problem of computerized searching is compounded at KSC by the volume of documentation that is produced during normal Space Shuttle operations. The Centralized Document Database (CDD) is designed to solve this problem. It provides a common interface to an unlimited number of files of various sizes, with the capability to perform any diversified types and levels of data searches. The heart of the CDD is the nature and capability of its search algorithms. The most complex form of search that the program uses is with the use of a domain-specific database of acronyms, abbreviations, synonyms, and word frequency tables. This database, along with basic sentence parsing, is used to convert a request for information into a relational network. This network is used as a filter on the original document file to determine the most likely locations for the data requested. This type of search will locate information that traditional techniques, (i.e., Boolean structured key-word searching), would not find.
with Search To search for a document, type a few descriptive words in the search box, and press the Enter key or click the search button. A results page appears with a list of documents and web pages that are related to your search terms, with the most relevant search results appearing at the top of the
NASA Technical Reports Server (NTRS)
Ambur, Manjula Y.; Adams, David L.; Trinidad, P. Paul
1997-01-01
NASA Langley Technical Library has been involved in developing systems for full-text information delivery of NACA/NASA technical reports since 1991. This paper will describe the two prototypes it has developed and the present production system configuration. The prototype systems are a NACA CD-ROM of thirty-three classic paper NACA reports and a network-based Full-text Electronic Reports Documents System (FEDS) constructed from both paper and electronic formats of NACA and NASA reports. The production system is the DigiDoc System (DIGItal Documents) presently being developed based on the experiences gained from the two prototypes. DigiDoc configuration integrates the on-line catalog database World Wide Web interface and PDF technology to provide a powerful and flexible search and retrieval system. It describes in detail significant achievements and lessons learned in terms of data conversion, storage technologies, full-text searching and retrieval, and image databases. The conclusions from the experiences of digitization and full- text access and future plans for DigiDoc system implementation are discussed.
Using Concept Relations to Improve Ranking in Information Retrieval
Price, Susan L.; Delcambre, Lois M.
2005-01-01
Despite improved search engine technology, most searches return numerous documents not directly related to the query. This problem is mitigated if relevant documents appear high on a ranked list of search results. We propose that some queries and the underlying information needs can be modeled as relationships between concepts (relations), and we match relations in queries to relations in documents to try to improve ranking of search results. We investigate four techniques to identify two relationships important in medicine, causes and treats, to improve the ranking of medical text documents relevant to clinical questions about causation and treatment. Preliminary results suggest that identifying relation instances can improve the ranking of search results. PMID:16779114
Features: Real-Time Adaptive Feature and Document Learning for Web Search.
ERIC Educational Resources Information Center
Chen, Zhixiang; Meng, Xiannong; Fowler, Richard H.; Zhu, Binhai
2001-01-01
Describes Features, an intelligent Web search engine that is able to perform real-time adaptive feature (i.e., keyword) and document learning. Explains how Features learns from users' document relevance feedback and automatically extracts and suggests indexing keywords relevant to a search query, and learns from users' keyword relevance feedback…
ERIC Educational Resources Information Center
Wilder, Dolores J., Comp.; Hines, Rella, Comp.
The Tennessee Research Coordinating Unit (RCU) has implemented a computerized information retrieval system known as "Query," which allows for the retrieval of documents indexed in Research in Education (RIE), Current Index to Journals in Education (CIJE), and Abstracts of Instructional and Research Materials (AIM/ARM). The document…
Federal Register 2010, 2011, 2012, 2013, 2014
2005-11-16
... Reference System (TRS) [see http://www.epa.gov/trs ] in order to better support future semantic Web needs... creation of glossaries for Web pages and documents, a common vocabulary for search engines, and in the...
System of HPC content archiving
NASA Astrophysics Data System (ADS)
Bogdanov, A.; Ivashchenko, A.
2017-12-01
This work is aimed to develop a system, that will effectively solve the problem of storing and analyzing files containing text data, by using modern software development tools, techniques and approaches. The main challenge of storing a large number of text documents defined at the problem formulation stage, have to be resolved with such functionality as full text search and document clustering depends on their contents. Main system features could be described with notions of distributed multilevel architecture, flexibility and interchangeability of components, achieved through the standard functionality incapsulation in independent executable modules.
Poster — Thur Eve — 52: A Web-based Platform for Collaborative Document Management in Radiotherapy
DOE Office of Scientific and Technical Information (OSTI.GOV)
Kildea, J.; Joseph, A.
We describe DepDocs, a web-based platform that we have developed to manage the committee meetings, policies, procedures and other documents within our otherwise paperless radiotherapy clinic. DepDocs is essentially a document management system based on the popular Drupal content management software. For security and confidentiality, it is hosted on a linux server internal to our hospital network such that documents are never sent to the cloud or outside of the hospital firewall. We used Drupal's in-built role-based user rights management system to assign a role, and associated document editing rights, to each user. Documents are accessed for viewing using eithermore » a simple Google-like search or by generating a list of related documents from a taxonomy of categorization terms. Our system provides document revision tracking and an document review and approval mechanism for all official policies and procedures. Committee meeting schedules, agendas and minutes are maintained by committee chairs and are restricted to committee members. DepDocs has been operational within our department for over six months and has already 45 unique users and an archive of over 1000 documents, mostly policies and procedures. Documents are easily retrievable from the system using any web browser within our hospital's network.« less
The Earth System Documentation (ES-DOC) project
NASA Astrophysics Data System (ADS)
Murphy, S.; Greenslade, M. A.; Treshansky, A.; DeLuca, C.; Guilyardi, E.; Denvil, S.
2013-12-01
Earth System Documentation (ES-DOC) is an international project supplying high quality tools and services in support of Earth system documentation creation, analysis and dissemination. It is nurturing a sustainable standards based documentation ecosystem that aims to become an integral part of the next generation of exa-scale dataset archives. ES-DOC leverages open source software, and applies a software development methodology that places end-user narratives at the heart of all it does. ES-DOC has initially focused upon nurturing the Earth System Model (ESM) documentation eco-system. Within this context ES-DOC leverages the emerging Common Information Model (CIM) metadata standard, which has supported the following projects: ** Coupled Model Inter-comparison Project Phase 5 (CMIP5); ** Dynamical Core Model Inter-comparison Project (DCMIP-2012); ** National Climate Predictions and Projections Platforms (NCPP) Quantitative Evaluation of Downscaling Workshop (QED-2013). This presentation will introduce the project to a wider audience and will demonstrate the current production level capabilities of the eco-system: ** An ESM documentation Viewer embeddable into any website; ** An ESM Questionnaire configurable on a project by project basis; ** An ESM comparison tool reusable across projects; ** An ESM visualization tool reusable across projects; ** A search engine for speedily accessing published documentation; ** Libraries for streamlining document creation, validation and publishing pipelines.
NASA Astrophysics Data System (ADS)
Brauer, U.
2007-08-01
The Open Navigator Framework (ONF) was developed to provide a unified and scalable platform for user interface integration. The main objective for the framework was to raise usability of monitoring and control consoles and to provide a reuse of software components in different application areas. ONF is currently applied for the Columbus onboard crew interface, the commanding application for the Columbus Control Centre, the Columbus user facilities specialized user interfaces, the Mission Execution Crew Assistant (MECA) study and EADS Astrium internal R&D projects. ONF provides a well documented and proven middleware for GUI components (Java plugin interface, simplified concept similar to Eclipse). The overall application configuration is performed within a graphical user interface for layout and component selection. The end-user does not have to work in the underlying XML configuration files. ONF was optimized to provide harmonized user interfaces for monitoring and command consoles. It provides many convenience functions designed together with flight controllers and onboard crew: user defined workspaces, incl. support for multi screens efficient communication mechanism between the components integrated web browsing and documentation search &viewing consistent and integrated menus and shortcuts common logging and application configuration (properties) supervision interface for remote plugin GUI access (web based) A large number of operationally proven ONF components have been developed: Command Stack & History: Release of commands and follow up the command acknowledges System Message Panel: Browse, filter and search system messages/events Unified Synoptic System: Generic synoptic display system Situational Awareness : Show overall subsystem status based on monitoring of key parameters System Model Browser: Browse mission database defintions (measurements, commands, events) Flight Procedure Executor: Execute checklist and logical flow interactive procedures Web Browser : Integrated browser reference documentation and operations data Timeline Viewer: View master timeline as Gantt chart Search: Local search of operations products (e.g. documentation, procedures, displays) All GUI components access the underlying spacecraft data (commanding, reporting data, events, command history) via a common library providing adaptors for the current deployments (Columbus MCS, Columbus onboard Data Management System, Columbus Trainer raw packet protocol). New Adaptors are easy to develop. Currently an adaptor to SCOS 2000 is developed as part of a study for the ESTEC standardization section ("USS for ESTEC Reference Facility").
US photovoltaic patents: 1991-1993
NASA Astrophysics Data System (ADS)
Pohle, L.
1995-03-01
This document contains US patents on terrestrial photovoltaic (PV) power applications, including systems, components, and materials as well as manufacturing and support functions. The patent entries in this document were issued from 1991 to 1993. The entries were located by searching USPA, the database of the US Patent Office. The final search retrieved all patents under the class 'Batteries, Thermoelectric and Photoelectric' and the subclasses 'Photoelectric,' 'Testing,' and 'Applications.' The search also located patents that contained the words 'photovoltaic(s)' or 'solar cell(s)' and their derivatives. After the initial list was compiled, most of the patents on the following subjects were excluded: space photovoltaic technology, use of the photovoltaic effect for detectors, and subjects only peripherally concerned with photovoltaic. Some patents on these three subjects were included when ft appeared that those inventions might be of use in terrestrial PV power technologies.
CINTEX: International Interoperability Extensions to EOSDIS
NASA Technical Reports Server (NTRS)
Graves, Sara J.
1997-01-01
A large part of the research under this cooperative agreement involved working with representatives of the DLR, NASDA, EDC, and NOAA-SAA data centers to propose a set of enhancements and additions to the EOSDIS Version 0 Information Management System (V0 IMS) Client/Server Message Protocol. Helen Conover of ITSL led this effort to provide for an additional geographic search specification (WRS Path/Row), data set- and data center-specific search criteria, search by granule ID, specification of data granule subsetting requests, data set-based ordering, and the addition of URLs to result messages. The V0 IMS Server Cookbook is an evolving document, providing resources and information to data centers setting up a VO IMS Server. Under this Cooperative Agreement, Helen Conover revised, reorganized, and expanded this document, and converted it to HTML. Ms. Conover has also worked extensively with the IRE RAS data center, CPSSI, in Russia. She served as the primary IMS contact for IRE-CPSSI and as IRE-CPSSI's liaison to other members of IMS and Web Gateway (WG) development teams. Her documentation of IMS problems in the IRE environment (Sun servers and low network bandwidth) led to a general restructuring of the V0 IMS Client message polling system. to the benefit of all IMS participants. In addition to the IMS server software and documentation. which are generally available to CINTEX sites, Ms. Conover also provided database design documentation and consulting, order tracking software, and hands-on testing and debug assistance to IRE. In the final pre-operational phase of IRE-CPSSI development, she also supplied information on configuration management, including ideas and processes in place at the Global Hydrology Resource Center (GHRC), an EOSDIS data center operated by ITSL.
Unified System Of Data On Materials And Processes
NASA Technical Reports Server (NTRS)
Key, Carlo F.
1989-01-01
Wide-ranging sets of data for aerospace industry described. Document describes Materials and Processes Technical Information System (MAPTIS), computerized set of integrated data bases for use by NASA and aerospace industry. Stores information in standard format for fast retrieval in searches and surveys of data. Helps engineers select materials and verify their properties. Promotes standardized nomenclature as well as standarized tests and presentation of data. Format of document of photographic projection slides used in lectures. Presents examples of reports from various data bases.
Method for gathering and summarizing internet information
Potok, Thomas E.; Elmore, Mark Thomas; Reed, Joel Wesley; Treadwell, Jim N.; Samatova, Nagiza Faridovna
2010-04-06
A computer method of gathering and summarizing large amounts of information comprises collecting information from a plurality of information sources (14, 51) according to respective maps (52) of the information sources (14), converting the collected information from a storage format to XML-language documents (26, 53) and storing the XML-language documents in a storage medium, searching for documents (55) according to a search query (13) having at least one term and identifying the documents (26) found in the search, and displaying the documents as nodes (33) of a tree structure (32) having links (34) and nodes (33) so as to indicate similarity of the documents to each other.
Method for gathering and summarizing internet information
Potok, Thomas E [Oak Ridge, TN; Elmore, Mark Thomas [Oak Ridge, TN; Reed, Joel Wesley [Knoxville, TN; Treadwell, Jim N [Louisville, TN; Samatova, Nagiza Faridovna [Oak Ridge, TN
2008-01-01
A computer method of gathering and summarizing large amounts of information comprises collecting information from a plurality of information sources (14, 51) according to respective maps (52) of the information sources (14), converting the collected information from a storage format to XML-language documents (26, 53) and storing the XML-language documents in a storage medium, searching for documents (55) according to a search query (13) having at least one term and identifying the documents (26) found in the search, and displaying the documents as nodes (33) of a tree structure (32) having links (34) and nodes (33) so as to indicate similarity of the documents to each other.
Querying archetype-based EHRs by search ontology-based XPath engineering.
Kropf, Stefan; Uciteli, Alexandr; Schierle, Katrin; Krücken, Peter; Denecke, Kerstin; Herre, Heinrich
2018-05-11
Legacy data and new structured data can be stored in a standardized format as XML-based EHRs on XML databases. Querying documents on these databases is crucial for answering research questions. Instead of using free text searches, that lead to false positive results, the precision can be increased by constraining the search to certain parts of documents. A search ontology-based specification of queries on XML documents defines search concepts and relates them to parts in the XML document structure. Such query specification method is practically introduced and evaluated by applying concrete research questions formulated in natural language on a data collection for information retrieval purposes. The search is performed by search ontology-based XPath engineering that reuses ontologies and XML-related W3C standards. The key result is that the specification of research questions can be supported by the usage of search ontology-based XPath engineering. A deeper recognition of entities and a semantic understanding of the content is necessary for a further improvement of precision and recall. Key limitation is that the application of the introduced process requires skills in ontology and software development. In future, the time consuming ontology development could be overcome by implementing a new clinical role: the clinical ontologist. The introduced Search Ontology XML extension connects Search Terms to certain parts in XML documents and enables an ontology-based definition of queries. Search ontology-based XPath engineering can support research question answering by the specification of complex XPath expressions without deep syntax knowledge about XPaths.
ERIC Educational Resources Information Center
Buckland, Lawrence F.; Madden, Mary
From experimental work performed, and reported upon in this document, it is concluded that converting the New York State Library (NYSL) shelf list sample to machine readable form, and searching this shelf list using a remote access catalog are technically sound concepts though the capital costs of data conversion and system installation will be…
77 FR 76012 - Privacy Act of 1974; System of Records
Federal Register 2010, 2011, 2012, 2013, 2014
2012-12-26
...) deletes eight systems of records from its existing inventory of systems of records subject to the Privacy... inventory of record systems subject to the Privacy Act (5 U.S.C. 552a). The deletions are not within the... documents of the Department published in the Federal Register by using the article search feature at: www...
Application Agreement and Integration Services
NASA Technical Reports Server (NTRS)
Driscoll, Kevin R.; Hall, Brendan; Schweiker, Kevin
2013-01-01
Application agreement and integration services are required by distributed, fault-tolerant, safety critical systems to assure required performance. An analysis of distributed and hierarchical agreement strategies are developed against the backdrop of observed agreement failures in fielded systems. The documented work was performed under NASA Task Order NNL10AB32T, Validation And Verification of Safety-Critical Integrated Distributed Systems Area 2. This document is intended to satisfy the requirements for deliverable 5.2.11 under Task 4.2.2.3. This report discusses the challenges of maintaining application agreement and integration services. A literature search is presented that documents previous work in the area of replica determinism. Sources of non-deterministic behavior are identified and examples are presented where system level agreement failed to be achieved. We then explore how TTEthernet services can be extended to supply some interesting application agreement frameworks. This document assumes that the reader is familiar with the TTEthernet protocol. The reader is advised to read the TTEthernet protocol standard [1] before reading this document. This document does not re-iterate the content of the standard.
Gupta, Amarnath; Bug, William; Marenco, Luis; Qian, Xufei; Condit, Christopher; Rangarajan, Arun; Müller, Hans Michael; Miller, Perry L.; Sanders, Brian; Grethe, Jeffrey S.; Astakhov, Vadim; Shepherd, Gordon; Sternberg, Paul W.; Martone, Maryann E.
2009-01-01
The overarching goal of the NIF (Neuroscience Information Framework) project is to be a one-stop-shop for Neuroscience. This paper provides a technical overview of how the system is designed. The technical goal of the first version of the NIF system was to develop an information system that a neuroscientist can use to locate relevant information from a wide variety of information sources by simple keyword queries. Although the user would provide only keywords to retrieve information, the NIF system is designed to treat them as concepts whose meanings are interpreted by the system. Thus, a search for term should find a record containing synonyms of the term. The system is targeted to find information from web pages, publications, databases, web sites built upon databases, XML documents and any other modality in which such information may be published. We have designed a system to achieve this functionality. A central element in the system is an ontology called NIFSTD (for NIF Standard) constructed by amalgamating a number of known and newly developed ontologies. NIFSTD is used by our ontology management module, called OntoQuest to perform ontology-based search over data sources. The NIF architecture currently provides three different mechanisms for searching heterogeneous data sources including relational databases, web sites, XML documents and full text of publications. Version 1.0 of the NIF system is currently in beta test and may be accessed through http://nif.nih.gov. PMID:18958629
Gupta, Amarnath; Bug, William; Marenco, Luis; Qian, Xufei; Condit, Christopher; Rangarajan, Arun; Müller, Hans Michael; Miller, Perry L; Sanders, Brian; Grethe, Jeffrey S; Astakhov, Vadim; Shepherd, Gordon; Sternberg, Paul W; Martone, Maryann E
2008-09-01
The overarching goal of the NIF (Neuroscience Information Framework) project is to be a one-stop-shop for Neuroscience. This paper provides a technical overview of how the system is designed. The technical goal of the first version of the NIF system was to develop an information system that a neuroscientist can use to locate relevant information from a wide variety of information sources by simple keyword queries. Although the user would provide only keywords to retrieve information, the NIF system is designed to treat them as concepts whose meanings are interpreted by the system. Thus, a search for term should find a record containing synonyms of the term. The system is targeted to find information from web pages, publications, databases, web sites built upon databases, XML documents and any other modality in which such information may be published. We have designed a system to achieve this functionality. A central element in the system is an ontology called NIFSTD (for NIF Standard) constructed by amalgamating a number of known and newly developed ontologies. NIFSTD is used by our ontology management module, called OntoQuest to perform ontology-based search over data sources. The NIF architecture currently provides three different mechanisms for searching heterogeneous data sources including relational databases, web sites, XML documents and full text of publications. Version 1.0 of the NIF system is currently in beta test and may be accessed through http://nif.nih.gov.
GENESIS: GPS Environmental and Earth Science Information System
NASA Technical Reports Server (NTRS)
Hajj, George
1999-01-01
This presentation reviews the GPS ENvironmental and Earth Science Information System (GENESIS). The objectives of GENESIS are outlined (1) Data Archiving, searching and distribution for science data products derived from Space borne TurboRogue Space Receivers for GPS science and other ground based GPS receivers, (2) Data browsing using integrated visualization tools, (3) Interactive web/java-based data search and retrieval, (4) Data subscription service, (5) Data migration from existing GPS archived data, (6) On-line help and documentation, and (7) participation in the WP-ESIP federation. The presentation reviews the products and services of Genesis, and the technology behind the system.
ERIC Educational Resources Information Center
Levy, David M.; Huttenlocher, Dan; Moll, Angela; Smith, MacKenzie; Hodge, Gail M.; Chandler, Adam; Foley, Dan; Hafez, Alaaeldin M.; Redalen, Aaron; Miller, Naomi
2000-01-01
Includes six articles focusing on the purpose of digital public libraries; encoding electronic documents through compression techniques; a distributed finding aid server; digital archiving practices in the framework of information life cycle management; converting metadata into MARC format and Dublin Core formats; and evaluating Web sites through…
Document Clustering Approach for Meta Search Engine
NASA Astrophysics Data System (ADS)
Kumar, Naresh, Dr.
2017-08-01
The size of WWW is growing exponentially with ever change in technology. This results in huge amount of information with long list of URLs. Manually it is not possible to visit each page individually. So, if the page ranking algorithms are used properly then user search space can be restricted up to some pages of searched results. But available literatures show that no single search system can provide qualitative results from all the domains. This paper provides solution to this problem by introducing a new meta search engine that determine the relevancy of query corresponding to web page and cluster the results accordingly. The proposed approach reduces the user efforts, improves the quality of results and performance of the meta search engine.
Boyer, C; Baujard, V; Scherrer, J R
2001-01-01
Any new user to the Internet will think that to retrieve the relevant document is an easy task especially with the wealth of sources available on this medium, but this is not the case. Even experienced users have difficulty formulating the right query for making the most of a search tool in order to efficiently obtain an accurate result. The goal of this work is to reduce the time and the energy necessary in searching and locating medical and health information. To reach this goal we have developed HONselect [1]. The aim of HONselect is not only to improve efficiency in retrieving documents but to respond to an increased need for obtaining a selection of relevant and accurate documents from a breadth of various knowledge databases including scientific bibliographical references, clinical trials, daily news, multimedia illustrations, conferences, forum, Web sites, clinical cases, and others. The authors based their approach on the knowledge representation using the National Library of Medicine's Medical Subject Headings (NLM, MeSH) vocabulary and classification [2,3]. The innovation is to propose a multilingual "one-stop searching" (one Web interface to databases currently in English, French and German) with full navigational and connectivity capabilities. The user may choose from a given selection of related terms the one that best suit his search, navigate in the term's hierarchical tree, and access directly to a selection of documents from high quality knowledge suppliers such as the MEDLINE database, the NLM's ClinicalTrials.gov server, the NewsPage's daily news, the HON's media gallery, conference listings and MedHunt's Web sites [4, 5, 6, 7, 8, 9]. HONselect, developed by HON, a non-profit organisation [10], is a free online available multilingual tool based on the MeSH thesaurus to index, select, retrieve and display accurate, up to date, high-level and quality documents.
Investigative change detection: identifying new topics using lexicon-based search
NASA Astrophysics Data System (ADS)
Hintz, Kenneth J.
2002-08-01
In law enforcement there is much textual data which needs to be searched in order to detect new threats. A new methodology which can be applied to this need is the automatic searching of the contents of documents from known sources to construct a lexicon of words used by that source. When analyzing future documents, the occurrence of words which have not been lexiconized are indicative of the introduction of a new topic into the source's lexicon which should be examined in its context by an analyst. A system analogous to this has been built and used to detect Fads and Categories on web sites. Fad refers to the first appearance of a word not in the lexicon; Category refers to the repeated appearance of a Fad word and the exceeding of some frequency or spatial occurrence metric indicating a permanence to the Category.
US photovoltaic patents: 1991--1993
DOE Office of Scientific and Technical Information (OSTI.GOV)
Pohle, L
1995-03-01
This document contains US patents on terrestrial photovoltaic (PV) power applications, including systems, components, and materials as well as manufacturing and support functions. The patent entries in this document were issued from 1991 to 1993. The entries were located by searching USPA, the database of the US Patent Office. The final search retrieved all patents under the class ``Batteries, Thermoelectric and Photoelectric`` and the subclasses ``Photoelectric,`` ``Testing,`` and ``Applications.`` The search also located patents that contained the words ``photovoltaic(s)`` or ``solar cell(s)`` and their derivatives. After the initial list was compiled, most of the patents on the following subjects weremore » excluded: space photovoltaic technology, use of the photovoltaic effect for detectors, and subjects only peripherally concerned with photovoltaic. Some patents on these three subjects were included when ft appeared that those inventions might be of use in terrestrial PV power technologies.« less
A Nugget-Based Test Collection Construction Paradigm
ERIC Educational Resources Information Center
Rajput, Shahzad K.
2012-01-01
The problem of building test collections is central to the development of information retrieval systems such as search engines. The primary use of test collections is the evaluation of IR systems. The widely employed "Cranfield paradigm" dictates that the information relevant to a topic be encoded at the level of documents, therefore…
Silicon Graphics' IRIS InSight: An SGML Success Story.
ERIC Educational Resources Information Center
Glushko, Robert J.; Kershner, Ken
1993-01-01
Offers a case history of the development of the Silicon Graphics "IRIS InSight" system, a system for viewing on-line documentation using Standard Generalized Markup Language. Notes that SGML's explicit encoding of structure and separation of structure and presentation make possible structure-based search, alternative structural views of…
Informed Consent: Does Anyone Really Understand What Is Contained In The Medical Record?
Fenton, S H; Manion, F; Hsieh, K; Harris, M
2015-01-01
Despite efforts to provide standard definitions of terms such as "medical record", "computer-based patient record", "electronic medical record" and "electronic health record", the terms are still used interchangeably. Initiatives like data and information governance, research biorepositories, and learning health systems require availability and reuse of data, as well as common understandings of the scope for specific purposes. Lacking widely shared definitions, utilization of the afore-mentioned terms in research informed consent documents calls to question whether all participants in the research process - patients, information technology and regulatory staff, and the investigative team - fully understand what data and information they are asking to obtain and agreeing to share. This descriptive study explored the terminology used in research informed consent documents when describing patient data and information, asking the question "Does the use of the term "medical record" in the context of a research informed consent document accurately represent the scope of the data involved?" Informed consent document templates found on 17 Institutional Review Board (IRB) websites with Clinical and Translational Science Awards (CTSA) were searched for terms that appeared to be describing the data resources to be accessed. The National Library of Medicine's (NLM) Terminology Services was searched for definitions provided by key standards groups that deposit terminologies with the NLM. The results suggest research consent documents are using outdated terms to describe patient information, health care terminology systems need to consider the context of research for use cases, and that there is significant work to be done to assure the HIPAA Omnibus Rule is applied to contemporary activities such as biorepositories and learning health systems. "Medical record", a term used extensively in research informed consent documents, is ambiguous and does not serve us well in the context of contemporary information management and governance.
Federal Register 2010, 2011, 2012, 2013, 2014
2012-11-14
..., timely account disablement/deletion, configuration management, maintenance, system and information... paper or CD-ROM submissions): Division of Dockets Management (HFA-305), Food and Drug Administration... document, into the ``Search'' box and follow the prompts and/or go to the Division of Dockets Management...
Federal Register 2010, 2011, 2012, 2013, 2014
2012-12-31
... development of accessible information technology including cloud computing, software, systems, and devices... documents of the Department published in the Federal Register by using the article search feature at: www...
Text Mining the History of Medicine.
Thompson, Paul; Batista-Navarro, Riza Theresa; Kontonatsios, Georgios; Carter, Jacob; Toon, Elizabeth; McNaught, John; Timmermann, Carsten; Worboys, Michael; Ananiadou, Sophia
2016-01-01
Historical text archives constitute a rich and diverse source of information, which is becoming increasingly readily accessible, due to large-scale digitisation efforts. However, it can be difficult for researchers to explore and search such large volumes of data in an efficient manner. Text mining (TM) methods can help, through their ability to recognise various types of semantic information automatically, e.g., instances of concepts (places, medical conditions, drugs, etc.), synonyms/variant forms of concepts, and relationships holding between concepts (which drugs are used to treat which medical conditions, etc.). TM analysis allows search systems to incorporate functionality such as automatic suggestions of synonyms of user-entered query terms, exploration of different concepts mentioned within search results or isolation of documents in which concepts are related in specific ways. However, applying TM methods to historical text can be challenging, according to differences and evolutions in vocabulary, terminology, language structure and style, compared to more modern text. In this article, we present our efforts to overcome the various challenges faced in the semantic analysis of published historical medical text dating back to the mid 19th century. Firstly, we used evidence from diverse historical medical documents from different periods to develop new resources that provide accounts of the multiple, evolving ways in which concepts, their variants and relationships amongst them may be expressed. These resources were employed to support the development of a modular processing pipeline of TM tools for the robust detection of semantic information in historical medical documents with varying characteristics. We applied the pipeline to two large-scale medical document archives covering wide temporal ranges as the basis for the development of a publicly accessible semantically-oriented search system. The novel resources are available for research purposes, while the processing pipeline and its modules may be used and configured within the Argo TM platform.
Text Mining the History of Medicine
Thompson, Paul; Batista-Navarro, Riza Theresa; Kontonatsios, Georgios; Carter, Jacob; Toon, Elizabeth; McNaught, John; Timmermann, Carsten; Worboys, Michael; Ananiadou, Sophia
2016-01-01
Historical text archives constitute a rich and diverse source of information, which is becoming increasingly readily accessible, due to large-scale digitisation efforts. However, it can be difficult for researchers to explore and search such large volumes of data in an efficient manner. Text mining (TM) methods can help, through their ability to recognise various types of semantic information automatically, e.g., instances of concepts (places, medical conditions, drugs, etc.), synonyms/variant forms of concepts, and relationships holding between concepts (which drugs are used to treat which medical conditions, etc.). TM analysis allows search systems to incorporate functionality such as automatic suggestions of synonyms of user-entered query terms, exploration of different concepts mentioned within search results or isolation of documents in which concepts are related in specific ways. However, applying TM methods to historical text can be challenging, according to differences and evolutions in vocabulary, terminology, language structure and style, compared to more modern text. In this article, we present our efforts to overcome the various challenges faced in the semantic analysis of published historical medical text dating back to the mid 19th century. Firstly, we used evidence from diverse historical medical documents from different periods to develop new resources that provide accounts of the multiple, evolving ways in which concepts, their variants and relationships amongst them may be expressed. These resources were employed to support the development of a modular processing pipeline of TM tools for the robust detection of semantic information in historical medical documents with varying characteristics. We applied the pipeline to two large-scale medical document archives covering wide temporal ranges as the basis for the development of a publicly accessible semantically-oriented search system. The novel resources are available for research purposes, while the processing pipeline and its modules may be used and configured within the Argo TM platform. PMID:26734936
A Spoken English Recognition Expert System.
1983-09-01
Davidson. "Representation of Knowledge," Handbook of Artificial Intelligence, edited by Avron Barr and Edward A. Felgenbaum. DTIC document number AD...Regents of the University of CalTorni, 1981. 9. Gardner, Anne. "Search," Handbook of Artificial Intelligence, edited by Avron Barr and Edward A...Felgenbaum, DTIC document number AD A074078, 1979. 10. Gardner, Anne,et al. "Natural Language Understanding," Handbook of Artificial Intelligence, edited
The NASA Air Traffic Management Ontology: Technical Documentation
NASA Technical Reports Server (NTRS)
Keller, Richard M.
2017-01-01
This document is intended to serve as comprehensive documentation for the NASA Air Traffic Management (ATM) Ontology. The ATM Ontology is a conceptual model that defines key classes of entities and relationships pertaining to the US National Airspace System (NAS) and the management of air traffic through that system. A wide variety of classes are represented in the ATM Ontology, including classes corresponding to flights, aircraft, manufacturers, airports, airlines, air routes, NAS facilities, air traffic control advisories, weather phenomena, and many others. The Ontology can be useful in the context of a variety of information management tasks relevant to NAS, including information exchange, data query and search, information organization, information integration, and terminology standardization.
Mars Rover imaging systems and directional filtering
NASA Technical Reports Server (NTRS)
Wang, Paul P.
1989-01-01
Computer literature searches were carried out at Duke University and NASA Langley Research Center. The purpose is to enhance personal knowledge based on the technical problems of pattern recognition and image understanding which must be solved for the Mars Rover and Sample Return Mission. Intensive study effort of a large collection of relevant literature resulted in a compilation of all important documents in one place. Furthermore, the documents are being classified into: Mars Rover; computer vision (theory); imaging systems; pattern recognition methodologies; and other smart techniques (AI, neural networks, fuzzy logic, etc).
Schwartze, J; Haarbrandt, B; Fortmeier, D; Haux, R; Seidel, C
2014-01-01
Integration of electronic signatures embedded in health care processes in Germany challenges health care service and supply facilities. The suitability of the signature level of an eligible authentication procedure is confirmed for a large part of documents in clinical practice. However, the concrete design of such a procedure remains unclear. To create a summary of usable user authentication systems suitable for clinical workflows. A Systematic literature review based on nine online bibliographic databases. Search keywords included authentication, access control, information systems, information security and biometrics with terms user authentication, user identification and login in title or abstract. Searches were run between 7 and 12 September 2011. Relevant conference proceedings were searched manually in February 2013. Backward reference search of selected results was done. Only publications fully describing authentication systems used or usable were included. Algorithms or purely theoretical concepts were excluded. Three authors did selection independently. DATA EXTRACTION AND ASSESSMENT: Semi-structured extraction of system characteristics was done by the main author. Identified procedures were assessed for security and fulfillment of relevant laws and guidelines as well as for applicability. Suitability for clinical workflows was derived from the assessments using a weighted sum proposed by Bonneau. Of 7575 citations retrieved, 55 publications meet our inclusion criteria. They describe 48 different authentication systems; 39 were biometric and nine graphical password systems. Assessment of authentication systems showed high error rates above European CENELEC standards and a lack of applicability of biometric systems. Graphical passwords did not add overall value compared to conventional passwords. Continuous authentication can add an additional layer of safety. Only few systems are suitable partially or entirely for use in clinical processes. Suitability strongly depends on national or institutional requirements. Four authentication systems seem to fulfill requirements of authentication procedures for clinical workflows. Research is needed in the area of continuous authentication with biometric methods. A proper authentication system should combine all factors of authentication implementing and connecting secure individual measures.
2017-01-01
Background In many Western countries, after a motor vehicle collision, those involved seek health care for the assessment of injuries and for insurance documentation purposes. In contrast, in many less wealthy countries, there may be limited access to care and no insurance or compensation system. Objective The purpose of this infodemiology study was to investigate the global pattern of evolving Internet usage in countries with and without insurance and the corresponding compensation systems for whiplash injury. Methods We used the Internet search engine analytics via Google Trends to study the health information-seeking behavior concerning whiplash injury at national population levels in Europe. Results We found that the search for “whiplash” is strikingly and consistently often associated with the search for “compensation” in countries or cultures with a tort system. Frequent or traumatic painful injuries; diseases or disorders such as arthritis, headache, radius, and hip fracture; depressive disorders; and fibromyalgia were not associated similarly with searches on “compensation.” Conclusions In this study, we present evidence from the evolving viewpoint of naturalistic Internet search engine analytics that the expectations for receiving compensation may influence Internet search behavior in relation to whiplash injury. PMID:28347974
Borkowski, A; Lee, D H; Sydnor, D L; Johnson, R J; Rabinovitch, A; Moore, G W
2001-01-01
The Pathology and Laboratory Medicine Service of the Veterans Affairs Maryland Health Care System is inspected biannually by the College of American Pathologists (CAP). As of the year 2000, all documentation in the Anatomic Pathology Section is available to all staff through the VA Intranet. Signed, supporting paper documents are on file in the office of the department chair. For the year 2000 CAP inspection, inspectors conducted their document review by use of these Web-based documents, in which each CAP question had a hyperlink to the corresponding section of the procedure manual. Thus inspectors were able to locate the documents relevant to each question quickly and efficiently. The procedure manuals consist of 87 procedures for surgical pathology, 52 procedures for cytopathology, and 25 procedures for autopsy pathology. Each CAP question requiring documentation had from one to three hyperlinks to the corresponding section of the procedure manual. Intranet documentation allows for easier sharing among decentralized institutions and for centralized updates of the laboratory documentation. These documents can be upgraded to allow for multimedia presentations, including text search for key words, hyperlinks to other documents, and images, audio, and video. Use of Web-based documents can improve the efficiency of the inspection process.
NASA Astrophysics Data System (ADS)
Manzella, Giuseppe M. R.; Bartolini, Andrea; Bustaffa, Franco; D'Angelo, Paolo; De Mattei, Maurizio; Frontini, Francesca; Maltese, Maurizio; Medone, Daniele; Monachini, Monica; Novellino, Antonio; Spada, Andrea
2016-04-01
The MAPS (Marine Planning and Service Platform) project is aiming at building a computer platform supporting a Marine Information and Knowledge System. One of the main objective of the project is to develop a repository that should gather, classify and structure marine scientific literature and data thus guaranteeing their accessibility to researchers and institutions by means of standard protocols. In oceanography the cost related to data collection is very high and the new paradigm is based on the concept to collect once and re-use many times (for re-analysis, marine environment assessment, studies on trends, etc). This concept requires the access to quality controlled data and to information that is provided in reports (grey literature) and/or in relevant scientific literature. Hence, creation of new technology is needed by integrating several disciplines such as data management, information systems, knowledge management. In one of the most important EC projects on data management, namely SeaDataNet (www.seadatanet.org), an initial example of knowledge management is provided through the Common Data Index, that is providing links to data and (eventually) to papers. There are efforts to develop search engines to find author's contributions to scientific literature or publications. This implies the use of persistent identifiers (such as DOI), as is done in ORCID. However very few efforts are dedicated to link publications to the data cited or used or that can be of importance for the published studies. This is the objective of MAPS. Full-text technologies are often unsuccessful since they assume the presence of specific keywords in the text; in order to fix this problem, the MAPS project suggests to use different semantic technologies for retrieving the text and data and thus getting much more complying results. The main parts of our design of the search engine are: • Syntactic parser - This module is responsible for the extraction of "rich words" from the text: the whole document gets parsed to extract the words which are more meaningful for the main argument of the document, and applies the extraction in the form of N-grams (mono-grams, bi-grams, tri-grams). • MAPS database - This module is a simple database which contains all the N-grams used by MAPS (physical parameters from SeaDataNet vocabularies) to define our marine "ontology". • Relation identifier - This module performs the most important task of identifying relationships between the N-gram extracted from the text by the parser and the provided oceanographic terminology. It checks N-grams supplied by the Syntactic parser and then matches them with the terms stored in the MAPS database. Found matches are returned back to the parser with flexed form appearing in the source text. • A "relaxed" extractor - This option can be activated when the search engine is launched. It was introduced to give the user a chance to create new N-grams combining existing mono-grams and bi-grams in the database with rich-words found within the source text. The innovation of a semantic engine lies in the fact that the process is not just about the retrieval of already known documents by means of a simple term query but rather the retrieval of a population of documents whose existence was unknown. The system answers by showing a screenshot of results ordered according to the following criteria: • Relevance - of the document with respect to the concept that is searched • Date - of publication of the paper • Source - data provider as defined in the SeaDataNet Common Data Index • Matrix - environmental matrices as defined in the oceanographic field • Geographic area - area specified in the text • Clustering - the process of organizing objects into groups whose members are similar The clustering returns as the output the related documents. For each document the MAPS visualization provides: • Title, author, source/provider of data, web address • Tagging of key terms or concepts • Summary of the document • Visualization of the whole document The possibility of inserting the number of citations for each document among the criteria of the advanced search is currently undergoing; in this case the engine should be able to connect to any of the existing bibliographic citation systems (such as Google Scholar, Scopus, etc.).
ICTNET at Web Track 2009 Diversity task
2009-11-01
performance. On the World Wide Web, there exist many documents which represents several implicit subtopics. We used commerce search engines to gather those...documents. In this task, our work can be divided into five steps. First, we collect documents returned by commerce search engines , and considered
NASA Astrophysics Data System (ADS)
Henze, F.; Magdalinski, N.; Schwarzbach, F.; Schulze, A.; Gerth, Ph.; Schäfer, F.
2013-07-01
Information systems play an important role in historical research as well as in heritage documentation. As part of a joint research project of the German Archaeological Institute, the Brandenburg University of Technology Cottbus and the Dresden University of Applied Sciences a web-based documentation system is currently being developed, which can easily be adapted to the needs of different projects with individual scientific concepts, methods and questions. Based on open source and standardized technologies it will focus on open and well-documented interfaces to ease the dissemination and re-use of its content via web-services and to communicate with desktop applications for further evaluation and analysis. Core of the system is a generic data model that represents a wide range of topics and methods of archaeological work. By the provision of a concerted amount of initial themes and attributes a cross project analysis of research data will be possible. The development of enhanced search and retrieval functionalities will simplify the processing and handling of large heterogeneous data sets. To achieve a high degree of interoperability with existing external data, systems and applications, standardized interfaces will be integrated. The analysis of spatial data shall be possible through the integration of web-based GIS functions. As an extension to this, customized functions for storage, processing and provision of 3D geo data are being developed. As part of the contribution system requirements and concepts will be presented and discussed. A particular focus will be on introducing the generic data model and the derived database schema. The research work on enhanced search and retrieval capabilities will be illustrated by prototypical developments, as well as concepts and first implementations for an integrated 2D/3D Web-GIS.
Is real world evidence influencing practice? A systematic review of CPRD research in NICE guidances.
Oyinlola, Jessie O; Campbell, Jennifer; Kousoulis, Antonis A
2016-07-26
There is currently limited evidence regarding the extent Real World Evidence (RWE) has directly impacted the health and social care systems. The aim of this review is to identify national guidelines or guidances published in England from 2000 onwards which have referenced studies using the governmental primary care data provider the Clinical Practice Research Datalink (CPRD). The methodology recommended by Preferred Reporting Items for Systematic Reviews and Meta-Analyses (PRISMA) was followed. Four databases were searched and documents of interest were identified through a search algorithm containing keywords relevant to CPRD. A search diary was maintained with the inclusion/exclusion decisions which were performed by two independent reviewers. Twenty-five guidance documents were included in the final review (following screening and assessment for eligibility), referencing 43 different CPRD/GPRD studies, all published since 2007. The documents covered 12 disease areas, with the majority (N =7) relevant to diseases of the Central Nervous system (CNS). The 43 studies provided evidence of disease epidemiology, incidence/prevalence, pharmacoepidemiology, pharmacovigilance and health utilisation. A slow uptake of RWE in clinical and therapeutic guidelines (as provided by UK governmental structures) was noticed. However, there seems to be an increasing trend in the use of healthcare system data to inform clinical practice, especially as the real world validity of clinical trials is being questioned. In order to accommodate this increasing demand and meet the paradigm shift expected, organisations need to work together to enable or improve data access, undertake translational and relevant research and establish sources of reliable evidence.
78 FR 68100 - Luminant Generation Company, LLC
Federal Register 2010, 2011, 2012, 2013, 2014
2013-11-13
... following methods: Federal Rulemaking Web site: Go to http://www.regulations.gov and search for Docket ID.../adams.html . To begin the search, select ``ADAMS Public Documents'' and then select ``Begin Web- based ADAMS Search.'' For problems with ADAMS, please contact the NRC's Public Document Room (PDR) reference...
Processable English: The Theory Behind the PENG System
2009-06-01
implicit - is often buried amongst masses of irrelevant data. Heralding from unstructured sources such as natural language documents, email, audio ...estimation and prediction, data-mining, social network analysis, and semantic search and visualisation . This report describes the theoretical
Information Resources; A Searcher's Manual. MOREL Regional Information System for Educators.
ERIC Educational Resources Information Center
Grimes, George; Doyle, James
This document is one of a series describing the background, functions, and utilization of the Regional Information System (RIS) developed by the Michigan-Ohio Regional Educational Laboratory (MOREL). The purpose of this manual is to detail a procedure for performing a productive search of information resources which can satisfy the informational…
Federal Register 2010, 2011, 2012, 2013, 2014
2012-10-09
... documents online in the NRC Library at http://www.nrc.gov/reading-rm/adams.html . To begin the search... of digital instrumentation and control system PRAs, including common cause failures in PRAs and uncertainty analysis associated with new reactor digital systems, and (4) incorporation of additional...
A bioinformatics knowledge discovery in text application for grid computing
Castellano, Marcello; Mastronardi, Giuseppe; Bellotti, Roberto; Tarricone, Gianfranco
2009-01-01
Background A fundamental activity in biomedical research is Knowledge Discovery which has the ability to search through large amounts of biomedical information such as documents and data. High performance computational infrastructures, such as Grid technologies, are emerging as a possible infrastructure to tackle the intensive use of Information and Communication resources in life science. The goal of this work was to develop a software middleware solution in order to exploit the many knowledge discovery applications on scalable and distributed computing systems to achieve intensive use of ICT resources. Methods The development of a grid application for Knowledge Discovery in Text using a middleware solution based methodology is presented. The system must be able to: perform a user application model, process the jobs with the aim of creating many parallel jobs to distribute on the computational nodes. Finally, the system must be aware of the computational resources available, their status and must be able to monitor the execution of parallel jobs. These operative requirements lead to design a middleware to be specialized using user application modules. It included a graphical user interface in order to access to a node search system, a load balancing system and a transfer optimizer to reduce communication costs. Results A middleware solution prototype and the performance evaluation of it in terms of the speed-up factor is shown. It was written in JAVA on Globus Toolkit 4 to build the grid infrastructure based on GNU/Linux computer grid nodes. A test was carried out and the results are shown for the named entity recognition search of symptoms and pathologies. The search was applied to a collection of 5,000 scientific documents taken from PubMed. Conclusion In this paper we discuss the development of a grid application based on a middleware solution. It has been tested on a knowledge discovery in text process to extract new and useful information about symptoms and pathologies from a large collection of unstructured scientific documents. As an example a computation of Knowledge Discovery in Database was applied on the output produced by the KDT user module to extract new knowledge about symptom and pathology bio-entities. PMID:19534749
A bioinformatics knowledge discovery in text application for grid computing.
Castellano, Marcello; Mastronardi, Giuseppe; Bellotti, Roberto; Tarricone, Gianfranco
2009-06-16
A fundamental activity in biomedical research is Knowledge Discovery which has the ability to search through large amounts of biomedical information such as documents and data. High performance computational infrastructures, such as Grid technologies, are emerging as a possible infrastructure to tackle the intensive use of Information and Communication resources in life science. The goal of this work was to develop a software middleware solution in order to exploit the many knowledge discovery applications on scalable and distributed computing systems to achieve intensive use of ICT resources. The development of a grid application for Knowledge Discovery in Text using a middleware solution based methodology is presented. The system must be able to: perform a user application model, process the jobs with the aim of creating many parallel jobs to distribute on the computational nodes. Finally, the system must be aware of the computational resources available, their status and must be able to monitor the execution of parallel jobs. These operative requirements lead to design a middleware to be specialized using user application modules. It included a graphical user interface in order to access to a node search system, a load balancing system and a transfer optimizer to reduce communication costs. A middleware solution prototype and the performance evaluation of it in terms of the speed-up factor is shown. It was written in JAVA on Globus Toolkit 4 to build the grid infrastructure based on GNU/Linux computer grid nodes. A test was carried out and the results are shown for the named entity recognition search of symptoms and pathologies. The search was applied to a collection of 5,000 scientific documents taken from PubMed. In this paper we discuss the development of a grid application based on a middleware solution. It has been tested on a knowledge discovery in text process to extract new and useful information about symptoms and pathologies from a large collection of unstructured scientific documents. As an example a computation of Knowledge Discovery in Database was applied on the output produced by the KDT user module to extract new knowledge about symptom and pathology bio-entities.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Zamora, Antonio
Advanced Natural Language Processing Tools for Web Information Retrieval, Content Analysis, and Synthesis. The goal of this SBIR was to implement and evaluate several advanced Natural Language Processing (NLP) tools and techniques to enhance the precision and relevance of search results by analyzing and augmenting search queries and by helping to organize the search output obtained from heterogeneous databases and web pages containing textual information of interest to DOE and the scientific-technical user communities in general. The SBIR investigated 1) the incorporation of spelling checkers in search applications, 2) identification of significant phrases and concepts using a combination of linguisticmore » and statistical techniques, and 3) enhancement of the query interface and search retrieval results through the use of semantic resources, such as thesauri. A search program with a flexible query interface was developed to search reference databases with the objective of enhancing search results from web queries or queries of specialized search systems such as DOE's Information Bridge. The DOE ETDE/INIS Joint Thesaurus was processed to create a searchable database. Term frequencies and term co-occurrences were used to enhance the web information retrieval by providing algorithmically-derived objective criteria to organize relevant documents into clusters containing significant terms. A thesaurus provides an authoritative overview and classification of a field of knowledge. By organizing the results of a search using the thesaurus terminology, the output is more meaningful than when the results are just organized based on the terms that co-occur in the retrieved documents, some of which may not be significant. An attempt was made to take advantage of the hierarchy provided by broader and narrower terms, as well as other field-specific information in the thesauri. The search program uses linguistic morphological routines to find relevant entries regardless of whether terms are stored in singular or plural form. Implementation of additional inflectional morphology processes for verbs can enhance retrieval further, but this has to be balanced by the possibility of broadening the results too much. In addition to the DOE energy thesaurus, other sources of specialized organized knowledge such as the Medical Subject Headings (MeSH), the Unified Medical Language System (UMLS), and Wikipedia were investigated. The supporting role of the NLP thesaurus search program was enhanced by incorporating spelling aid and a part-of-speech tagger to cope with misspellings in the queries and to determine the grammatical roles of the query words and identify nouns for special processing. To improve precision, multiple modes of searching were implemented including Boolean operators, and field-specific searches. Programs to convert a thesaurus or reference file into searchable support files can be deployed easily, and the resulting files are immediately searchable to produce relevance-ranked results with builtin spelling aid, morphological processing, and advanced search logic. Demonstration systems were built for several databases, including the DOE energy thesaurus.« less
Optimum tuned mass damper design using harmony search with comparison of classical methods
NASA Astrophysics Data System (ADS)
Nigdeli, Sinan Melih; Bekdaş, Gebrail; Sayin, Baris
2017-07-01
As known, tuned mass dampers (TMDs) are added to mechanical systems in order to obtain a good vibration damping. The main aim is to reduce the maximum amplitude at the resonance state. In this study, a metaheuristic algorithm called harmony search employed for the optimum design of TMDs. As the optimization objective, the transfer function of the acceleration of the system with respect to ground acceleration was minimized. The numerical trails were conducted for 4 single degree of freedom systems and the results were compared with classical methods. As a conclusion, the proposed method is feasible and more effective than the other documented methods.
Documents News FAQs Internet Resources Glossary Home » Glossary/Acronyms Search Glossary: Search Glossary Facility EISs | Documents News | FAQs | Internet Resources | Glossary Help | Mailing Services | Contact Us
ERIC Educational Resources Information Center
Liu, Chang
2012-01-01
When using information retrieval (IR) systems, users often pose short and ambiguous query terms. It is critical for IR systems to obtain more accurate representation of users' information need, their document preferences, and the context they are working in, and then incorporate them into the design of the systems to tailor retrieval to…
Müller, H-M; Van Auken, K M; Li, Y; Sternberg, P W
2018-03-09
The biomedical literature continues to grow at a rapid pace, making the challenge of knowledge retrieval and extraction ever greater. Tools that provide a means to search and mine the full text of literature thus represent an important way by which the efficiency of these processes can be improved. We describe the next generation of the Textpresso information retrieval system, Textpresso Central (TPC). TPC builds on the strengths of the original system by expanding the full text corpus to include the PubMed Central Open Access Subset (PMC OA), as well as the WormBase C. elegans bibliography. In addition, TPC allows users to create a customized corpus by uploading and processing documents of their choosing. TPC is UIMA compliant, to facilitate compatibility with external processing modules, and takes advantage of Lucene indexing and search technology for efficient handling of millions of full text documents. Like Textpresso, TPC searches can be performed using keywords and/or categories (semantically related groups of terms), but to provide better context for interpreting and validating queries, search results may now be viewed as highlighted passages in the context of full text. To facilitate biocuration efforts, TPC also allows users to select text spans from the full text and annotate them, create customized curation forms for any data type, and send resulting annotations to external curation databases. As an example of such a curation form, we describe integration of TPC with the Noctua curation tool developed by the Gene Ontology (GO) Consortium. Textpresso Central is an online literature search and curation platform that enables biocurators and biomedical researchers to search and mine the full text of literature by integrating keyword and category searches with viewing search results in the context of the full text. It also allows users to create customized curation interfaces, use those interfaces to make annotations linked to supporting evidence statements, and then send those annotations to any database in the world. Textpresso Central URL: http://www.textpresso.org/tpc.
Sauer, Ursula G; Wächter, Thomas; Hareng, Lars; Wareing, Britta; Langsch, Angelika; Zschunke, Matthias; Alvers, Michael R; Landsiedel, Robert
2014-06-01
The knowledge-based search engine Go3R, www.Go3R.org, has been developed to assist scientists from industry and regulatory authorities in collecting comprehensive toxicological information with a special focus on identifying available alternatives to animal testing. The semantic search paradigm of Go3R makes use of expert knowledge on 3Rs methods and regulatory toxicology, laid down in the ontology, a network of concepts, terms, and synonyms, to recognize the contents of documents. Search results are automatically sorted into a dynamic table of contents presented alongside the list of documents retrieved. This table of contents allows the user to quickly filter the set of documents by topics of interest. Documents containing hazard information are automatically assigned to a user interface following the endpoint-specific IUCLID5 categorization scheme required, e.g. for REACH registration dossiers. For this purpose, complex endpoint-specific search queries were compiled and integrated into the search engine (based upon a gold standard of 310 references that had been assigned manually to the different endpoint categories). Go3R sorts 87% of the references concordantly into the respective IUCLID5 categories. Currently, Go3R searches in the 22 million documents available in the PubMed and TOXNET databases. However, it can be customized to search in other databases including in-house databanks. Copyright © 2013 Elsevier Ltd. All rights reserved.
Centrality based Document Ranking
2014-11-01
clinical domain and very uncommon elsewhere. A regular IR system may fail to rank documents from such a domain, dealing with symptoms, diagnosis and...description). We prepared a hand-crafted list of synonyms for each of the query types, viz. diagnosis , test and treatment. This list was used to expand the...Miller. Semantic search. In INTERNATIONAL WORLD WIDE WEB CONFERENCE, pages 700–709. ACM, 2003. 8. A. Hanbury and M. Lupu . Toward a Model of Domain
Critical Infrastructure References: Documented Literature Search
2012-10-01
the literature search document can be a resource for DRDC and external partners. Future plans: At present, the electronic copies of the reference...Personal Information Protection and Electronic Documents Act (S.C. 2000, c. 5) Title: Personal Information Protection and Electronic Documents Act (S.C...2011 Overview: • "An Act to support and promote electronic commerce by protecting personal information that is collected, used or disclosed in
BIOREMEDIATION IN THE FIELD SEARCH SYSTEM (BFSS) - USER DOCUMENTATION
The Bioremediation Field Initiative is a cooperative effort of the U.S. EPA's Office of Research and Development (ORD), Office of Solid Waste and Emergency Response (OSWER), and regional offices, and other federal agencies, state agencies, industry, and universities to ...
A knowledgebase system to enhance scientific discovery: Telemakus
Fuller, Sherrilynne S; Revere, Debra; Bugni, Paul F; Martin, George M
2004-01-01
Background With the rapid expansion of scientific research, the ability to effectively find or integrate new domain knowledge in the sciences is proving increasingly difficult. Efforts to improve and speed up scientific discovery are being explored on a number of fronts. However, much of this work is based on traditional search and retrieval approaches and the bibliographic citation presentation format remains unchanged. Methods Case study. Results The Telemakus KnowledgeBase System provides flexible new tools for creating knowledgebases to facilitate retrieval and review of scientific research reports. In formalizing the representation of the research methods and results of scientific reports, Telemakus offers a potential strategy to enhance the scientific discovery process. While other research has demonstrated that aggregating and analyzing research findings across domains augments knowledge discovery, the Telemakus system is unique in combining document surrogates with interactive concept maps of linked relationships across groups of research reports. Conclusion Based on how scientists conduct research and read the literature, the Telemakus KnowledgeBase System brings together three innovations in analyzing, displaying and summarizing research reports across a domain: (1) research report schema, a document surrogate of extracted research methods and findings presented in a consistent and structured schema format which mimics the research process itself and provides a high-level surrogate to facilitate searching and rapid review of retrieved documents; (2) research findings, used to index the documents, allowing searchers to request, for example, research studies which have studied the relationship between neoplasms and vitamin E; and (3) visual exploration interface of linked relationships for interactive querying of research findings across the knowledgebase and graphical displays of what is known as well as, through gaps in the map, what is yet to be tested. The rationale and system architecture are described and plans for the future are discussed. PMID:15507158
Development and Evaluation of Thesauri-Based Bibliographic Biomedical Search Engine
ERIC Educational Resources Information Center
Alghoson, Abdullah
2017-01-01
Due to the large volume and exponential growth of biomedical documents (e.g., books, journal articles), it has become increasingly challenging for biomedical search engines to retrieve relevant documents based on users' search queries. Part of the challenge is the matching mechanism of free-text indexing that performs matching based on…
Federal Register 2010, 2011, 2012, 2013, 2014
2012-03-09
... doing a detailed search for all applicable airworthiness related documents that apply to any airplane that has an incorrectly marked data plate and take necessary corrective actions based on the search... detailed search for all applicable airworthiness related documents that apply to any airplane that has an...
Semantic retrieval and navigation in clinical document collections.
Kreuzthaler, Markus; Daumke, Philipp; Schulz, Stefan
2015-01-01
Patients with chronic diseases undergo numerous in- and outpatient treatment periods, and therefore many documents accumulate in their electronic records. We report on an on-going project focussing on the semantic enrichment of medical texts, in order to support recall-oriented navigation across a patient's complete documentation. A document pool of 1,696 de-identified discharge summaries was used for prototyping. A natural language processing toolset for document annotation (based on the text-mining framework UIMA) and indexing (Solr) was used to support a browser-based platform for document import, search and navigation. The integrated search engine combines free text and concept-based querying, supported by dynamically generated facets (diagnoses, procedures, medications, lab values, and body parts). The prototype demonstrates the feasibility of semantic document enrichment within document collections of a single patient. Originally conceived as an add-on for the clinical workplace, this technology could also be adapted to support personalised health record platforms, as well as cross-patient search for cohort building and other secondary use scenarios.
Applications of Fourier transform Raman and infrared spectroscopy in forensic sciences
NASA Astrophysics Data System (ADS)
Kuptsov, Albert N.
2000-02-01
First in the world literature comprehensive digital complementary vibrational spectra collection of polymer materials and search system was developed. Non-destructive combined analysis using complementary FT-Raman and FTIR spectra followed by cross-parallel searching on digital spectral libraries, was applied in different fields of forensic sciences. Some unique possibilities of Raman spectroscopy has been shown in the fields of examination of questioned documents, paper, paints, polymer materials, gemstones and other physical evidences.
Search of medical literature for indoor carbon monoxide exposure
DOE Office of Scientific and Technical Information (OSTI.GOV)
Brennan, T.; Ivanovich, M.
1995-12-01
This report documents a literature search on carbon monoxide. The search was limited to the medical and toxicological databases at the National Library of Medicine (MEDLARS). The databases searched were Medline, Toxline and TOXNET. Searches were performed using a variety of strategies. Combinations of the following keywords were used: carbon, monoxide, accidental, residential, occult, diagnosis, misdiagnosis, heating, furnace, and indoor. The literature was searched from 1966 to the present. Over 1000 references were identified and summarized using the following abbreviations: The major findings of the search are: (1) Acute and subacute carbon monoxide exposures result in a large number ofmore » symptoms affecting the brain, kidneys, respiratory system, retina, and motor functions. (2) Acute and subacute carbon monoxide (CO) poisonings have been misdiagnosed on many occasions. (3) Very few systematic investigations have been made into the frequency and consequences of carbon monoxide poisonings.« less
Survey of Knowledge Representation and Reasoning Systems
2009-07-01
processing large volumes of unstructured information such as natural language documents, email, audio , images and video [Ferrucci et al. 2006]. Using this...information we hope to obtain improved es- timation and prediction, data-mining, social network analysis, and semantic search and visualisation . Knowledge
Fast title extraction method for business documents
NASA Astrophysics Data System (ADS)
Katsuyama, Yutaka; Naoi, Satoshi
1997-04-01
Conventional electronic document filing systems are inconvenient because the user must specify the keywords in each document for later searches. To solve this problem, automatic keyword extraction methods using natural language processing and character recognition have been developed. However, these methods are slow, especially for japanese documents. To develop a practical electronic document filing system, we focused on the extraction of keyword areas from a document by image processing. Our fast title extraction method can automatically extract titles as keywords from business documents. All character strings are evaluated for similarity by rating points associated with title similarity. We classified these points as four items: character sitting size, position of character strings, relative position among character strings, and string attribution. Finally, the character string that has the highest rating is selected as the title area. The character recognition process is carried out on the selected area. It is fast because this process must recognize a small number of patterns in the restricted area only, and not throughout the entire document. The mean performance of this method is an accuracy of about 91 percent and a 1.8 sec. processing time for an examination of 100 Japanese business documents.
Noll-Hussong, Michael
2017-03-27
In many Western countries, after a motor vehicle collision, those involved seek health care for the assessment of injuries and for insurance documentation purposes. In contrast, in many less wealthy countries, there may be limited access to care and no insurance or compensation system. The purpose of this infodemiology study was to investigate the global pattern of evolving Internet usage in countries with and without insurance and the corresponding compensation systems for whiplash injury. We used the Internet search engine analytics via Google Trends to study the health information-seeking behavior concerning whiplash injury at national population levels in Europe. We found that the search for "whiplash" is strikingly and consistently often associated with the search for "compensation" in countries or cultures with a tort system. Frequent or traumatic painful injuries; diseases or disorders such as arthritis, headache, radius, and hip fracture; depressive disorders; and fibromyalgia were not associated similarly with searches on "compensation." In this study, we present evidence from the evolving viewpoint of naturalistic Internet search engine analytics that the expectations for receiving compensation may influence Internet search behavior in relation to whiplash injury. ©Michael Noll-Hussong. Originally published in JMIR Public Health and Surveillance (http://publichealth.jmir.org), 27.03.2017.
The BioPrompt-box: an ontology-based clustering tool for searching in biological databases.
Corsi, Claudio; Ferragina, Paolo; Marangoni, Roberto
2007-03-08
High-throughput molecular biology provides new data at an incredible rate, so that the increase in the size of biological databanks is enormous and very rapid. This scenario generates severe problems not only at indexing time, where suitable algorithmic techniques for data indexing and retrieval are required, but also at query time, since a user query may produce such a large set of results that their browsing and "understanding" becomes humanly impractical. This problem is well known to the Web community, where a new generation of Web search engines is being developed, like Vivisimo. These tools organize on-the-fly the results of a user query in a hierarchy of labeled folders that ease their browsing and knowledge extraction. We investigate this approach on biological data, and propose the so called The BioPrompt-boxsoftware system which deploys ontology-driven clustering strategies for making the searching process of biologists more efficient and effective. The BioPrompt-box (Bpb) defines a document as a biological sequence plus its associated meta-data taken from the underneath databank--like references to ontologies or to external databanks, and plain texts as comments of researchers and (title, abstracts or even body of) papers. Bpboffers several tools to customize the search and the clustering process over its indexed documents. The user can search a set of keywords within a specific field of the document schema, or can execute Blastto find documents relative to homologue sequences. In both cases the search task returns a set of documents (hits) which constitute the answer to the user query. Since the number of hits may be large, Bpbclusters them into groups of homogenous content, organized as a hierarchy of labeled clusters. The user can actually choose among several ontology-based hierarchical clustering strategies, each offering a different "view" of the returned hits. Bpbcomputes these views by exploiting the meta-data present within the retrieved documents such as the references to Gene Ontology, the taxonomy lineage, the organism and the keywords. Of course, the approach is flexible enough to leave room for future additions of other meta-information. The ultimate goal of the clustering process is to provide the user with several different readings of the (maybe numerous) query results and show possible hidden correlations among them, thus improving their browsing and understanding. Bpb is a powerful search engine that makes it very easy to perform complex queries over the indexed databanks (currently only UNIPROT is considered). The ontology-based clustering approach is efficient and effective, and could thus be applied successfully to larger databanks, like GenBank or EMBL.
The BioPrompt-box: an ontology-based clustering tool for searching in biological databases
Corsi, Claudio; Ferragina, Paolo; Marangoni, Roberto
2007-01-01
Background High-throughput molecular biology provides new data at an incredible rate, so that the increase in the size of biological databanks is enormous and very rapid. This scenario generates severe problems not only at indexing time, where suitable algorithmic techniques for data indexing and retrieval are required, but also at query time, since a user query may produce such a large set of results that their browsing and "understanding" becomes humanly impractical. This problem is well known to the Web community, where a new generation of Web search engines is being developed, like Vivisimo. These tools organize on-the-fly the results of a user query in a hierarchy of labeled folders that ease their browsing and knowledge extraction. We investigate this approach on biological data, and propose the so called The BioPrompt-boxsoftware system which deploys ontology-driven clustering strategies for making the searching process of biologists more efficient and effective. Results The BioPrompt-box (Bpb) defines a document as a biological sequence plus its associated meta-data taken from the underneath databank – like references to ontologies or to external databanks, and plain texts as comments of researchers and (title, abstracts or even body of) papers. Bpboffers several tools to customize the search and the clustering process over its indexed documents. The user can search a set of keywords within a specific field of the document schema, or can execute Blastto find documents relative to homologue sequences. In both cases the search task returns a set of documents (hits) which constitute the answer to the user query. Since the number of hits may be large, Bpbclusters them into groups of homogenous content, organized as a hierarchy of labeled clusters. The user can actually choose among several ontology-based hierarchical clustering strategies, each offering a different "view" of the returned hits. Bpbcomputes these views by exploiting the meta-data present within the retrieved documents such as the references to Gene Ontology, the taxonomy lineage, the organism and the keywords. Of course, the approach is flexible enough to leave room for future additions of other meta-information. The ultimate goal of the clustering process is to provide the user with several different readings of the (maybe numerous) query results and show possible hidden correlations among them, thus improving their browsing and understanding. Conclusion Bpb is a powerful search engine that makes it very easy to perform complex queries over the indexed databanks (currently only UNIPROT is considered). The ontology-based clustering approach is efficient and effective, and could thus be applied successfully to larger databanks, like GenBank or EMBL. PMID:17430575
29 CFR 1610.15 - Schedule of fees and method of payment for services rendered.
Code of Federal Regulations, 2012 CFR
2012-07-01
... search, review and duplication. The Commission shall not charge for review at the administrative appeal... shall charge the direct costs for document search time after the first two hours and the direct costs for document duplication after the first 100 pages. The first two hours of search time and the first...
29 CFR 1610.15 - Schedule of fees and method of payment for services rendered.
Code of Federal Regulations, 2013 CFR
2013-07-01
... search, review and duplication. The Commission shall not charge for review at the administrative appeal... shall charge the direct costs for document search time after the first two hours and the direct costs for document duplication after the first 100 pages. The first two hours of search time and the first...
29 CFR 1610.15 - Schedule of fees and method of payment for services rendered.
Code of Federal Regulations, 2014 CFR
2014-07-01
... search, review and duplication. The Commission shall not charge for review at the administrative appeal... shall charge the direct costs for document search time after the first two hours and the direct costs for document duplication after the first 100 pages. The first two hours of search time and the first...
ERIC Educational Resources Information Center
Darrah, Brenda
Researchers for small businesses, which may have no access to expensive databases or market research reports, must often rely on information found on the Internet, which can be difficult to find. Although current conventional Internet search engines are now able to index over on billion documents, there are many more documents existing in…
Federal Register 2010, 2011, 2012, 2013, 2014
2012-08-21
... any of the following methods: Federal Rulemaking Web Site: Go to http://www.regulations.gov and search.../reading-rm/adams.html . To begin the search, select ``ADAMS Public Documents'' and then select ``Begin Web- based ADAMS Search.'' For problems with ADAMS, please contact the NRC's Public Document Room (PDR...
Federal Register 2010, 2011, 2012, 2013, 2014
2012-09-12
....regulations.gov and search for Docket ID NRC-2012-0212. Address questions about NRC dockets to Carol Gallagher....gov/reading-rm/adams.html . To begin the search, select ``ADAMS Public Documents'' and then select ``Begin Web- based ADAMS Search.'' For problems with ADAMS, please contact the NRC's Public Document Room...
Patel, Amisha; Prabhakaran, Dorairaj; Berendsen, Mark; Mohanan, P P; Huffman, Mark D
2017-04-01
Ischemic heart disease is the leading cause of death in India. In high-income countries, pre-hospital systems of care have been developed to manage acute manifestations of ischemic heart disease, such as acute coronary syndrome (ACS). However, it is unknown whether guidelines, policies, regulations, or laws exist to guide pre-hospital ACS care in India. We undertook a nation-wide document analysis to address this gap in knowledge. From November 2014 to May 2016, we searched for publicly available emergency care guidelines and legislation addressing pre-hospital ACS care in all 29 Indian states and 7 Union Territories via Internet search and direct correspondence. We found two documents addressing pre-hospital ACS care. Though India has legislation mandating acute care for emergencies such as trauma, regulations or laws to guide pre-hospital ACS care are largely absent. Policy makers urgently need to develop comprehensive, multi-stakeholder policies for pre-hospital emergency cardiovascular care in India. Copyright © 2016. Published by Elsevier B.V.
Support Vector Machines: Relevance Feedback and Information Retrieval.
ERIC Educational Resources Information Center
Drucker, Harris; Shahrary, Behzad; Gibbon, David C.
2002-01-01
Compares support vector machines (SVMs) to Rocchio, Ide regular and Ide dec-hi algorithms in information retrieval (IR) of text documents using relevancy feedback. If the preliminary search is so poor that one has to search through many documents to find at least one relevant document, then SVM is preferred. Includes nine tables. (Contains 24…
Hanauer, David A; Wu, Danny T Y; Yang, Lei; Mei, Qiaozhu; Murkowski-Steffy, Katherine B; Vydiswaran, V G Vinod; Zheng, Kai
2017-03-01
The utility of biomedical information retrieval environments can be severely limited when users lack expertise in constructing effective search queries. To address this issue, we developed a computer-based query recommendation algorithm that suggests semantically interchangeable terms based on an initial user-entered query. In this study, we assessed the value of this approach, which has broad applicability in biomedical information retrieval, by demonstrating its application as part of a search engine that facilitates retrieval of information from electronic health records (EHRs). The query recommendation algorithm utilizes MetaMap to identify medical concepts from search queries and indexed EHR documents. Synonym variants from UMLS are used to expand the concepts along with a synonym set curated from historical EHR search logs. The empirical study involved 33 clinicians and staff who evaluated the system through a set of simulated EHR search tasks. User acceptance was assessed using the widely used technology acceptance model. The search engine's performance was rated consistently higher with the query recommendation feature turned on vs. off. The relevance of computer-recommended search terms was also rated high, and in most cases the participants had not thought of these terms on their own. The questions on perceived usefulness and perceived ease of use received overwhelmingly positive responses. A vast majority of the participants wanted the query recommendation feature to be available to assist in their day-to-day EHR search tasks. Challenges persist for users to construct effective search queries when retrieving information from biomedical documents including those from EHRs. This study demonstrates that semantically-based query recommendation is a viable solution to addressing this challenge. Published by Elsevier Inc.
Document Delivery from Full-Text Online Files: A Pilot Project.
ERIC Educational Resources Information Center
Gillikin, David P.
1990-01-01
Describes the Electronic Journal Retrieval Project (EJRP) developed at the University of Tennessee, Knoxville Libraries, to provide full-text journal articles from online systems. Highlights include costs of various search strategies; implications for library services; collection development and interlibrary loan considerations; and suggestions…
MISSE in the Materials and Processes Technical Information System (MAPTIS )
NASA Technical Reports Server (NTRS)
Burns, DeWitt; Finckenor, Miria; Henrie, Ben
2013-01-01
Materials International Space Station Experiment (MISSE) data is now being collected and distributed through the Materials and Processes Technical Information System (MAPTIS) at Marshall Space Flight Center in Huntsville, Alabama. MISSE data has been instrumental in many programs and continues to be an important source of data for the space community. To facilitate great access to the MISSE data the International Space Station (ISS) program office and MAPTIS are working to gather this data into a central location. The MISSE database contains information about materials, samples, and flights along with pictures, pdfs, excel files, word documents, and other files types. Major capabilities of the system are: access control, browsing, searching, reports, and record comparison. The search capabilities will search within any searchable files so even if the desired meta-data has not been associated data can still be retrieved. Other functionality will continue to be added to the MISSE database as the Athena Platform is expanded
Supporting online learning with games
NASA Astrophysics Data System (ADS)
Yao, JingTao; Kim, DongWon; Herbert, Joseph P.
2007-04-01
This paper presents a study on Web-based learning support systems that is enhanced with two major subsystems: a Web-based learning game and a learning-oriented Web search. The Internet and theWeb may be considered as a first resource for students seeking for information and help. However, much of the information available online is not related to the course contents or is wrong in the worse case. The search subsystem aims to provide students with precise, relative and adaptable documents about certain courses or classes. Therefore, students do not have to spend time to verify the relationship of documents to the class. The learning game subsystem stimulates students to study, enables students to review their studies and to perform self-evaluation through a Web-based learning game such as a treasure hunt game. During the challenge and entertaining learning and evaluation process, it is hoped that students will eventually understand and master the course concepts easily. The goal of developing such a system is to provide students with an efficient and effective learning environment.
Wieland, L Susan; Rutkow, Lainie; Vedula, S Swaroop; Kaufmann, Christopher N; Rosman, Lori M; Twose, Claire; Mahendraratnam, Nirosha; Dickersin, Kay
2014-01-01
To describe the sources of internal company documents used in public health and healthcare research. We searched PubMed and Embase for articles using internal company documents to address a research question about a health-related topic. Our primary interest was where authors obtained internal company documents for their research. We also extracted information on type of company, type of research question, type of internal documents, and funding source. Our searches identified 9,305 citations of which 357 were eligible. Scanning of reference lists and consultation with colleagues identified 4 additional articles, resulting in 361 included articles. Most articles examined internal tobacco company documents (325/361; 90%). Articles using documents from pharmaceutical companies (20/361; 6%) were the next most common. Tobacco articles used documents from repositories; pharmaceutical documents were from a range of sources. Most included articles relied upon internal company documents obtained through litigation (350/361; 97%). The research questions posed were primarily about company strategies to promote or position the company and its products (326/361; 90%). Most articles (346/361; 96%) used information from miscellaneous documents such as memos or letters, or from unspecified types of documents. When explicit information about study funding was provided (290/361 articles), the most common source was the US-based National Cancer Institute. We developed an alternative and more sensitive search targeted at identifying additional research articles using internal pharmaceutical company documents, but the search retrieved an impractical number of citations for review. Internal company documents provide an excellent source of information on health topics (e.g., corporate behavior, study data) exemplified by articles based on tobacco industry documents. Pharmaceutical and other industry documents appear to have been less used for research, indicating a need for funding for this type of research and well-indexed and curated repositories to provide researchers with ready access to the documents.
Wieland, L. Susan; Rutkow, Lainie; Vedula, S. Swaroop; Kaufmann, Christopher N.; Rosman, Lori M.; Twose, Claire; Mahendraratnam, Nirosha; Dickersin, Kay
2014-01-01
Objective To describe the sources of internal company documents used in public health and healthcare research. Methods We searched PubMed and Embase for articles using internal company documents to address a research question about a health-related topic. Our primary interest was where authors obtained internal company documents for their research. We also extracted information on type of company, type of research question, type of internal documents, and funding source. Results Our searches identified 9,305 citations of which 357 were eligible. Scanning of reference lists and consultation with colleagues identified 4 additional articles, resulting in 361 included articles. Most articles examined internal tobacco company documents (325/361; 90%). Articles using documents from pharmaceutical companies (20/361; 6%) were the next most common. Tobacco articles used documents from repositories; pharmaceutical documents were from a range of sources. Most included articles relied upon internal company documents obtained through litigation (350/361; 97%). The research questions posed were primarily about company strategies to promote or position the company and its products (326/361; 90%). Most articles (346/361; 96%) used information from miscellaneous documents such as memos or letters, or from unspecified types of documents. When explicit information about study funding was provided (290/361 articles), the most common source was the US-based National Cancer Institute. We developed an alternative and more sensitive search targeted at identifying additional research articles using internal pharmaceutical company documents, but the search retrieved an impractical number of citations for review. Conclusions Internal company documents provide an excellent source of information on health topics (e.g., corporate behavior, study data) exemplified by articles based on tobacco industry documents. Pharmaceutical and other industry documents appear to have been less used for research, indicating a need for funding for this type of research and well-indexed and curated repositories to provide researchers with ready access to the documents. PMID:24800999
Hart, Jan; Drost, Charles
2008-01-01
We conducted a baseline inventory for avian species at Manzanar National Historic Site, Inyo County, Calif., from 2002 to 2005. Under the guidelines of the Mojave Network Biological Inventory Program, the primary objectives for this study were to (1) inventory and document the occurrence of avian species at Manzanar, with the goal of documenting at least 90 percent of the species present; (2) provide a geographic information system (GIS)-referenced list of sensitive species occurring at Manzanar that are rare, on Federal or State lists, or otherwise worthy of special consideration; and (3) enter all species data into the National Park Service NPSpecies database. Survey methods included general area searches, variable circular plot point-count censusing, nocturnal surveys, and nest searching. During 13 year-round survey sessions, we documented the occurrence of 132 bird species at Manzanar and confirmed breeding by 19 of these. Based on our findings, as well as review of the literature and searches for records of species occurrence, we estimate inventory completeness for regularly occurring bird species at Manzanar to be near 90 percent. No sensitive species on Federal or State lists were found. The distribution and relative abundance of common bird species at this site is now well enough known to begin development of a monitoring protocol for this group.
Earth System Documentation (ES-DOC) Preparation for CMIP6
NASA Astrophysics Data System (ADS)
Denvil, S.; Murphy, S.; Greenslade, M. A.; Lawrence, B.; Guilyardi, E.; Pascoe, C.; Treshanksy, A.; Elkington, M.; Hibling, E.; Hassell, D.
2015-12-01
During the course of 2015 the Earth System Documentation (ES-DOC) project began its preparations for CMIP6 (Coupled Model Inter-comparison Project 6) by further extending the ES-DOC tooling ecosystem in support of Earth System Model (ESM) documentation creation, search, viewing & comparison. The ES-DOC online questionnaire, the ES-DOC desktop notebook, and the ES-DOC python toolkit will serve as multiple complementary pathways to generating CMIP6 documentation. It is envisaged that institutes will leverage these tools at different points of the CMIP6 lifecycle. Institutes will be particularly interested to know that the documentation burden will be either streamlined or completely automated.As all the tools are tightly integrated with the ES-DOC web-service, institutes can be confident that the latency between documentation creation & publishing will be reduced to a minimum. Published documents will be viewable with the online ES-DOC Viewer (accessible via citable URL's). Model inter-comparison scenarios will be supported using the ES-DOC online Comparator tool. The Comparator is being extended to:• Support comparison of both Model descriptions & Simulation runs;• Greatly streamline the effort involved in compiling official tables.The entire ES-DOC ecosystem is open source and built upon open standards such as the Common Information Model (CIM) (versions 1 and 2).
XML Flight/Ground Data Dictionary Management
NASA Technical Reports Server (NTRS)
Wright, Jesse; Wiklow, Colette
2007-01-01
A computer program generates Extensible Markup Language (XML) files that effect coupling between the command- and telemetry-handling software running aboard a spacecraft and the corresponding software running in ground support systems. The XML files are produced by use of information from the flight software and from flight-system engineering. The XML files are converted to legacy ground-system data formats for command and telemetry, transformed into Web-based and printed documentation, and used in developing new ground-system data-handling software. Previously, the information about telemetry and command was scattered in various paper documents that were not synchronized. The process of searching and reading the documents was time-consuming and introduced errors. In contrast, the XML files contain all of the information in one place. XML structures can evolve in such a manner as to enable the addition, to the XML files, of the metadata necessary to track the changes and the associated documentation. The use of this software has reduced the extent of manual operations in developing a ground data system, thereby saving considerable time and removing errors that previously arose in the translation and transcription of software information from the flight to the ground system.
Joint Force Quarterly. Issue 55, 4th Quarter 2009
2009-01-01
average 1 hour per response, including the time for reviewing instructions, searching existing data sources , gathering and maintaining the data needed...to critique systems theory from some of its source documents, he could have accessed the many works of scientists and philosophers such as Russell...numerous sources , including writings of some leading systems theorists. It is simply false to claim, as Professor Czarnecki does (and many systems
Human Systems Integration (HSI) in Acquisition. Acquisition Phase Guide
2009-08-01
information is estimated to average 1 hour per response, including the time for reviewing instructions, searching existing data sources, gathering and...maintaining the data needed, and completing and reviewing the collection of information. Send comments regarding this burden estimate or any other aspect of...available Concept of Operations (CONOPS) and other available data 1.1 Select and review Baseline Comparison System(s) (BCS) documentation 1.2 Assess
ERIC Educational Resources Information Center
Gantt, Linda, Comp.; Schmal, Marilyn Strauss, Comp.
The bibliography on art therapy presents 1175 citations (1940-1973) drawn from searches of the medical indexes, computer systems of the National Library of Medicine and the National Institute of Mental Health, other bibliographies, Centre International de Documentation Concernant les Expressions Plastiques, and the American Journal of Art Therapy.…
PlateRunner: A Search Engine to Identify EMR Boilerplates.
Divita, Guy; Workman, T Elizabeth; Carter, Marjorie E; Redd, Andrew; Samore, Matthew H; Gundlapalli, Adi V
2016-01-01
Medical text contains boilerplated content, an artifact of pull-down forms from EMRs. Boilerplated content is the source of challenges for concept extraction on clinical text. This paper introduces PlateRunner, a search engine on boilerplates from the US Department of Veterans Affairs (VA) EMR. Boilerplates containing concepts should be identified and reviewed to recognize challenging formats, identify high yield document titles, and fine tune section zoning. This search engine has the capability to filter negated and asserted concepts, save and search query results. This tool can save queries, search results, and documents found for later analysis.
Nursing documentation with NANDA and NIC in a comprehensive HIS/EPR system.
Flø, Kåre
2006-01-01
DIPS nursing documentation system facilitates that nurses can write several types of notes into the EPR. Within these notes the nurses can register NANDA diagnoses and NIC interventions with nursing activities. To choose NANDA and NIC the nurse can use a search engine, or she can choose a relevant Care plan guideline and pick the suggested diagnoses and interventions from there. Diagnoses and interventions with nursing activities registered are presented in a Care plan. When a nurse writes a note for a patient she will always be presented the Care plan and she can easy evaluate and update the Care plan.
Forrest, Lynne F; Adams, Jean; Ben-Shlomo, Yoav; Buckner, Stefanie; Payne, Nick; Rimmer, Melanie; Salway, Sarah; Sowden, Sarah; Walters, Kate; White, Martin
2017-05-01
older people may be less likely to receive interventions than younger people. Age bias in national guidance may influence entire public health and health care systems. We examined how English National Institute for Health & Care Excellence (NICE) guidance and guidelines consider age. we undertook a documentary analysis of NICE public health (n = 33) and clinical (n = 114) guidelines and technology appraisals (n = 212). We systematically searched for age-related terms, and conducted thematic analysis of the paragraphs in which these occurred ('age-extracts'). Quantitative analysis explored frequency of age-extracts between and within document types. Illustrative quotes were used to elaborate and explain quantitative findings. 2,314 age-extracts were identified within three themes: age documented as an a-priori consideration at scope-setting (518 age-extracts, 22.4%); documentation of differential effectiveness, cost-effectiveness or other outcomes by age (937 age-extracts, 40.5%); and documentation of age-specific recommendations (859 age-extracts, 37.1%). Public health guidelines considered age most comprehensively. There were clear examples of older-age being considered in both evidence searching and in making recommendations, suggesting that this can be achieved within current processes. we found inconsistencies in how age is considered in NICE guidance and guidelines. More effort may be required to ensure age is consistently considered. Future NICE committees should search for and document evidence of age-related differences in receipt of interventions. Where evidence relating to effectiveness and cost-effectiveness in older populations is available, more explicit age-related recommendations should be made. Where there is a lack of evidence, it should be stated what new research is needed. © The Author 2016. Published by Oxford University Press on behalf of the British Geriatrics Society.
NASA Technical Reports Server (NTRS)
Maluf, David A.; Tran, Peter B.
2003-01-01
Object-Relational database management system is an integrated hybrid cooperative approach to combine the best practices of both the relational model utilizing SQL queries and the object-oriented, semantic paradigm for supporting complex data creation. In this paper, a highly scalable, information on demand database framework, called NETMARK, is introduced. NETMARK takes advantages of the Oracle 8i object-relational database using physical addresses data types for very efficient keyword search of records spanning across both context and content. NETMARK was originally developed in early 2000 as a research and development prototype to solve the vast amounts of unstructured and semistructured documents existing within NASA enterprises. Today, NETMARK is a flexible, high-throughput open database framework for managing, storing, and searching unstructured or semi-structured arbitrary hierarchal models, such as XML and HTML.
An Extensible Schema-less Database Framework for Managing High-throughput Semi-Structured Documents
NASA Technical Reports Server (NTRS)
Maluf, David A.; Tran, Peter B.; La, Tracy; Clancy, Daniel (Technical Monitor)
2002-01-01
Object-Relational database management system is an integrated hybrid cooperative approach to combine the best practices of both the relational model utilizing SQL queries and the object oriented, semantic paradigm for supporting complex data creation. In this paper, a highly scalable, information on demand database framework, called NETMARK is introduced. NETMARK takes advantages of the Oracle 8i object-relational database using physical addresses data types for very efficient keyword searches of records for both context and content. NETMARK was originally developed in early 2000 as a research and development prototype to solve the vast amounts of unstructured and semi-structured documents existing within NASA enterprises. Today, NETMARK is a flexible, high throughput open database framework for managing, storing, and searching unstructured or semi structured arbitrary hierarchal models, XML and HTML.
Efficient automatic OCR word validation using word partial format derivation and language model
NASA Astrophysics Data System (ADS)
Chen, Siyuan; Misra, Dharitri; Thoma, George R.
2010-01-01
In this paper we present an OCR validation module, implemented for the System for Preservation of Electronic Resources (SPER) developed at the U.S. National Library of Medicine.1 The module detects and corrects suspicious words in the OCR output of scanned textual documents through a procedure of deriving partial formats for each suspicious word, retrieving candidate words by partial-match search from lexicons, and comparing the joint probabilities of N-gram and OCR edit transformation corresponding to the candidates. The partial format derivation, based on OCR error analysis, efficiently and accurately generates candidate words from lexicons represented by ternary search trees. In our test case comprising a historic medico-legal document collection, this OCR validation module yielded the correct words with 87% accuracy and reduced the overall OCR word errors by around 60%.
Document similarity measures and document browsing
NASA Astrophysics Data System (ADS)
Ahmadullin, Ildus; Fan, Jian; Damera-Venkata, Niranjan; Lim, Suk Hwan; Lin, Qian; Liu, Jerry; Liu, Sam; O'Brien-Strain, Eamonn; Allebach, Jan
2011-03-01
Managing large document databases is an important task today. Being able to automatically com- pare document layouts and classify and search documents with respect to their visual appearance proves to be desirable in many applications. We measure single page documents' similarity with respect to distance functions between three document components: background, text, and saliency. Each document component is represented as a Gaussian mixture distribution; and distances between dierent documents' components are calculated as probabilistic similarities between corresponding distributions. The similarity measure between documents is represented as a weighted sum of the components' distances. Using this document similarity measure, we propose a browsing mechanism operating on a document dataset. For these purposes, we use a hierarchical browsing environment which we call the document similarity pyramid. It allows the user to browse a large document dataset and to search for documents in the dataset that are similar to the query. The user can browse the dataset on dierent levels of the pyramid, and zoom into the documents that are of interest.
Semantic Clustering of Search Engine Results
Soliman, Sara Saad; El-Sayed, Maged F.; Hassan, Yasser F.
2015-01-01
This paper presents a novel approach for search engine results clustering that relies on the semantics of the retrieved documents rather than the terms in those documents. The proposed approach takes into consideration both lexical and semantics similarities among documents and applies activation spreading technique in order to generate semantically meaningful clusters. This approach allows documents that are semantically similar to be clustered together rather than clustering documents based on similar terms. A prototype is implemented and several experiments are conducted to test the prospered solution. The result of the experiment confirmed that the proposed solution achieves remarkable results in terms of precision. PMID:26933673
Texas Conservation Guide for Municipal Services: A Report to Texas Cities.
ERIC Educational Resources Information Center
Walker, Jacilyn G.; DeHaven, Martha L.
This document presents a collection of energy conservation ideas gathered from a survey of Texas cities and via a national search of recent technical information. The conservation ideas presented are grouped into categories of: administrative concerns, public buildings, public vehicle fleets, transportation systems, municipal utilities, alternate…
Highlights: Selected Statements from the Literature on Collaboration and Coordination.
ERIC Educational Resources Information Center
Roberts, Jane
Thirty documents selected from the literature on educational change, social systems, and management and organization were searched for statements considered relevant to collaboration and coordination between organizational units. These statements, either quoted or paraphrased, are listed in this report along with their sources. The statements are…
The state of the art of general aviation autopilots
NASA Technical Reports Server (NTRS)
See, M. J.; Levy, D.
1980-01-01
The study is based on the information obtained from a general literature search, product literature, and visitations and interviews with manufacturers, users, and service centers. State of the art autopilots are documented with respect to total systems, components, and functions. Recommendations concerning potential areas of further research are also presented.
SEARCHING FOR A WATER SUPPLY CONNECTION IN THE CABOOL, MISSOURI OUTBREAK
A recent disease outbreak resulting in 4 deaths, 32 hospitalization and a total of 243 documented cases of diarrhea was linked epidemiologically and by on-site data gathering supported by the use of a distribution system model to the public water supply. The pathogenic agent, Esc...
The Number of Scholarly Documents on the Public Web
Khabsa, Madian; Giles, C. Lee
2014-01-01
The number of scholarly documents available on the web is estimated using capture/recapture methods by studying the coverage of two major academic search engines: Google Scholar and Microsoft Academic Search. Our estimates show that at least 114 million English-language scholarly documents are accessible on the web, of which Google Scholar has nearly 100 million. Of these, we estimate that at least 27 million (24%) are freely available since they do not require a subscription or payment of any kind. In addition, at a finer scale, we also estimate the number of scholarly documents on the web for fifteen fields: Agricultural Science, Arts and Humanities, Biology, Chemistry, Computer Science, Economics and Business, Engineering, Environmental Sciences, Geosciences, Material Science, Mathematics, Medicine, Physics, Social Sciences, and Multidisciplinary, as defined by Microsoft Academic Search. In addition, we show that among these fields the percentage of documents defined as freely available varies significantly, i.e., from 12 to 50%. PMID:24817403
The number of scholarly documents on the public web.
Khabsa, Madian; Giles, C Lee
2014-01-01
The number of scholarly documents available on the web is estimated using capture/recapture methods by studying the coverage of two major academic search engines: Google Scholar and Microsoft Academic Search. Our estimates show that at least 114 million English-language scholarly documents are accessible on the web, of which Google Scholar has nearly 100 million. Of these, we estimate that at least 27 million (24%) are freely available since they do not require a subscription or payment of any kind. In addition, at a finer scale, we also estimate the number of scholarly documents on the web for fifteen fields: Agricultural Science, Arts and Humanities, Biology, Chemistry, Computer Science, Economics and Business, Engineering, Environmental Sciences, Geosciences, Material Science, Mathematics, Medicine, Physics, Social Sciences, and Multidisciplinary, as defined by Microsoft Academic Search. In addition, we show that among these fields the percentage of documents defined as freely available varies significantly, i.e., from 12 to 50%.
Modelling and Simulation of Search Engine
NASA Astrophysics Data System (ADS)
Nasution, Mahyuddin K. M.
2017-01-01
The best tool currently used to access information is a search engine. Meanwhile, the information space has its own behaviour. Systematically, an information space needs to be familiarized with mathematics so easily we identify the characteristics associated with it. This paper reveal some characteristics of search engine based on a model of document collection, which are then estimated the impact on the feasibility of information. We reveal some of characteristics of search engine on the lemma and theorem about singleton and doubleton, then computes statistically characteristic as simulating the possibility of using search engine. In this case, Google and Yahoo. There are differences in the behaviour of both search engines, although in theory based on the concept of documents collection.
Document Exploration and Automatic Knowledge Extraction for Unstructured Biomedical Text
NASA Astrophysics Data System (ADS)
Chu, S.; Totaro, G.; Doshi, N.; Thapar, S.; Mattmann, C. A.; Ramirez, P.
2015-12-01
We describe our work on building a web-browser based document reader with built-in exploration tool and automatic concept extraction of medical entities for biomedical text. Vast amounts of biomedical information are offered in unstructured text form through scientific publications and R&D reports. Utilizing text mining can help us to mine information and extract relevant knowledge from a plethora of biomedical text. The ability to employ such technologies to aid researchers in coping with information overload is greatly desirable. In recent years, there has been an increased interest in automatic biomedical concept extraction [1, 2] and intelligent PDF reader tools with the ability to search on content and find related articles [3]. Such reader tools are typically desktop applications and are limited to specific platforms. Our goal is to provide researchers with a simple tool to aid them in finding, reading, and exploring documents. Thus, we propose a web-based document explorer, which we called Shangri-Docs, which combines a document reader with automatic concept extraction and highlighting of relevant terms. Shangri-Docsalso provides the ability to evaluate a wide variety of document formats (e.g. PDF, Words, PPT, text, etc.) and to exploit the linked nature of the Web and personal content by performing searches on content from public sites (e.g. Wikipedia, PubMed) and private cataloged databases simultaneously. Shangri-Docsutilizes Apache cTAKES (clinical Text Analysis and Knowledge Extraction System) [4] and Unified Medical Language System (UMLS) to automatically identify and highlight terms and concepts, such as specific symptoms, diseases, drugs, and anatomical sites, mentioned in the text. cTAKES was originally designed specially to extract information from clinical medical records. Our investigation leads us to extend the automatic knowledge extraction process of cTAKES for biomedical research domain by improving the ontology guided information extraction process. We will describe our experience and implementation of our system and share lessons learned from our development. We will also discuss ways in which this could be adapted to other science fields. [1] Funk et al., 2014. [2] Kang et al., 2014. [3] Utopia Documents, http://utopiadocs.com [4] Apache cTAKES, http://ctakes.apache.org
MR-Compatible Integrated Eye Tracking System
2016-03-10
SECURITY CLASSIFICATION OF: This instrumentation grant was used to purchase state-of-the-art, high-resolution video eye tracker that can be used to...P.O. Box 12211 Research Triangle Park, NC 27709-2211 video eye tracking, eye movments, visual search; camouflage-breaking REPORT DOCUMENTATION PAGE...Report: MR-Compatible Integrated Eye Tracking System Report Title This instrumentation grant was used to purchase state-of-the-art, high-resolution video
Digital Archive Issues from the Perspective of an Earth Science Data Producer
NASA Technical Reports Server (NTRS)
Barkstrom, Bruce R.
2004-01-01
Contents include the following: Introduction. A Producer Perspective on Earth Science Data. Data Producers as Members of a Scientific Community. Some Unique Characteristics of Scientific Data. Spatial and Temporal Sampling for Earth (or Space) Science Data. The Influence of the Data Production System Architecture. The Spatial and Temporal Structures Underlying Earth Science Data. Earth Science Data File (or Relation) Schemas. Data Producer Configuration Management Complexities. The Topology of Earth Science Data Inventories. Some Thoughts on the User Perspective. Science Data User Communities. Spatial and Temporal Structure Needs of Different Users. User Spatial Objects. Data Search Services. Inventory Search. Parameter (Keyword) Search. Metadata Searches. Documentation Search. Secondary Index Search. Print Technology and Hypertext. Inter-Data Collection Configuration Management Issues. An Archive View. Producer Data Ingest and Production. User Data Searching and Distribution. Subsetting and Supersetting. Semantic Requirements for Data Interchange. Tentative Conclusions. An Object Oriented View of Archive Information Evolution. Scientific Data Archival Issues. A Perspective on the Future of Digital Archives for Scientific Data. References Index for this paper.
Raising the IQ in full-text searching via intelligent querying
DOE Office of Scientific and Technical Information (OSTI.GOV)
Kero, R.; Russell, L.; Swietlik, C.
1994-11-01
Current Information Retrieval (IR) technologies allow for efficient access to relevant information, provided that user selected query terms coincide with the specific linguistical choices made by the authors whose works constitute the text-base. Therefore, the challenge is to enhance the limited searching capability of state-of-the-practice IR. This can be done either with augmented clients that overcome current server searching deficiencies, or with added capabilities that can augment searching algorithms on the servers. The technology being investigated is that of deductive databases, with a set of new techniques called cooperative answering. This technology utilizes semantic networks to allow for navigation betweenmore » possible query search term alternatives. The augmented search terms are passed to an IR engine and the results can be compared. The project utilizes the OSTI Environment, Safety and Health Thesaurus to populate the domain specific semantic network and the text base of ES&H related documents from the Facility Profile Information Management System as the domain specific search space.« less
Word spotting for handwritten documents using Chamfer Distance and Dynamic Time Warping
NASA Astrophysics Data System (ADS)
Saabni, Raid M.; El-Sana, Jihad A.
2011-01-01
A large amount of handwritten historical documents are located in libraries around the world. The desire to access, search, and explore these documents paves the way for a new age of knowledge sharing and promotes collaboration and understanding between human societies. Currently, the indexes for these documents are generated manually, which is very tedious and time consuming. Results produced by state of the art techniques, for converting complete images of handwritten documents into textual representations, are not yet sufficient. Therefore, word-spotting methods have been developed to archive and index images of handwritten documents in order to enable efficient searching within documents. In this paper, we present a new matching algorithm to be used in word-spotting tasks for historical Arabic documents. We present a novel algorithm based on the Chamfer Distance to compute the similarity between shapes of word-parts. Matching results are used to cluster images of Arabic word-parts into different classes using the Nearest Neighbor rule. To compute the distance between two word-part images, the algorithm subdivides each image into equal-sized slices (windows). A modified version of the Chamfer Distance, incorporating geometric gradient features and distance transform data, is used as a similarity distance between the different slices. Finally, the Dynamic Time Warping (DTW) algorithm is used to measure the distance between two images of word-parts. By using the DTW we enabled our system to cluster similar word-parts, even though they are transformed non-linearly due to the nature of handwriting. We tested our implementation of the presented methods using various documents in different writing styles, taken from Juma'a Al Majid Center - Dubai, and obtained encouraging results.
NASA Technical Reports Server (NTRS)
1997-01-01
This report summarizes work done under Cooperative Agreement (CA) on the following testbed projects: TERRIERS - The development of the ground systems to support the TERRIERS satellite mission at Boston University (BU). HSTS - The application of ARC's Heuristic Scheduling Testbed System (HSTS) to the EUVE satellite mission. SELMON - The application of NASA's Jet Propulsion Laboratory's (JPL) Selective Monitoring (SELMON) system to the EUVE satellite mission. EVE - The development of the EUVE Virtual Environment (EVE), a prototype three-dimensional (3-D) visualization environment for the EUVE satellite and its sensors, instruments, and communications antennae. FIDO - The development of the Fault-Induced Document Officer (FIDO) system, a prototype application to respond to anomalous conditions by automatically searching for, retrieving, and displaying relevant documentation for an operators use.
EXTENSIBLE DATABASE FRAMEWORK FOR MANAGEMENT OF UNSTRUCTURED AND SEMI-STRUCTURED DOCUMENTS
NASA Technical Reports Server (NTRS)
Gawdiak, Yuri O. (Inventor); La, Tracy T. (Inventor); Lin, Shu-Chun Y. (Inventor); Malof, David A. (Inventor); Tran, Khai Peter B. (Inventor)
2005-01-01
Method and system for querying a collection of Unstructured or semi-structured documents to identify presence of, and provide context and/or content for, keywords and/or keyphrases. The documents are analyzed and assigned a node structure, including an ordered sequence of mutually exclusive node segments or strings. Each node has an associated set of at least four, five or six attributes with node information and can represent a format marker or text, with the last node in any node segment usually being a text node. A keyword (or keyphrase) is specified. and the last node in each node segment is searched for a match with the keyword. When a match is found at a query node, or at a node determined with reference to a query node, the system displays the context andor the content of the query node.
Intellectual property in holographic interferometry
NASA Astrophysics Data System (ADS)
Reingand, Nadya; Hunt, David
2006-08-01
This paper presents an overview of patents and patent applications on holographic interferometry, and highlights the possibilities offered by patent searching and analysis. Thousands of patent documents relevant to holographic interferometry were uncovered by the study. The search was performed in the following databases: U.S. Patent Office, European Patent Office, Japanese Patent Office and Korean Patent Office for the time frame from 1971 through May 2006. The patent analysis unveils trends in patent temporal distribution, patent families formation, significant technological coverage within the market of system that employ holographic interferometry and other interesting insights.
Large Scale Document Inversion using a Multi-threaded Computing System
Jung, Sungbo; Chang, Dar-Jen; Park, Juw Won
2018-01-01
Current microprocessor architecture is moving towards multi-core/multi-threaded systems. This trend has led to a surge of interest in using multi-threaded computing devices, such as the Graphics Processing Unit (GPU), for general purpose computing. We can utilize the GPU in computation as a massive parallel coprocessor because the GPU consists of multiple cores. The GPU is also an affordable, attractive, and user-programmable commodity. Nowadays a lot of information has been flooded into the digital domain around the world. Huge volume of data, such as digital libraries, social networking services, e-commerce product data, and reviews, etc., is produced or collected every moment with dramatic growth in size. Although the inverted index is a useful data structure that can be used for full text searches or document retrieval, a large number of documents will require a tremendous amount of time to create the index. The performance of document inversion can be improved by multi-thread or multi-core GPU. Our approach is to implement a linear-time, hash-based, single program multiple data (SPMD), document inversion algorithm on the NVIDIA GPU/CUDA programming platform utilizing the huge computational power of the GPU, to develop high performance solutions for document indexing. Our proposed parallel document inversion system shows 2-3 times faster performance than a sequential system on two different test datasets from PubMed abstract and e-commerce product reviews. CCS Concepts •Information systems➝Information retrieval • Computing methodologies➝Massively parallel and high-performance simulations. PMID:29861701
Large Scale Document Inversion using a Multi-threaded Computing System.
Jung, Sungbo; Chang, Dar-Jen; Park, Juw Won
2017-06-01
Current microprocessor architecture is moving towards multi-core/multi-threaded systems. This trend has led to a surge of interest in using multi-threaded computing devices, such as the Graphics Processing Unit (GPU), for general purpose computing. We can utilize the GPU in computation as a massive parallel coprocessor because the GPU consists of multiple cores. The GPU is also an affordable, attractive, and user-programmable commodity. Nowadays a lot of information has been flooded into the digital domain around the world. Huge volume of data, such as digital libraries, social networking services, e-commerce product data, and reviews, etc., is produced or collected every moment with dramatic growth in size. Although the inverted index is a useful data structure that can be used for full text searches or document retrieval, a large number of documents will require a tremendous amount of time to create the index. The performance of document inversion can be improved by multi-thread or multi-core GPU. Our approach is to implement a linear-time, hash-based, single program multiple data (SPMD), document inversion algorithm on the NVIDIA GPU/CUDA programming platform utilizing the huge computational power of the GPU, to develop high performance solutions for document indexing. Our proposed parallel document inversion system shows 2-3 times faster performance than a sequential system on two different test datasets from PubMed abstract and e-commerce product reviews. •Information systems➝Information retrieval • Computing methodologies➝Massively parallel and high-performance simulations.
Lossef, S V; Schwartz, L H
1990-09-01
A computerized reference system for radiology journal articles was developed by using an IBM-compatible personal computer with a hand-held optical scanner and optical character recognition software. This allows direct entry of scanned text from printed material into word processing or data-base files. Additionally, line diagrams and photographs of radiographs can be incorporated into these files. A text search and retrieval software program enables rapid searching for keywords in scanned documents. The hand scanner and software programs are commercially available, relatively inexpensive, and easily used. This permits construction of a personalized radiology literature file of readily accessible text and images requiring minimal typing or keystroke entry.
Smith, R F; Wiese, B A; Wojzynski, M K; Davison, D B; Worley, K C
1996-05-01
The BCM Search Launcher is an integrated set of World Wide Web (WWW) pages that organize molecular biology-related search and analysis services available on the WWW by function, and provide a single point of entry for related searches. The Protein Sequence Search Page, for example, provides a single sequence entry form for submitting sequences to WWW servers that offer remote access to a variety of different protein sequence search tools, including BLAST, FASTA, Smith-Waterman, BEAUTY, PROSITE, and BLOCKS searches. Other Launch pages provide access to (1) nucleic acid sequence searches, (2) multiple and pair-wise sequence alignments, (3) gene feature searches, (4) protein secondary structure prediction, and (5) miscellaneous sequence utilities (e.g., six-frame translation). The BCM Search Launcher also provides a mechanism to extend the utility of other WWW services by adding supplementary hypertext links to results returned by remote servers. For example, links to the NCBI's Entrez data base and to the Sequence Retrieval System (SRS) are added to search results returned by the NCBI's WWW BLAST server. These links provide easy access to auxiliary information, such as Medline abstracts, that can be extremely helpful when analyzing BLAST data base hits. For new or infrequent users of sequence data base search tools, we have preset the default search parameters to provide the most informative first-pass sequence analysis possible. We have also developed a batch client interface for Unix and Macintosh computers that allows multiple input sequences to be searched automatically as a background task, with the results returned as individual HTML documents directly to the user's system. The BCM Search Launcher and batch client are available on the WWW at URL http:@gc.bcm.tmc.edu:8088/search-launcher.html.
The use of a GIS Red-Amber-Green (RAG) system to define search priorities for burials
NASA Astrophysics Data System (ADS)
Somma, Roberta; Silvestro, Massimiliano; Cascio, Maria; Dawson, Lorna; Donnelly, Laurance; Harrison, Mark; McKinley, Jennifer; Ruffell, Alastair
2016-04-01
The aim of this research is to promote among the Italian police, magistrates, and geologists, the applications of a Geographical Information System (GIS)-based RAG system for use in ground searches for burials. To date the RAG system has not been used and documented in Italy and would potentially be useful for searches related to clandestine burial sites. This technique, was originally documented by the British Army in the 1st World War. The RAG method is based on the construction of theme maps. RAG maps can facilitate the deployment of appropriate search assets (such as geophysics, probe or search dogs) and therefore applied to ground searches for the potential location of homicide graves or other buried objects (including weapons, explosives, etc.). RAG maps also may assist in the management of resources such as the deployment of search personnel, search teams and dogs. A GIS RAG (Red-Amber-Green) system related to a search for a homicide grave was applied to a test site in Italy, simulating the concealment of a victim in the area of Alì. This is an area of hill in Sicily, characterized by Palaeozoic phyllites. It was assumed during this test that information was provided by an observer who saw a suspect carrying tools on his land during daylight hours. A desktop study of the rural area was first implemented. Data was collated from previous geological, geomorphological, hydrogeological, geophysical and land use surveys. All these data were stored and independently analysed in a GIS using ArcGIS software. For the development of the GIS-based RAG map a digital elevation model (DEM) including a digital surface model (DTS) and digital terrain model (DTM) types were used. These were integrated with data from soil surveys to provide a preliminary assessment of "diggability" - including the possible thickness of loose superficial deposits and soils. Data were stored in different layers within the GIS. These included the delineation of the search area with consideration of access/exit points, diggability (easy: red, difficult: green), ground slope (<27°: red, >27°: green), vegetation type (easy access: red, difficult access: green), geomorphology (stable area: red, unstable area: green), anthropogenic structures (not present: red, present: green), visibility of the site from a potential eyewitnesses perspective (not visible: red, visible: green). Overlaying these layers, using the ArcGIS tools, enabled the RAG map to be composed with red showing the high priority search areas, amber the intermediate priority search areas and green the low priority search areas. The GIS-based RAG map of the simulated test-site allowed the original extent of the search area of 39.315m2, to be significantly reduced to 7.45% (2.930m2: extension red area) by desktop study and to 2.93% (1.152m2) with a further reconnaissance site visit. During subsequent field training conducted by forensic geology students at Messina University, the grave was found after 2 hours of searching, both using the RAG map and a soil probe and observing topographic disturbances. A subsidence of some centimeters and an anomalous growth of vegetation was found on the superficial surface of the grave (75cm deep).
Guided Text Search Using Adaptive Visual Analytics
DOE Office of Scientific and Technical Information (OSTI.GOV)
Steed, Chad A; Symons, Christopher T; Senter, James K
This research demonstrates the promise of augmenting interactive visualizations with semi- supervised machine learning techniques to improve the discovery of significant associations and insights in the search and analysis of textual information. More specifically, we have developed a system called Gryffin that hosts a unique collection of techniques that facilitate individualized investigative search pertaining to an ever-changing set of analytical questions over an indexed collection of open-source documents related to critical national infrastructure. The Gryffin client hosts dynamic displays of the search results via focus+context record listings, temporal timelines, term-frequency views, and multiple coordinate views. Furthermore, as the analyst interactsmore » with the display, the interactions are recorded and used to label the search records. These labeled records are then used to drive semi-supervised machine learning algorithms that re-rank the unlabeled search records such that potentially relevant records are moved to the top of the record listing. Gryffin is described in the context of the daily tasks encountered at the US Department of Homeland Security s Fusion Center, with whom we are collaborating in its development. The resulting system is capable of addressing the analysts information overload that can be directly attributed to the deluge of information that must be addressed in the search and investigative analysis of textual information.« less
Sampling criteria in multicollection searching.
NASA Astrophysics Data System (ADS)
Gilio, A.; Scozzafava, R.; Marchetti, P. G.
In the first stage of the document retrieval process, no information concerning relevance of a particular document is available. On the other hand, computer implementation requires that the analysis be made only for a sample of retrieved documents. This paper addresses the significance and suitability of two different sampling criteria for a multicollection online search facility. The inevitability of resorting to a logarithmic criterion in order to achieve a "spread of representativeness" from the multicollection is demonstrated.
TSAFE Interface Control Document v 2.0
NASA Technical Reports Server (NTRS)
Paielli, Russell A.; Bach, Ralph E.
2013-01-01
This document specifies the data interface for TSAFE, the Tactical Separation-Assured Flight Environment. TSAFE is a research prototype of a software application program for alerting air traffic controllers to imminent conflicts in enroute airspace. It is intended for Air Route Traffic Control Centers ("Centers") in the U.S. National Airspace System. It predicts trajectories for approximately 3 minutes into the future, searches for conflicts, and sends data about predicted conflicts to the client, which uses the data to alert an air traffic controller of conflicts. TSAFE itself does not provide a graphical user interface.
Methods for Documenting Systematic Review Searches: A Discussion of Common Issues
ERIC Educational Resources Information Center
Rader, Tamara; Mann, Mala; Stansfield, Claire; Cooper, Chris; Sampson, Margaret
2014-01-01
Introduction: As standardized reporting requirements for systematic reviews are being adopted more widely, review authors are under greater pressure to accurately record their search process. With careful planning, documentation to fulfill the Preferred Reporting Items for Systematic Reviews and Meta-Analyses requirements can become a valuable…
Recommendations for a Habitability Data Base.
ERIC Educational Resources Information Center
Illinois Univ., Urbana. Library Research Center.
A prototype Habitability Data Base was developed for the United States Army Corps of Engineers. From a review of selected Army documents, standards in the form of goals or architectural criteria were identified as significant to man-environment relations (MER). A search of appropriate information systems was conducted to retrieve a minimum of 500…
77 FR 784 - Statewide Longitudinal Data Systems; Reopening Fiscal Year (FY) 2012 Competition
Federal Register 2010, 2011, 2012, 2013, 2014
2012-01-06
... following address: U.S. Department of Education, Application Control Center, Attention: (CFDA Number: 84... Center, Attention: (CFDA Number: 84.372A), 550 12th Street SW., Room 7041, Potomac Center Plaza... also access documents of the Department published in the Federal Register by using the article search...
Documenting the Conversation: A Systematic Review of Library Discovery Layers
ERIC Educational Resources Information Center
Bossaller, Jenny S.; Sandy, Heather Moulaison
2017-01-01
This article describes the results of a systematic review of peer-reviewed, published research articles about "discovery layers," user-friendly interfaces or systems that provide single-search box access to library content. Focusing on articles in LISTA published 2009-2013, a set of 80 articles was coded for community of users, journal…
48 CFR 31.205-30 - Patent costs.
Code of Federal Regulations, 2012 CFR
2012-10-01
... 48 Federal Acquisition Regulations System 1 2012-10-01 2012-10-01 false Patent costs. 31.205-30....205-30 Patent costs. (a) The following patent costs are allowable to the extent that they are incurred... disclosures, reports, and other documents. (2) Costs for searching the art to the extent necessary to make the...
48 CFR 31.205-30 - Patent costs.
Code of Federal Regulations, 2013 CFR
2013-10-01
... 48 Federal Acquisition Regulations System 1 2013-10-01 2013-10-01 false Patent costs. 31.205-30....205-30 Patent costs. (a) The following patent costs are allowable to the extent that they are incurred... disclosures, reports, and other documents. (2) Costs for searching the art to the extent necessary to make the...
48 CFR 31.205-30 - Patent costs.
Code of Federal Regulations, 2011 CFR
2011-10-01
... 48 Federal Acquisition Regulations System 1 2011-10-01 2011-10-01 false Patent costs. 31.205-30....205-30 Patent costs. (a) The following patent costs are allowable to the extent that they are incurred... disclosures, reports, and other documents. (2) Costs for searching the art to the extent necessary to make the...
48 CFR 31.205-30 - Patent costs.
Code of Federal Regulations, 2010 CFR
2010-10-01
... 48 Federal Acquisition Regulations System 1 2010-10-01 2010-10-01 false Patent costs. 31.205-30....205-30 Patent costs. (a) The following patent costs are allowable to the extent that they are incurred... disclosures, reports, and other documents. (2) Costs for searching the art to the extent necessary to make the...
48 CFR 31.205-30 - Patent costs.
Code of Federal Regulations, 2014 CFR
2014-10-01
... 48 Federal Acquisition Regulations System 1 2014-10-01 2014-10-01 false Patent costs. 31.205-30....205-30 Patent costs. (a) The following patent costs are allowable to the extent that they are incurred... disclosures, reports, and other documents. (2) Costs for searching the art to the extent necessary to make the...
Sentence-Based Metadata: An Approach and Tool for Viewing Database Designs.
ERIC Educational Resources Information Center
Boyle, John M.; Gunge, Jakob; Bryden, John; Librowski, Kaz; Hanna, Hsin-Yi
2002-01-01
Describes MARS (Museum Archive Retrieval System), a research tool which enables organizations to exchange digital images and documents by means of a common thesaurus structure, and merge the descriptive data and metadata of their collections. Highlights include theoretical basis; searching the MARS database; and examples in European museums.…
Computer-Based Education. The Best of ERIC, June 1976-August 1980.
ERIC Educational Resources Information Center
Hall, Keith A.
This bibliography contains annotations of reports, reviews, conference proceedings, other documents, and journal articles on computer based education (CBE), most of which were derived from a search of the Educational Resources Information Center (ERIC) system. Covering June 1976 through August 1980, this compilation serves as an update to two…
BIOZON: a system for unification, management and analysis of heterogeneous biological data.
Birkland, Aaron; Yona, Golan
2006-02-15
Integration of heterogeneous data types is a challenging problem, especially in biology, where the number of databases and data types increase rapidly. Amongst the problems that one has to face are integrity, consistency, redundancy, connectivity, expressiveness and updatability. Here we present a system (Biozon) that addresses these problems, and offers biologists a new knowledge resource to navigate through and explore. Biozon unifies multiple biological databases consisting of a variety of data types (such as DNA sequences, proteins, interactions and cellular pathways). It is fundamentally different from previous efforts as it uses a single extensive and tightly connected graph schema wrapped with hierarchical ontology of documents and relations. Beyond warehousing existing data, Biozon computes and stores novel derived data, such as similarity relationships and functional predictions. The integration of similarity data allows propagation of knowledge through inference and fuzzy searches. Sophisticated methods of query that span multiple data types were implemented and first-of-a-kind biological ranking systems were explored and integrated. The Biozon system is an extensive knowledge resource of heterogeneous biological data. Currently, it holds more than 100 million biological documents and 6.5 billion relations between them. The database is accessible through an advanced web interface that supports complex queries, "fuzzy" searches, data materialization and more, online at http://biozon.org.
A novel architecture for information retrieval system based on semantic web
NASA Astrophysics Data System (ADS)
Zhang, Hui
2011-12-01
Nowadays, the web has enabled an explosive growth of information sharing (there are currently over 4 billion pages covering most areas of human endeavor) so that the web has faced a new challenge of information overhead. The challenge that is now before us is not only to help people locating relevant information precisely but also to access and aggregate a variety of information from different resources automatically. Current web document are in human-oriented formats and they are suitable for the presentation, but machines cannot understand the meaning of document. To address this issue, Berners-Lee proposed a concept of semantic web. With semantic web technology, web information can be understood and processed by machine. It provides new possibilities for automatic web information processing. A main problem of semantic web information retrieval is that when these is not enough knowledge to such information retrieval system, the system will return to a large of no sense result to uses due to a huge amount of information results. In this paper, we present the architecture of information based on semantic web. In addiction, our systems employ the inference Engine to check whether the query should pose to Keyword-based Search Engine or should pose to the Semantic Search Engine.
DOE Office of Scientific and Technical Information (OSTI.GOV)
IRIS is a search tool plug-in that is used to implement latent topic feedback for enhancing text navigation. It accepts a list of returned documents from an information retrieval wywtem that is generated from keyword search queries. Data is pulled directly from a topic information database and processed by IRIS to determine the most prominent and relevant topics, along with topic-ngrams, associated with the list of returned documents. User selected topics are then used to expand the query and presumabley refine the search results.
Health search engine with e-document analysis for reliable search results.
Gaudinat, Arnaud; Ruch, Patrick; Joubert, Michel; Uziel, Philippe; Strauss, Anne; Thonnet, Michèle; Baud, Robert; Spahni, Stéphane; Weber, Patrick; Bonal, Juan; Boyer, Celia; Fieschi, Marius; Geissbuhler, Antoine
2006-01-01
After a review of the existing practical solution available to the citizen to retrieve eHealth document, the paper describes an original specialized search engine WRAPIN. WRAPIN uses advanced cross lingual information retrieval technologies to check information quality by synthesizing medical concepts, conclusions and references contained in the health literature, to identify accurate, relevant sources. Thanks to MeSH terminology [1] (Medical Subject Headings from the U.S. National Library of Medicine) and advanced approaches such as conclusion extraction from structured document, reformulation of the query, WRAPIN offers to the user a privileged access to navigate through multilingual documents without language or medical prerequisites. The results of an evaluation conducted on the WRAPIN prototype show that results of the WRAPIN search engine are perceived as informative 65% (59% for a general-purpose search engine), reliable and trustworthy 72% (41% for the other engine) by users. But it leaves room for improvement such as the increase of database coverage, the explanation of the original functionalities and an audience adaptability. Thanks to evaluation outcomes, WRAPIN is now in exploitation on the HON web site (http://www.healthonnet.org), free of charge. Intended to the citizen it is a good alternative to general-purpose search engines when the user looks up trustworthy health and medical information or wants to check automatically a doubtful content of a Web page.
Improving PHENIX search with Solr, Nutch and Drupal.
NASA Astrophysics Data System (ADS)
Morrison, Dave; Sourikova, Irina
2012-12-01
During its 20 years of R&D, construction and operation the PHENIX experiment at the Relativistic Heavy Ion Collider (RHIC) has accumulated large amounts of proprietary collaboration data that is hosted on many servers around the world and is not open for commercial search engines for indexing and searching. The legacy search infrastructure did not scale well with the fast growing PHENIX document base and produced results inadequate in both precision and recall. After considering the possible alternatives that would provide an aggregated, fast, full text search of a variety of data sources and file formats we decided to use Nutch [1] as a web crawler and Solr [2] as a search engine. To present XML-based Solr search results in a user-friendly format we use Drupal [3] as a web interface to Solr. We describe the experience of building a federated search for a heterogeneous collection of 10 million PHENIX documents with Nutch, Solr and Drupal.
Three-dimensional measurement system for crime scene documentation
NASA Astrophysics Data System (ADS)
Adamczyk, Marcin; Hołowko, Elwira; Lech, Krzysztof; Michoński, Jakub; MÄ czkowski, Grzegorz; Bolewicki, Paweł; Januszkiewicz, Kamil; Sitnik, Robert
2017-10-01
Three dimensional measurements (such as photogrammetry, Time of Flight, Structure from Motion or Structured Light techniques) are becoming a standard in the crime scene documentation process. The usage of 3D measurement techniques provide an opportunity to prepare more insightful investigation and helps to show every trace in the context of the entire crime scene. In this paper we would like to present a hierarchical, three-dimensional measurement system that is designed for crime scenes documentation process. Our system reflects the actual standards in crime scene documentation process - it is designed to perform measurement in two stages. First stage of documentation, the most general, is prepared with a scanner with relatively low spatial resolution but also big measuring volume - it is used for the whole scene documentation. Second stage is much more detailed: high resolution but smaller size of measuring volume for areas that required more detailed approach. The documentation process is supervised by a specialised application CrimeView3D, that is a software platform for measurements management (connecting with scanners and carrying out measurements, automatic or semi-automatic data registration in the real time) and data visualisation (3D visualisation of documented scenes). It also provides a series of useful tools for forensic technicians: virtual measuring tape, searching for sources of blood spatter, virtual walk on the crime scene and many others. In this paper we present our measuring system and the developed software. We also provide an outcome from research on metrological validation of scanners that was performed according to VDI/VDE standard. We present a CrimeView3D - a software-platform that was developed to manage the crime scene documentation process. We also present an outcome from measurement sessions that were conducted on real crime scenes with cooperation with Technicians from Central Forensic Laboratory of Police.
JSC Search System Usability Case Study
NASA Technical Reports Server (NTRS)
Meza, David; Berndt, Sarah
2014-01-01
The advanced nature of "search" has facilitated the movement from keyword match to the delivery of every conceivable information topic from career, commerce, entertainment, learning... the list is infinite. At NASA Johnson Space Center (JSC ) the Search interface is an important means of knowledge transfer. By indexing multiple sources between directorates and organizations, the system's potential is culture changing in that through search, knowledge of the unique accomplishments in engineering and science can be seamlessly passed between generations. This paper reports the findings of an initial survey, the first of a four part study to help determine user sentiment on the intranet, or local (JSC) enterprise search environment as well as the larger NASA enterprise. The survey is a means through which end users provide direction on the development and transfer of knowledge by way of the search experience. The ideal is to identify what is working and what needs to be improved from the users' vantage point by documenting: (1) Where users are satisfied/dissatisfied (2) Perceived value of interface components (3) Gaps which cause any disappointment in search experience. The near term goal is it to inform JSC search in order to improve users' ability to utilize existing services and infrastructure to perform tasks with a shortened life cycle. Continuing steps include an agency based focus with modified questions to accomplish a similar purpose
Concept Based Tie-breaking and Maximal Marginal Relevance Retrieval in Microblog Retrieval
2014-11-01
the same score, another singal will be used to rank these documents to break the ties , but the relative orders of other documents against these...documents remain the same. The tie- breaking step above is repeatedly applied to further break ties until all candidate signals are applied and the ranking...searched it on the Yahoo! search engine, which returned some query sug- gestions for the query. The original queries as well as their query suggestions
Luo, Jake; Chen, Weiheng; Wu, Min; Weng, Chunhua
2018-01-01
Background Prior studies of clinical trial planning indicate that it is crucial to search and screen recruitment sites before starting to enroll participants. However, currently there is no systematic method developed to support clinical investigators to search candidate recruitment sites according to their interested clinical trial factors. Objective In this study, we aim at developing a new approach to integrating the location data of over one million heterogeneous recruitment sites that are stored in clinical trial documents. The integrated recruitment location data can be searched and visualized using a map-based information retrieval method. The method enables systematic search and analysis of recruitment sites across a large amount of clinical trials. Methods The location data of more than 1.4 million recruitment sites of over 183,000 clinical trials was normalized and integrated using a geocoding method. The integrated data can be used to support geographic information retrieval of recruitment sites. Additionally, the information of over 6000 clinical trial target disease conditions and close to 4000 interventions was also integrated into the system and linked to the recruitment locations. Such data integration enabled the construction of a novel map-based query system. The system will allow clinical investigators to search and visualize candidate recruitment sites for clinical trials based on target conditions and interventions. Results The evaluation results showed that the coverage of the geographic location mapping for the 1.4 million recruitment sites was 99.8%. The evaluation of 200 randomly retrieved recruitment sites showed that the correctness of geographic information mapping was 96.5%. The recruitment intensities of the top 30 countries were also retrieved and analyzed. The data analysis results indicated that the recruitment intensity varied significantly across different countries and geographic areas. Conclusion This study contributed a new data processing framework to extract and integrate the location data of heterogeneous recruitment sites from clinical trial documents. The developed system can support effective retrieval and analysis of potential recruitment sites using target clinical trial factors. PMID:29132636
Luo, Jake; Chen, Weiheng; Wu, Min; Weng, Chunhua
2017-12-01
Prior studies of clinical trial planning indicate that it is crucial to search and screen recruitment sites before starting to enroll participants. However, currently there is no systematic method developed to support clinical investigators to search candidate recruitment sites according to their interested clinical trial factors. In this study, we aim at developing a new approach to integrating the location data of over one million heterogeneous recruitment sites that are stored in clinical trial documents. The integrated recruitment location data can be searched and visualized using a map-based information retrieval method. The method enables systematic search and analysis of recruitment sites across a large amount of clinical trials. The location data of more than 1.4 million recruitment sites of over 183,000 clinical trials was normalized and integrated using a geocoding method. The integrated data can be used to support geographic information retrieval of recruitment sites. Additionally, the information of over 6000 clinical trial target disease conditions and close to 4000 interventions was also integrated into the system and linked to the recruitment locations. Such data integration enabled the construction of a novel map-based query system. The system will allow clinical investigators to search and visualize candidate recruitment sites for clinical trials based on target conditions and interventions. The evaluation results showed that the coverage of the geographic location mapping for the 1.4 million recruitment sites was 99.8%. The evaluation of 200 randomly retrieved recruitment sites showed that the correctness of geographic information mapping was 96.5%. The recruitment intensities of the top 30 countries were also retrieved and analyzed. The data analysis results indicated that the recruitment intensity varied significantly across different countries and geographic areas. This study contributed a new data processing framework to extract and integrate the location data of heterogeneous recruitment sites from clinical trial documents. The developed system can support effective retrieval and analysis of potential recruitment sites using target clinical trial factors. Copyright © 2017 Elsevier B.V. All rights reserved.
Full Text and Figure Display Improves Bioscience Literature Search
Divoli, Anna; Wooldridge, Michael A.; Hearst, Marti A.
2010-01-01
When reading bioscience journal articles, many researchers focus attention on the figures and their captions. This observation led to the development of the BioText literature search engine [1], a freely available Web-based application that allows biologists to search over the contents of Open Access Journals, and see figures from the articles displayed directly in the search results. This article presents a qualitative assessment of this system in the form of a usability study with 20 biologist participants using and commenting on the system. 19 out of 20 participants expressed a desire to use a bioscience literature search engine that displays articles' figures alongside the full text search results. 15 out of 20 participants said they would use a caption search and figure display interface either frequently or sometimes, while 4 said rarely and 1 said undecided. 10 out of 20 participants said they would use a tool for searching the text of tables and their captions either frequently or sometimes, while 7 said they would use it rarely if at all, 2 said they would never use it, and 1 was undecided. This study found evidence, supporting results of an earlier study, that bioscience literature search systems such as PubMed should show figures from articles alongside search results. It also found evidence that full text and captions should be searched along with the article title, metadata, and abstract. Finally, for a subset of users and information needs, allowing for explicit search within captions for figures and tables is a useful function, but it is not entirely clear how to cleanly integrate this within a more general literature search interface. Such a facility supports Open Access publishing efforts, as it requires access to full text of documents and the lifting of restrictions in order to show figures in the search interface. PMID:20418942
Stansfield, Claire; O'Mara-Eves, Alison; Thomas, James
2017-09-01
Using text mining to aid the development of database search strings for topics described by diverse terminology has potential benefits for systematic reviews; however, methods and tools for accomplishing this are poorly covered in the research methods literature. We briefly review the literature on applications of text mining for search term development for systematic reviewing. We found that the tools can be used in 5 overarching ways: improving the precision of searches; identifying search terms to improve search sensitivity; aiding the translation of search strategies across databases; searching and screening within an integrated system; and developing objectively derived search strategies. Using a case study and selected examples, we then reflect on the utility of certain technologies (term frequency-inverse document frequency and Termine, term frequency, and clustering) in improving the precision and sensitivity of searches. Challenges in using these tools are discussed. The utility of these tools is influenced by the different capabilities of the tools, the way the tools are used, and the text that is analysed. Increased awareness of how the tools perform facilitates the further development of methods for their use in systematic reviews. Copyright © 2017 John Wiley & Sons, Ltd.
BioEve Search: A Novel Framework to Facilitate Interactive Literature Search
Ahmed, Syed Toufeeq; Davulcu, Hasan; Tikves, Sukru; Nair, Radhika; Zhao, Zhongming
2012-01-01
Background. Recent advances in computational and biological methods in last two decades have remarkably changed the scale of biomedical research and with it began the unprecedented growth in both the production of biomedical data and amount of published literature discussing it. An automated extraction system coupled with a cognitive search and navigation service over these document collections would not only save time and effort, but also pave the way to discover hitherto unknown information implicitly conveyed in the texts. Results. We developed a novel framework (named “BioEve”) that seamlessly integrates Faceted Search (Information Retrieval) with Information Extraction module to provide an interactive search experience for the researchers in life sciences. It enables guided step-by-step search query refinement, by suggesting concepts and entities (like genes, drugs, and diseases) to quickly filter and modify search direction, and thereby facilitating an enriched paradigm where user can discover related concepts and keywords to search while information seeking. Conclusions. The BioEve Search framework makes it easier to enable scalable interactive search over large collection of textual articles and to discover knowledge hidden in thousands of biomedical literature articles with ease. PMID:22693501
UnCover on the Web: search hints and applications in library environments.
Galpern, N F; Albert, K M
1997-01-01
Among the huge maze of resources available on the Internet, UnCoverWeb stands out as a valuable tool for medical libraries. This up-to-date, free-access, multidisciplinary database of periodical references is searched through an easy-to-learn graphical user interface that is a welcome improvement over the telnet version. This article reviews the basic and advanced search techniques for UnCoverWeb, as well as providing information on the document delivery functions and table of contents alerting service called Reveal. UnCover's currency is evaluated and compared with other current awareness resources. System deficiencies are discussed, with the conclusion that although UnCoverWeb lacks the sophisticated features of many commercial database search services, it is nonetheless a useful addition to the repertoire of information sources available in a library.
Image and information management system
NASA Technical Reports Server (NTRS)
Robertson, Tina L. (Inventor); Raney, Michael C. (Inventor); Dougherty, Dennis M. (Inventor); Kent, Peter C. (Inventor); Brucker, Russell X. (Inventor); Lampert, Daryl A. (Inventor)
2009-01-01
A system and methods through which pictorial views of an object's configuration, arranged in a hierarchical fashion, are navigated by a person to establish a visual context within the configuration. The visual context is automatically translated by the system into a set of search parameters driving retrieval of structured data and content (images, documents, multimedia, etc.) associated with the specific context. The system places ''hot spots'', or actionable regions, on various portions of the pictorials representing the object. When a user interacts with an actionable region, a more detailed pictorial from the hierarchy is presented representing that portion of the object, along with real-time feedback in the form of a popup pane containing information about that region, and counts-by-type reflecting the number of items that are available within the system associated with the specific context and search filters established at that point in time.
Image and information management system
NASA Technical Reports Server (NTRS)
Robertson, Tina L. (Inventor); Kent, Peter C. (Inventor); Raney, Michael C. (Inventor); Dougherty, Dennis M. (Inventor); Brucker, Russell X. (Inventor); Lampert, Daryl A. (Inventor)
2007-01-01
A system and methods through which pictorial views of an object's configuration, arranged in a hierarchical fashion, are navigated by a person to establish a visual context within the configuration. The visual context is automatically translated by the system into a set of search parameters driving retrieval of structured data and content (images, documents, multimedia, etc.) associated with the specific context. The system places hot spots, or actionable regions, on various portions of the pictorials representing the object. When a user interacts with an actionable region, a more detailed pictorial from the hierarchy is presented representing that portion of the object, along with real-time feedback in the form of a popup pane containing information about that region, and counts-by-type reflecting the number of items that are available within the system associated with the specific context and search filters established at that point in time.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Abe Lederman
This report contains the comprehensive summary of the work performed on the SBIR Phase II project (“Distributed Relevance Ranking in Heterogeneous Document Collections”) at Deep Web Technologies (http://www.deepwebtech.com). We have successfully completed all of the tasks defined in our SBIR Proposal work plan (See Table 1 - Phase II Tasks Status). The project was completed on schedule and we have successfully deployed an initial production release of the software architecture at DOE-OSTI for the Science.gov Alliance's search portal (http://www.science.gov). We have implemented a set of grid services that supports the extraction, filtering, aggregation, and presentation of search results from numerousmore » heterogeneous document collections. Illustration 3 depicts the services required to perform QuickRank™ filtering of content as defined in our architecture documentation. Functionality that has been implemented is indicated by the services highlighted in green. We have successfully tested our implementation in a multi-node grid deployment both within the Deep Web Technologies offices, and in a heterogeneous geographically distributed grid environment. We have performed a series of load tests in which we successfully simulated 100 concurrent users submitting search requests to the system. This testing was performed on deployments of one, two, and three node grids with services distributed in a number of different configurations. The preliminary results from these tests indicate that our architecture will scale well across multi-node grid deployments, but more work will be needed, beyond the scope of this project, to perform testing and experimentation to determine scalability and resiliency requirements. We are pleased to report that a production quality version (1.4) of the science.gov Alliance's search portal based on our grid architecture was released in June of 2006. This demonstration portal is currently available at http://science.gov/search30 . The portal allows the user to select from a number of collections grouped by category and enter a query expression (See Illustration 1 - Science.gov 3.0 Search Page). After the user clicks “search” a results page is displayed that provides a list of results from the selected collections ordered by relevance based on the query expression the user provided. Our grid based solution to deep web search and document ranking has already gained attention within DOE, other Government Agencies and a fortune 50 company. We are committed to the continued development of grid based solutions to large scale data access, filtering, and presentation problems within the domain of Information Retrieval and the more general categories of content management, data mining and data analysis.« less
Effects of Three Different Hypertextual Menu Designs on Various Information Searching Activities.
ERIC Educational Resources Information Center
Lai, Yee-Rong, Waugh, Michael L.
1995-01-01
Examines the influence of three combinations of document structures and menu designs on users' attitude, performance, and learning in search tasks. Three versions of a document introducing UNIX, Internet, BITNET, and general computing and networking resources at a university were implemented using HyperCard. Results showed that only the accuracy…
Federal Register 2010, 2011, 2012, 2013, 2014
2012-08-15
....regulations.gov . You can search for the document by selecting ``Notice'' under Document Type, entering the... ``Search.'' If necessary, use the ``Narrow by Agency'' option on the Results page. Email: [email protected] the burden of the proposed collection, including the validity of the methodology and assumptions used...
Federal Register 2010, 2011, 2012, 2013, 2014
2013-03-12
... NUCLEAR REGULATORY COMMISSION [NRC-2013-0047] Compendium of Analyses To Investigate Select Level 1...) has issued for public comment a document entitled: Compendium of Analyses to Investigate Select Level... begin the search, select ``ADAMS Public Documents'' and then select ``Begin Web- based ADAMS Search...
Computerized training management system
Rice, H.B.; McNair, R.C.; White, K.; Maugeri, T.
1998-08-04
A Computerized Training Management System (CTMS) is disclosed for providing a procedurally defined process that is employed to develop accreditable performance based training programs for job classifications that are sensitive to documented regulations and technical information. CTMS is a database that links information needed to maintain a five-phase approach to training-analysis, design, development, implementation, and evaluation independent of training program design. CTMS is designed using R-Base{trademark}, an-SQL compliant software platform. Information is logically entered and linked in CTMS. Each task is linked directly to a performance objective, which, in turn, is linked directly to a learning objective; then, each enabling objective is linked to its respective test items. In addition, tasks, performance objectives, enabling objectives, and test items are linked to their associated reference documents. CTMS keeps all information up to date since it automatically sorts, files and links all data; CTMS includes key word and reference document searches. 18 figs.
Computerized training management system
Rice, Harold B.; McNair, Robert C.; White, Kenneth; Maugeri, Terry
1998-08-04
A Computerized Training Management System (CTMS) for providing a procedurally defined process that is employed to develop accreditable performance based training programs for job classifications that are sensitive to documented regulations and technical information. CTMS is a database that links information needed to maintain a five-phase approach to training-analysis, design, development, implementation, and evaluation independent of training program design. CTMS is designed using R-Base.RTM., an-SQL compliant software platform. Information is logically entered and linked in CTMS. Each task is linked directly to a performance objective, which, in turn, is linked directly to a learning objective; then, each enabling objective is linked to its respective test items. In addition, tasks, performance objectives, enabling objectives, and test items are linked to their associated reference documents. CTMS keeps all information up to date since it automatically sorts, files and links all data; CTMS includes key word and reference document searches.
Office of Diversity Management and Equal Opportunity
Skip to main content (Press Enter). Toggle navigation Office of Diversity Management and Equal Opportunity Search Search ODMEO: Search Search ODMEO: Search Office of Diversity Management and Equal Opportunity Office of Diversity Management and Equal Opportunity Home About ODMEO Leadership Documents News
Usability/Sentiment for the Enterprise and ENTERPRISE
NASA Technical Reports Server (NTRS)
Meza, David; Berndt, Sarah
2014-01-01
The purpose of the Sentiment of Search Study for NASA Johnson Space Center (JSC) is to gain insight into the intranet search environment. With an initial usability survey, the authors were able to determine a usability score based on the Systems Usability Scale (SUS). Created in 1986, the freely available, well cited, SUS is commonly used to determine user perceptions of a system (in this case the intranet search environment). As with any improvement initiative, one must first examine and document the current reality of the situation. In this scenario, a method was needed to determine the usability of a search interface in addition to the user's perception on how well the search system was providing results. The use of the SUS provided a mechanism to quickly ascertain information in both areas, by adding one additional open-ended question at the end. The first ten questions allowed us to examine the usability of the system, while the last questions informed us on how the users rated the performance of the search results. The final analysis provides us with a better understanding of the current situation and areas to focus on for improvement. The power of search applications to enhance knowledge transfer is indisputable. The performance impact for any user unable to find needed information undermines project lifecycle, resource and scheduling requirements. Ever-increasing complexity of content and the user interface make usability considerations for the intranet, especially for search, a necessity instead of a 'nice-to-have'. Despite these arguments, intranet usability is largely disregarded due to lack of attention beyond the functionality of the infrastructure (White, 2013). The data collected from users of the JSC search system revealed their overall sentiment by means of the widely-known System Usability Scale. Results of the scores suggest 75%, +/-0.04, of the population rank the search system below average. In terms of a grading scaled, this equated to D or lower. It is obvious JSC users are not satisfied with the current situation, however they are eager to provide information and assistance in improving the search system. A majority of the respondents provided feedback on the issues most troubling them. This information will be used to enrich the next phase, root cause analysis and solution creation.
Ansari, Sereena; Boyle, Adrian
2017-02-01
Domestic abuse represents a serious public health and human rights concern. Interventions to reduce the risk of abuse include staff training and standardized documentation improving detection and adherence to referral pathways. Interventional studies have been conducted in primary care, maternity and outpatient settings. Women disclosing abuse in emergency departments differ from women attending other healthcare settings, and it is unclear whether these interventions can be transferred to the emergency care setting. This review examines interventional studies to evaluate the effectiveness of emergency department-based interventions in reducing domestic abuse-related morbidity. Medline, EMBASE, CINAHL, PsycINFO and Cochrane Library were searched, according to prespecified selection criteria. Study quality was assessed using the Jadad scale. Of 273 search results, nine were eligible for review. Interventions involving staff training demonstrated benefits in subjective measures, such as staff knowledge regarding abuse, but no changes in clinical practice, based on detection and referral rates. When staff training was implemented in conjunction with supporting system changes - for example, standardized documentation for assessment and referral - clinically relevant improvements were noted. Interventions centred around staff training are insufficient to bring about improvements in the management and, thus, outcome of patients suffering abuse. Instead, system changes, such as standardized documentation and referral pathways, supported by training, may bring about beneficial changes. It remains uncertain whether surrogate outcomes employed by most studies translate to changes in abuse-related morbidity: the ultimate goal.
Dynamic "inline" images: context-sensitive retrieval and integration of images into Web documents.
Kahn, Charles E
2008-09-01
Integrating relevant images into web-based information resources adds value for research and education. This work sought to evaluate the feasibility of using "Web 2.0" technologies to dynamically retrieve and integrate pertinent images into a radiology web site. An online radiology reference of 1,178 textual web documents was selected as the set of target documents. The ARRS GoldMiner image search engine, which incorporated 176,386 images from 228 peer-reviewed journals, retrieved images on demand and integrated them into the documents. At least one image was retrieved in real-time for display as an "inline" image gallery for 87% of the web documents. Each thumbnail image was linked to the full-size image at its original web site. Review of 20 randomly selected Collaborative Hypertext of Radiology documents found that 69 of 72 displayed images (96%) were relevant to the target document. Users could click on the "More" link to search the image collection more comprehensively and, from there, link to the full text of the article. A gallery of relevant radiology images can be inserted easily into web pages on any web server. Indexing by concepts and keywords allows context-aware image retrieval, and searching by document title and subject metadata yields excellent results. These techniques allow web developers to incorporate easily a context-sensitive image gallery into their documents.
Federal Register 2010, 2011, 2012, 2013, 2014
2012-09-19
... to reduce poverty; a system to combat corruption and bribery; and protection of internationally... ``Document Type'' on search- results page and click on the link entitled ``Submit a Comment.'' (For further... site provides the option of making submissions by filling in a ``Type Comment'' field, or by attaching...
Federal Register 2010, 2011, 2012, 2013, 2014
2011-09-22
... policies to reduce poverty; a system to combat corruption and bribery; and protection of internationally... selecting ``notice'' under ``Document Type'' on search- results page and click on the link entitled ``Submit... http://www.regulations.gov Web site provides the option of making submissions by filling in a ``Type...
78 FR 37927 - List of Approved Spent Fuel Storage Casks: MAGNASTOR® System
Federal Register 2010, 2011, 2012, 2013, 2014
2013-06-25
...;Prices of new books are listed in the first FEDERAL REGISTER issue of each #0;week. #0; #0; #0; #0;#0... ADAMS Search.'' For problems with ADAMS, please contact the NRC's Public Document Room (PDR) reference... not have a significant economic impact on a substantial number of small entities. This final rule...
Text Mining in Biomedical Domain with Emphasis on Document Clustering.
Renganathan, Vinaitheerthan
2017-07-01
With the exponential increase in the number of articles published every year in the biomedical domain, there is a need to build automated systems to extract unknown information from the articles published. Text mining techniques enable the extraction of unknown knowledge from unstructured documents. This paper reviews text mining processes in detail and the software tools available to carry out text mining. It also reviews the roles and applications of text mining in the biomedical domain. Text mining processes, such as search and retrieval of documents, pre-processing of documents, natural language processing, methods for text clustering, and methods for text classification are described in detail. Text mining techniques can facilitate the mining of vast amounts of knowledge on a given topic from published biomedical research articles and draw meaningful conclusions that are not possible otherwise.
NASA Strategic Roadmap Summary Report
NASA Technical Reports Server (NTRS)
Wilson, Scott; Bauer, Frank; Stetson, Doug; Robey, Judee; Smith, Eric P.; Capps, Rich; Gould, Dana; Tanner, Mike; Guerra, Lisa; Johnston, Gordon
2005-01-01
In response to the Vision, NASA commissioned strategic and capability roadmap teams to develop the pathways for turning the Vision into a reality. The strategic roadmaps were derived from the Vision for Space Exploration and the Aldrich Commission Report dated June 2004. NASA identified 12 strategic areas for roadmapping. The Agency added a thirteenth area on nuclear systems because the topic affects the entire program portfolio. To ensure long-term public visibility and engagement, NASA established a committee for each of the 13 areas. These committees - made up of prominent members of the scientific and aerospace industry communities and senior government personnel - worked under the Federal Advisory Committee Act. A committee was formed for each of the following program areas: 1) Robotic and Human Lunar Exploration; 2) Robotic and Human Exploration of Mars; 3) Solar System Exploration; 4) Search for Earth-Like Planets; 5) Exploration Transportation System; 6) International Space Station; 7) Space Shuttle; 8) Universe Exploration; 9) Earth Science and Applications from Space; 10) Sun-Solar System Connection; 11) Aeronautical Technologies; 12) Education; 13) Nuclear Systems. This document contains roadmap summaries for 10 of these 13 program areas; The International Space Station, Space Shuttle, and Education are excluded. The completed roadmaps for the following committees: Robotic and Human Exploration of Mars; Solar System Exploration; Search for Earth-Like Planets; Universe Exploration; Earth Science and Applications from Space; Sun-Solar System Connection are collected in a separate Strategic Roadmaps volume. This document contains memebership rosters and charters for all 13 committees.
Search automation of the generalized method of device operational characteristics improvement
NASA Astrophysics Data System (ADS)
Petrova, I. Yu; Puchkova, A. A.; Zaripova, V. M.
2017-01-01
The article presents brief results of analysis of existing search methods of the closest patents, which can be applied to determine generalized methods of device operational characteristics improvement. There were observed the most widespread clustering algorithms and metrics for determining the proximity degree between two documents. The article proposes the technique of generalized methods determination; it has two implementation variants and consists of 7 steps. This technique has been implemented in the “Patents search” subsystem of the “Intellect” system. Also the article gives an example of the use of the proposed technique.
Spiders and Worms and Crawlers, Oh My: Searching on the World Wide Web.
ERIC Educational Resources Information Center
Eagan, Ann; Bender, Laura
Searching on the world wide web can be confusing. A myriad of search engines exist, often with little or no documentation, and many of these search engines work differently from the standard search engines people are accustomed to using. Intended for librarians, this paper defines search engines, directories, spiders, and robots, and covers basics…
In Search of a Better Search Engine
ERIC Educational Resources Information Center
Kolowich, Steve
2009-01-01
Early this decade, the number of Web-based documents stored on the servers of the University of Florida hovered near 300,000. By the end of 2006, that number had leapt to four million. Two years later, the university hosts close to eight million Web documents. Web sites for colleges and universities everywhere have become repositories for data…
32 CFR 806.27 - Samples of Air Force FOIA processing documents.
Code of Federal Regulations, 2014 CFR
2014-07-01
... scientific institution or news media/all others). As a result, you (if commercial category) are required to pay all document search, review and duplication costs over $15.00. (or) As a result, you (if... result, you will be provided the first 2 hours of search time and the first 100 pages free of charge; you...
32 CFR 806.27 - Samples of Air Force FOIA processing documents.
Code of Federal Regulations, 2011 CFR
2011-07-01
... scientific institution or news media/all others). As a result, you (if commercial category) are required to pay all document search, review and duplication costs over $15.00. (or) As a result, you (if... result, you will be provided the first 2 hours of search time and the first 100 pages free of charge; you...
32 CFR 806.27 - Samples of Air Force FOIA processing documents.
Code of Federal Regulations, 2013 CFR
2013-07-01
... scientific institution or news media/all others). As a result, you (if commercial category) are required to pay all document search, review and duplication costs over $15.00. (or) As a result, you (if... result, you will be provided the first 2 hours of search time and the first 100 pages free of charge; you...
32 CFR 806.27 - Samples of Air Force FOIA processing documents.
Code of Federal Regulations, 2010 CFR
2010-07-01
... scientific institution or news media/all others). As a result, you (if commercial category) are required to pay all document search, review and duplication costs over $15.00. (or) As a result, you (if... result, you will be provided the first 2 hours of search time and the first 100 pages free of charge; you...
32 CFR 806.27 - Samples of Air Force FOIA processing documents.
Code of Federal Regulations, 2012 CFR
2012-07-01
... scientific institution or news media/all others). As a result, you (if commercial category) are required to pay all document search, review and duplication costs over $15.00. (or) As a result, you (if... result, you will be provided the first 2 hours of search time and the first 100 pages free of charge; you...
Young, Kenneth J
2014-12-01
The purpose of this study was to evaluate publicly available information of chiropractic technique systems that advocate radiography for subluxation detection to identify links between chiropractic technique systems and to describe claims made of the health effects of the osseous misalignment component of the chiropractic subluxation and radiographic paradigms. The Internet and publicly available documents were searched for information representing chiropractic technique systems that advocate radiography for subluxation detection. Key phrases including chiropractic, x-ray, radiography, and technique were identified from a Google search between April 2013 and March 2014. Phrases in Web sites and public documents were examined for any information about origins and potential links between these techniques, including the type of connection to BJ Palmer, who was the first chiropractor to advocate radiography for subluxation detection. Quotes were gathered to identify claims of health effects from osseous misalignment (subluxation) and paradigms of radiography. Techniques were grouped by region of the spine and how they could be traced back to B.J Palmer. A genealogy model and summary table of information on each technique were created. Patterns in year of origination and radiographic paradigms were noted, and percentages were calculated on elements of the techniques' characteristics in comparison to the entire group. Twenty-three techniques were identified on the Internet: 6 full spine, 17 upper cervical, and 2 techniques generating other lineage. Most of the upper cervical techniques (14/16) traced their origins to a time when the Palmer School was teaching upper cervical technique, and all the full spine techniques (6/6) originated before or after this phase. All the technique systems' documents attributed broad health effects to their methods. Many (21/23) of the techniques used spinal realignment on radiographs as one of their outcome measures. Chiropractic technique systems in this study (ie, those that advocate for radiography for subluxation misalignment detection) seem to be closely related by descent, their claims of a variety of health effects associated with chiropractic subluxation, and their radiographic paradigms.
Kozlowski, L T; Dreschel, N A; Stellman, S D; Wilkenfeld, J; Weiss, E B; Goldberg, M E
2005-02-01
The Barclay cigarette (Brown & Williamson) was introduced in 1980 in the USA in the most expensive launch in history. In the USA and around the world, Barclay was later determined to have a grooved filter design that was compromised by human smokers in the normal act of smoking, but that was measured as ultra-low tar using the standard tar testing protocol. To evaluate whether Brown & Williamson knew of the compensatability of Barclay during the design process and before it was released; to evaluate initial responses of competing tobacco companies to Barclay, before complaints were made to the Federal Trade Commission in 1981. Internet databases of industry documents (Tobacco Documents Online, Legacy Tobacco Documents Library, Brown & Williamson Litigation discovery website, Guildford and major company websites) were searched using key words, key dates, and targeted searches. Documents related specifically to the development, evaluation and release of the Barclay cigarette and related to the responses by competing tobacco companies were examined. Documents indicate the manufacturer was aware of Barclay design problems and was planning, before release, to respond to criticism. Competing companies quickly detected the filter groove stratagem and considered developing their own similar filter, but eventually backed off. The design problems with Barclay were readily understood by cigarette manufacturers, including the maker of Barclay, before official governmental evaluations occurred. Testing involving measured exposures to human smokers may in the end be crucial to identifying problems with novel cigarette designs.
Maintaining the momentum of Open Search in Earth Science Data discovery
NASA Astrophysics Data System (ADS)
Newman, D. J.; Lynnes, C.
2013-12-01
Federated Search for Earth Observation data has been a hallmark of EOSDIS (Earth Observing System Data and Information System) for two decades. Originally, the EOSDIS Version 0 system provided both data-collection-level and granule/file-level search in the mid 1990s with EOSDIS-specific socket protocols and message formats. Since that time, the advent of several standards has helped to simplify EOSDIS federated search, beginning with HTTP as the transfer protocol. Most recently, OpenSearch (www.opensearch.org) was employed for the EOS Clearinghouse (ECHO), based on a set of conventions that had been developed within the Earth Science Information Partners (ESIP) Federation. The ECHO OpenSearch API has evolved to encompass the ESIP RFC and the Open Geospatial Consortium (OGC) Open Search standard. Uptake of the ECHO Open Search API has been significant and has made ECHO accessible to client developers that found the previous ECHO SOAP API and current REST API too complex. Client adoption of the OpenSearch API appears to be largely driven by the simplicity of the OpenSearch convention. This simplicity is thus important to retain as the standard and convention evolve. For example, ECHO metrics indicate that the vast majority of ECHO users favor the following search criteria when using the REST API, - Spatial - bounding box, polygon, line and point - Temporal - start and end time - Keywords - free text Fewer than 10% of searches use additional constraints, particularly those requiring a controlled vocabulary, such as instrument, sensor, etc. This suggests that ongoing standardization efforts around OpenSearch usage for Earth Observation data may be more productive if oriented toward improving support for the Spatial, Temporal and Keyword search aspects. Areas still requiring improvement include support of - Concrete requirements for keyword constraints - Phrasal search for keyword constraints - Temporal constraint relations - Terminological symmetry between search URLs and response documents for both temporal and spatial terms - Best practices for both servers and clients. Over the past year we have seen several ongoing efforts to further standardize Open Search in the earth science domain such as, - Federation of Earth Science Information Partners (ESIP) - Open Geospatial Consortium (OGC) - Committee on Earth Observation Satellites (CEOS)
Atmospheric Science Data Center
2018-06-07
... Search and Order: Earthdata Search Order Data: Earthdata Search: Order Data Guide Documents: ... for instrument validation purposes only and not suited for science publications. SCAR-B Block: ...
Atmospheric Science Data Center
2018-06-07
... Search and Order: Earthdata Search Order Data: Earthdata Search: Order Data Guide Documents: ... for instrument validation purposes only and not suited for science publications. SCAR-B Block: ...
Atmospheric Science Data Center
2018-06-07
... Search and Order: Earthdata Search Order Data: Earthdata Search: Order Data Guide Documents: ... for instrument validation purposes only and not suited for science publications. SCAR-B Block: ...
Atmospheric Science Data Center
2018-06-05
... Search and Order: Earthdata Search Order Data: Earthdata Search: Order Data Guide Documents: ... for instrument validation purposes only and not suited for science publications. SCAR-B Block: ...
Atmospheric Science Data Center
2018-05-13
... Search and Order: Earthdata Search Order Data: Earthdata Search: Order Data Guide Documents: ... for instrument validation purposes only and not suited for science publications. SCAR-B Block: ...
Litigation in Argentina: challenging the tobacco industry
Flores, M L; Barnoya, J; Mejia, R; Alderete, E; Pérez‐Stable, E J
2006-01-01
Objective To evaluate the processes and outcomes of tobacco litigation in Argentina and to analyse the strategies of the tobacco industry to oppose litigation using tobacco industry documents. Methods A systematic search of tobacco industry documents on the internet dating from 1978 to 2002. Law library searches using Argentinean official and unofficial reports systems were combined with computerised online searches. Results There have been at least 15 failed litigation cases in Argentina and the tobacco industry presented a concerted defence in every claim regardless of cost. We categorised 11 cases as product liability and nicotine addiction, two as health care reimbursement, and two as criminal law and secondhand smoke. Industry strategies included hiring legal consultants from prestigious international and Argentinean law firms and developing litigation prevention programmes. Industry monitored legal academic meetings, controlled the development of new product liability legislation, obtained favourable opinions from experts, and closely observed the development of litigation in Argentina. Conclusion The strategies used by the industry have been successful in preventing recovery for tobacco injuries through litigation. Argentinean health advocates and lawyers need to be aware of the roles and strategies of the tobacco industry in order to develop effective litigation in Argentina. PMID:16565455
Litigation in Argentina: challenging the tobacco industry.
Flores, M L; Barnoya, J; Mejia, R; Alderete, E; Pérez-Stable, E J
2006-04-01
To evaluate the processes and outcomes of tobacco litigation in Argentina and to analyse the strategies of the tobacco industry to oppose litigation using tobacco industry documents. A systematic search of tobacco industry documents on the internet dating from 1978 to 2002. Law library searches using Argentinean official and unofficial reports systems were combined with computerised online searches. There have been at least 15 failed litigation cases in Argentina and the tobacco industry presented a concerted defence in every claim regardless of cost. We categorised 11 cases as product liability and nicotine addiction, two as health care reimbursement, and two as criminal law and secondhand smoke. Industry strategies included hiring legal consultants from prestigious international and Argentinean law firms and developing litigation prevention programmes. Industry monitored legal academic meetings, controlled the development of new product liability legislation, obtained favourable opinions from experts, and closely observed the development of litigation in Argentina. The strategies used by the industry have been successful in preventing recovery for tobacco injuries through litigation. Argentinean health advocates and lawyers need to be aware of the roles and strategies of the tobacco industry in order to develop effective litigation in Argentina.
NASA Astrophysics Data System (ADS)
Hughes, J. S.; Crichton, D. J.; Hardman, S. H.; Mattman, C. A.; Ramirez, P. M.
2009-12-01
Experience suggests that no single search paradigm will meet all of a community’s search requirements. Traditional forms based search is still considered critical by a significant percentage of most science communities. However text base and facet based search are improving the community’s perception that search can be easy and that the data is available and can be located. Finally semantic search promises ways to find data that were not conceived when the metadata was first captured and organized. This situation suggests that successful science information systems must be able to deploy new search applications quickly, efficiently, and often for ad-hoc purposes. Federated registries allow data to be packaged or associated with their metadata and managed as simple registry objects. Standard reference models for federated registries now exist that ensure registry objects are uniquely identified at registration and that versioning, classification, and cataloging are addressed automatically. Distributed but locally governed, federated registries also provide notification of registry events and federated query, linking, and replication of registry objects. Key principles for shared ontology development in the space sciences are that the ontology remains independent of its implementation and be extensible, flexible and scalable. The dichotomy between digital things and physical/conceptual things in the domain need to be unified under a standard model, such as the Open Archive Information System (OAIS) Information Object. Finally the fact must be accepted that ontology development is a difficult task that requires time, patience and experts in both the science domain and information modeling. The Planetary Data System (PDS) has adopted this architecture for it next generation information system, PDS 2010. The authors will report on progress, briefly describe key elements, and illustrate how the new system will be phased into operations to handle both legacy and new science data. In particular the shared ontology is being used to drive system implementation through the generation of standards documents and software configuration files. The resulting information system will help meet the expectations of modern scientists by providing more of the information interconnectedness, correlative science, and system interoperability that they desire. Fig.1 - Data Driven Architecture
ADP and brucellosis indemnity systems development
DOE Office of Scientific and Technical Information (OSTI.GOV)
Sanders, W.M.; Harlan, B.L.
1976-01-01
Our initial study of the USDA/TAHC Brucellosis Indemnity Program in Texas has shown that both the efficiency and rate of claim payments can be increased by the application of present day computer technologies. Two main factors contribute to these increases: the number of discrepancies that are caused by poor penmanship, transposition of numbers, and other human errors can be monitored and minimized; and the documented information can be indexed, sorted, and searched faster, more efficiently, and without human error. The overall flow of documentation that is used to control the movement of infected or exposed animals through commerce should bemore » studied. A new system should be designed that fully utilizes present day computer and electronic technologies.« less
Nursing record systems: effects on nursing practice and health care outcomes.
Currell, R; Wainwright, P; Urquhart, C
2000-01-01
A nursing record system is the record of care planned and/or given to individual patients/clients by qualified nurses or other caregivers under the direction of a qualified nurse. Nursing record systems may be an effective way of influencing nurse practice. To assess the effects of nursing record systems on nursing practice and patient outcomes. We searched The Cochrane Library, MEDLINE, Cinahl, Sigle, and databases of the Royal College of Nursing, King's Fund, the NHS Centre for Reviews and Dissemination, and the Institute of Electrical Engineers up to August 1999; and OCLC First Search, Department of Health database, NHS Register of Computer Applications and the Health Visitors' Association database up to the end of 1995. We hand searched the Journal of Nursing Administration (1971-1999), Computers in Nursing (1984-1999), Information Technology in Nursing (1989-1999) and reference lists of articles. We also hand searched the major health informatics conference proceedings. We contacted experts in the field of nursing informatics, suppliers of nursing computer systems, and relevant Internet groups. Randomised trials, controlled before and after studies and interrupted time series comparing one kind of nursing record system with another, in hospital, community or primary care settings. The participants were qualified nurses, students or health care assistants working under the direction of a qualified nurse and patients receiving care recorded and/or planned using nursing record systems. Two reviewers independently assessed trial quality and extracted data. Six trials involving 1407 people were included. In three studies of client held records, there were no overall positive or negative effects, although some administrative benefits through fewer missing notes were suggested. A paediatric pain management sheet study showed a positive effect on the children's pain intensity. A computerised nursing care planning study showed a negative effect on documented nursing care planning. A controlled before-and-after study of two paper nursing record systems showed improvement in meeting documentation standards. No evidence was found of effects on practice attributable to changes in record systems. Although there is a paucity of studies of sufficient methodological rigour to yield reliable results in this area, it is clear from the literature that it is possible to set up randomised trials or other quasi-experimental designs needed to produce evidence for practice. The research undertaken so far may have suffered both from methodological problems and faulty hypotheses.
Tobacco documents research methodology
McCandless, Phyra M; Klausner, Kim; Taketa, Rachel; Yerger, Valerie B
2011-01-01
Tobacco documents research has developed into a thriving academic enterprise since its inception in 1995. The technology supporting tobacco documents archiving, searching and retrieval has improved greatly since that time, and consequently tobacco documents researchers have considerably more access to resources than was the case when researchers had to travel to physical archives and/or electronically search poorly and incompletely indexed documents. The authors of the papers presented in this supplement all followed the same basic research methodology. Rather than leave the reader of the supplement to read the same discussion of methods in each individual paper, presented here is an overview of the methods all authors followed. In the individual articles that follow in this supplement, the authors present the additional methodological information specific to their topics. This brief discussion also highlights technological capabilities in the Legacy Tobacco Documents Library and updates methods for organising internal tobacco documents data and findings. PMID:21504933
A model for enhancing Internet medical document retrieval with "medical core metadata".
Malet, G; Munoz, F; Appleyard, R; Hersh, W
1999-01-01
Finding documents on the World Wide Web relevant to a specific medical information need can be difficult. The goal of this work is to define a set of document content description tags, or metadata encodings, that can be used to promote disciplined search access to Internet medical documents. The authors based their approach on a proposed metadata standard, the Dublin Core Metadata Element Set, which has recently been submitted to the Internet Engineering Task Force. Their model also incorporates the National Library of Medicine's Medical Subject Headings (MeSH) vocabulary and MEDLINE-type content descriptions. The model defines a medical core metadata set that can be used to describe the metadata for a wide variety of Internet documents. The authors propose that their medical core metadata set be used to assign metadata to medical documents to facilitate document retrieval by Internet search engines.
A Model for Enhancing Internet Medical Document Retrieval with “Medical Core Metadata”
Malet, Gary; Munoz, Felix; Appleyard, Richard; Hersh, William
1999-01-01
Objective: Finding documents on the World Wide Web relevant to a specific medical information need can be difficult. The goal of this work is to define a set of document content description tags, or metadata encodings, that can be used to promote disciplined search access to Internet medical documents. Design: The authors based their approach on a proposed metadata standard, the Dublin Core Metadata Element Set, which has recently been submitted to the Internet Engineering Task Force. Their model also incorporates the National Library of Medicine's Medical Subject Headings (MeSH) vocabulary and Medline-type content descriptions. Results: The model defines a medical core metadata set that can be used to describe the metadata for a wide variety of Internet documents. Conclusions: The authors propose that their medical core metadata set be used to assign metadata to medical documents to facilitate document retrieval by Internet search engines. PMID:10094069
Semantic Document Model to Enhance Data and Knowledge Interoperability
NASA Astrophysics Data System (ADS)
Nešić, Saša
To enable document data and knowledge to be efficiently shared and reused across application, enterprise, and community boundaries, desktop documents should be completely open and queryable resources, whose data and knowledge are represented in a form understandable to both humans and machines. At the same time, these are the requirements that desktop documents need to satisfy in order to contribute to the visions of the Semantic Web. With the aim of achieving this goal, we have developed the Semantic Document Model (SDM), which turns desktop documents into Semantic Documents as uniquely identified and semantically annotated composite resources, that can be instantiated into human-readable (HR) and machine-processable (MP) forms. In this paper, we present the SDM along with an RDF and ontology-based solution for the MP document instance. Moreover, on top of the proposed model, we have built the Semantic Document Management System (SDMS), which provides a set of services that exploit the model. As an application example that takes advantage of SDMS services, we have extended MS Office with a set of tools that enables users to transform MS Office documents (e.g., MS Word and MS PowerPoint) into Semantic Documents, and to search local and distant semantic document repositories for document content units (CUs) over Semantic Web protocols.
MetaSpider: Meta-Searching and Categorization on the Web.
ERIC Educational Resources Information Center
Chen, Hsinchun; Fan, Haiyan; Chau, Michael; Zeng, Daniel
2001-01-01
Discusses the difficulty of locating relevant information on the Web and studies two approaches to addressing the low precision and poor presentation of search results: meta-search and document categorization. Introduces MetaSpider, a meta-search engine, and presents results of a user evaluation study that compared three search engines.…
The NASA Astrophysics Data System: Capabilities and Roadmap for the 2020s
NASA Astrophysics Data System (ADS)
Accomazzi, Alberto; ADS Team
2018-06-01
The NASA Astrophysics Data System (ADS) is used daily by researchers and curators as a discovery platform for the Astronomy literature. Over the past several years, the ADS has been adding to the breadth and depth of its contents. Scholarly astronomy articles are now indexed as full-text documents, allowing for complete and accurate literature searches. High-level data products, data links, and software used in refereed astronomy papers are now also being ingested and indexed in our database. All the search functionality exposed in the new ADS interface is also available via its API, which we are continuing to develop and enhance. In this talk I will describe the current system, our current roadmap, and solicit input from the community regarding what additional data, services, and discovery capabilities the ADS should support.
Ontology-Driven Search and Triage: Design of a Web-Based Visual Interface for MEDLINE.
Demelo, Jonathan; Parsons, Paul; Sedig, Kamran
2017-02-02
Diverse users need to search health and medical literature to satisfy open-ended goals such as making evidence-based decisions and updating their knowledge. However, doing so is challenging due to at least two major difficulties: (1) articulating information needs using accurate vocabulary and (2) dealing with large document sets returned from searches. Common search interfaces such as PubMed do not provide adequate support for exploratory search tasks. Our objective was to improve support for exploratory search tasks by combining two strategies in the design of an interactive visual interface by (1) using a formal ontology to help users build domain-specific knowledge and vocabulary and (2) providing multi-stage triaging support to help mitigate the information overload problem. We developed a Web-based tool, Ontology-Driven Visual Search and Triage Interface for MEDLINE (OVERT-MED), to test our design ideas. We implemented a custom searchable index of MEDLINE, which comprises approximately 25 million document citations. We chose a popular biomedical ontology, the Human Phenotype Ontology (HPO), to test our solution to the vocabulary problem. We implemented multistage triaging support in OVERT-MED, with the aid of interactive visualization techniques, to help users deal with large document sets returned from searches. Formative evaluation suggests that the design features in OVERT-MED are helpful in addressing the two major difficulties described above. Using a formal ontology seems to help users articulate their information needs with more accurate vocabulary. In addition, multistage triaging combined with interactive visualizations shows promise in mitigating the information overload problem. Our strategies appear to be valuable in addressing the two major problems in exploratory search. Although we tested OVERT-MED with a particular ontology and document collection, we anticipate that our strategies can be transferred successfully to other contexts. ©Jonathan Demelo, Paul Parsons, Kamran Sedig. Originally published in JMIR Medical Informatics (http://medinform.jmir.org), 02.02.2017.
Ontology-Driven Search and Triage: Design of a Web-Based Visual Interface for MEDLINE
2017-01-01
Background Diverse users need to search health and medical literature to satisfy open-ended goals such as making evidence-based decisions and updating their knowledge. However, doing so is challenging due to at least two major difficulties: (1) articulating information needs using accurate vocabulary and (2) dealing with large document sets returned from searches. Common search interfaces such as PubMed do not provide adequate support for exploratory search tasks. Objective Our objective was to improve support for exploratory search tasks by combining two strategies in the design of an interactive visual interface by (1) using a formal ontology to help users build domain-specific knowledge and vocabulary and (2) providing multi-stage triaging support to help mitigate the information overload problem. Methods We developed a Web-based tool, Ontology-Driven Visual Search and Triage Interface for MEDLINE (OVERT-MED), to test our design ideas. We implemented a custom searchable index of MEDLINE, which comprises approximately 25 million document citations. We chose a popular biomedical ontology, the Human Phenotype Ontology (HPO), to test our solution to the vocabulary problem. We implemented multistage triaging support in OVERT-MED, with the aid of interactive visualization techniques, to help users deal with large document sets returned from searches. Results Formative evaluation suggests that the design features in OVERT-MED are helpful in addressing the two major difficulties described above. Using a formal ontology seems to help users articulate their information needs with more accurate vocabulary. In addition, multistage triaging combined with interactive visualizations shows promise in mitigating the information overload problem. Conclusions Our strategies appear to be valuable in addressing the two major problems in exploratory search. Although we tested OVERT-MED with a particular ontology and document collection, we anticipate that our strategies can be transferred successfully to other contexts. PMID:28153818
The Impact of Online Bibliographic Databases on Teaching and Research in Political Science.
ERIC Educational Resources Information Center
Reichel, Mary
The availability of online bibliographic databases greatly facilitates literature searching in political science. The advantages to searching databases online include combination of concepts, comprehensiveness, multiple database searching, free-text searching, currency, current awareness services, document delivery service, and convenience.…
CITE NLM: Natural-Language Searching in an Online Catalog.
ERIC Educational Resources Information Center
Doszkocs, Tamas E.
1983-01-01
The National Library of Medicine's Current Information Transfer in English public access online catalog offers unique subject search capabilities--natural-language query input, automatic medical subject headings display, closest match search strategy, ranked document output, dynamic end user feedback for search refinement. References, description…
Nursing record systems: effects on nursing practice and health care outcomes.
Currell, R; Urquhart, C
2003-01-01
A nursing record system is the record of care planned and/or given to individual patients/clients by qualified nurses or other caregivers under the direction of a qualified nurse. Nursing record systems may be an effective way of influencing nurse practice. To assess the effects of nursing record systems on nursing practice and patient outcomes. We searched The Cochrane Library, the EPOC trial register (October 2002), MEDLINE, Cinahl, Sigle, and databases of the Royal College of Nursing, King's Fund, the NHS Centre for Reviews and Dissemination, and the Institute of Electrical Engineers up to August 1999; and OCLC First Search, Department of Health database, NHS Register of Computer Applications and the Health Visitors' Association database up to the end of 1995. We hand searched the Journal of Nursing Administration (1971-1999), Computers in Nursing (1984-1999), Information Technology in Nursing (1989-1999) and reference lists of articles. We also hand searched the major health informatics conference proceedings. We contacted experts in the field of nursing informatics, suppliers of nursing computer systems, and relevant Internet groups. To update the review the Medline, Cinahl, British Nursing Index, Aslib Index to Theses databases were all searched from 1998 to 2002. The Journal of Nursing Administration, Computers in Nursing, Information Technology in Nursing were all hand searched up to 2002. The searches of the other databases and grey literature included in the original review, were not updated (except for Health Care Computing Conference and Med Info) as the original searches produced little relevant material. Randomised trials, controlled before and after studies and interrupted time series comparing one kind of nursing record system with another, in hospital, community or primary care settings. The participants were qualified nurses, students or health care assistants working under the direction of a qualified nurse and patients receiving care recorded and/or planned using nursing record systems. Two reviewers independently assessed trial quality and extracted data. Eight trials involving 1497 people were included. In three studies of client held records, there were no overall positive or negative effects, although some administrative benefits through fewer missing notes were suggested. A paediatric pain management sheet study showed a positive effect on the children's pain intensity. A computerised nursing care planning study showed a negative effect on documented nursing care planning, although two other computerised nursing information studies showed an increase in recording but no change in patient outcomes. Care planning took longer with these computerised systems, but the numbers of patients and nurses included in these studies was small. A controlled before-and-after study of two paper nursing record systems showed improvement in meeting documentation standards. No evidence was found of effects on practice attributable to changes in record systems. Although there is a paucity of studies of sufficient methodological rigour to yield reliable results in this area, it is clear from the literature that it is possible to set up randomised trials or other quasi-experimental designs needed to produce evidence for practice. The research undertaken so far may have suffered both from methodological problems and faulty hypotheses. Qualitative nursing research to explore the relationship between practice and information use, could be used as a precursor to the design and testing of nursing information systems.
A Scalable Monitoring for the CMS Filter Farm Based on Elasticsearch
DOE Office of Scientific and Technical Information (OSTI.GOV)
Andre, J.M.; et al.
2015-12-23
A flexible monitoring system has been designed for the CMS File-based Filter Farm making use of modern data mining and analytics components. All the metadata and monitoring information concerning data flow and execution of the HLT are generated locally in the form of small documents using the JSON encoding. These documents are indexed into a hierarchy of elasticsearch (es) clusters along with process and system log information. Elasticsearch is a search server based on Apache Lucene. It provides a distributed, multitenant-capable search and aggregation engine. Since es is schema-free, any new information can be added seamlessly and the unstructured informationmore » can be queried in non-predetermined ways. The leaf es clusters consist of the very same nodes that form the Filter Farm thus providing natural horizontal scaling. A separate central” es cluster is used to collect and index aggregated information. The fine-grained information, all the way to individual processes, remains available in the leaf clusters. The central es cluster provides quasi-real-time high-level monitoring information to any kind of client. Historical data can be retrieved to analyse past problems or correlate them with external information. We discuss the design and performance of this system in the context of the CMS DAQ commissioning for LHC Run 2.« less
Iterative Integration of Visual Insights during Scalable Patent Search and Analysis.
Koch, S; Bosch, H; Giereth, M; Ertl, T
2011-05-01
Patents are of growing importance in current economic markets. Analyzing patent information has, therefore, become a common task for many interest groups. As a prerequisite for patent analysis, extensive search for relevant patent information is essential. Unfortunately, the complexity of patent material inhibits a straightforward retrieval of all relevant patent documents and leads to iterative, time-consuming approaches in practice. Already the amount of patent data to be analyzed poses challenges with respect to scalability. Further scalability issues arise concerning the diversity of users and the large variety of analysis tasks. With "PatViz", a system for interactive analysis of patent information has been developed addressing scalability at various levels. PatViz provides a visual environment allowing for interactive reintegration of insights into subsequent search iterations, thereby bridging the gap between search and analytic processes. Because of its extensibility, we expect that the approach we have taken can be employed in different problem domains that require high quality of search results regarding their completeness.
An XML-based system for the flexible classification and retrieval of clinical practice guidelines.
Ganslandt, T.; Mueller, M. L.; Krieglstein, C. F.; Senninger, N.; Prokosch, H. U.
2002-01-01
Beneficial effects of clinical practice guidelines (CPGs) have not yet reached expectations due to limited routine adoption. Electronic distribution and reminder systems have the potential to overcome implementation barriers. Existing electronic CPG repositories like the National Guideline Clearinghouse (NGC) provide individual access but lack standardized computer-readable interfaces necessary for automated guideline retrieval. The aim of this paper was to facilitate automated context-based selection and presentation of CPGs. Using attributes from the NGC classification scheme, an XML-based metadata repository was successfully implemented, providing document storage, classification and retrieval functionality. Semi-automated extraction of attributes was implemented for the import of XML guideline documents using XPath. A hospital information system interface was exemplarily implemented for diagnosis-based guideline invocation. Limitations of the implemented system are discussed and possible future work is outlined. Integration of standardized computer-readable search interfaces into existing CPG repositories is proposed. PMID:12463831
A new information architecture, website and services for the CMS experiment
DOE Office of Scientific and Technical Information (OSTI.GOV)
Taylor, Lucas; Rusack, Eleanor; Zemleris, Vidmantas
2012-01-01
The age and size of the CMS collaboration at the LHC means it now has many hundreds of inhomogeneous web sites and services, and hundreds of thousands of documents. We describe a major initiative to create a single coherent CMS internal and public web site. This uses the Drupal web Content Management System (now supported by CERN/IT) on top of a standard LAMP stack (Linux, Apache, MySQL, and php/perl). The new navigation, content and search services are coherently integrated with numerous existing CERN services (CDS, EDMS, Indico, phonebook, Twiki) as well as many CMS internal Web services. We describe themore » information architecture, the system design, implementation and monitoring, the document and content database, security aspects, and our deployment strategy, which ensured continual smooth operation of all systems at all times.« less
A new Information Architecture, Website and Services for the CMS Experiment
NASA Astrophysics Data System (ADS)
Taylor, Lucas; Rusack, Eleanor; Zemleris, Vidmantas
2012-12-01
The age and size of the CMS collaboration at the LHC means it now has many hundreds of inhomogeneous web sites and services, and hundreds of thousands of documents. We describe a major initiative to create a single coherent CMS internal and public web site. This uses the Drupal web Content Management System (now supported by CERN/IT) on top of a standard LAMP stack (Linux, Apache, MySQL, and php/perl). The new navigation, content and search services are coherently integrated with numerous existing CERN services (CDS, EDMS, Indico, phonebook, Twiki) as well as many CMS internal Web services. We describe the information architecture; the system design, implementation and monitoring; the document and content database; security aspects; and our deployment strategy, which ensured continual smooth operation of all systems at all times.
Title list of documents made publicly available, November 1-30, 1995
DOE Office of Scientific and Technical Information (OSTI.GOV)
NONE
1996-01-01
The Title List of Documents Made Publicly Available is a monthly publication. It contains descriptions of the information received and generated by the U.S. Nuclear Regulatory Commission (NRC). This information includes (1) docketed material associated with civilian nuclear power plants and other uses of radioactive materials and (2) nondocketed material received and generated by NRC pertinent to its role as a regulatory agency. As used here, docketed does not refer to Court dockets; it refers to the system by which NRC maintains its regulatory records. This series of documents is indexed by a Personal Author Index, a Corporate Source Indexmore » and a Report Number Index. The docketed information contained in the Title List includes the information formerly issued through the Department of Energy publication Power Reactor Docker Information, last published in January 1979. NRC documents that are publicly available may be examined without charge at the NRC Public Document Room (PDR). Duplicate copies may be obtained for a fee. Standing orders for certain categories of documents are also available. Clients may search for and order desired titles through the PDR computerized Bibliographic Retrieval System, which is accessible both at the PDR and remotely. The PDR is staffed by professional technical librarians, who provide reference assistance to users.« less
Engineering Lessons Learned and Systems Engineering Applications
NASA Technical Reports Server (NTRS)
Gill, Paul S.; Garcia, Danny; Vaughan, William W.
2005-01-01
Systems Engineering is fundamental to good engineering, which in turn depends on the integration and application of engineering lessons learned. Thus, good Systems Engineering also depends on systems engineering lessons learned from within the aerospace industry being documented and applied. About ten percent of the engineering lessons learned documented in the NASA Lessons Learned Information System are directly related to Systems Engineering. A key issue associated with lessons learned datasets is the communication and incorporation of this information into engineering processes. As part of the NASA Technical Standards Program activities, engineering lessons learned datasets have been identified from a number of sources. These are being searched and screened for those having a relation to Technical Standards. This paper will address some of these Systems Engineering Lessons Learned and how they are being related to Technical Standards within the NASA Technical Standards Program, including linking to the Agency's Interactive Engineering Discipline Training Courses and the life cycle for a flight vehicle development program.
Organ donation in the ICU: A document analysis of institutional policies, protocols, and order sets.
Oczkowski, Simon J W; Centofanti, John E; Durepos, Pamela; Arseneau, Erika; Kelecevic, Julija; Cook, Deborah J; Meade, Maureen O
2018-04-01
To better understand how local policies influence organ donation rates. We conducted a document analysis of our ICU organ donation policies, protocols and order sets. We used a systematic search of our institution's policy library to identify documents related to organ donation. We used Mindnode software to create a publication timeline, basic statistics to describe document characteristics, and qualitative content analysis to extract document themes. Documents were retrieved from Hamilton Health Sciences, an academic hospital system with a high volume of organ donation, from database inception to October 2015. We retrieved 12 active organ donation documents, including six protocols, two policies, two order sets, and two unclassified documents, a majority (75%) after the introduction of donation after circulatory death in 2006. Four major themes emerged: organ donation process, quality of care, patient and family-centred care, and the role of the institution. These themes indicate areas where documented institutional standards may be beneficial. Further research is necessary to determine the relationship of local policies, protocols, and order sets to actual organ donation practices, and to identify barriers and facilitators to improving donation rates. Copyright © 2017 Elsevier Ltd. All rights reserved.
Comparing image search behaviour in the ARRS GoldMiner search engine and a clinical PACS/RIS.
De-Arteaga, Maria; Eggel, Ivan; Do, Bao; Rubin, Daniel; Kahn, Charles E; Müller, Henning
2015-08-01
Information search has changed the way we manage knowledge and the ubiquity of information access has made search a frequent activity, whether via Internet search engines or increasingly via mobile devices. Medical information search is in this respect no different and much research has been devoted to analyzing the way in which physicians aim to access information. Medical image search is a much smaller domain but has gained much attention as it has different characteristics than search for text documents. While web search log files have been analysed many times to better understand user behaviour, the log files of hospital internal systems for search in a PACS/RIS (Picture Archival and Communication System, Radiology Information System) have rarely been analysed. Such a comparison between a hospital PACS/RIS search and a web system for searching images of the biomedical literature is the goal of this paper. Objectives are to identify similarities and differences in search behaviour of the two systems, which could then be used to optimize existing systems and build new search engines. Log files of the ARRS GoldMiner medical image search engine (freely accessible on the Internet) containing 222,005 queries, and log files of Stanford's internal PACS/RIS search called radTF containing 18,068 queries were analysed. Each query was preprocessed and all query terms were mapped to the RadLex (Radiology Lexicon) terminology, a comprehensive lexicon of radiology terms created and maintained by the Radiological Society of North America, so the semantic content in the queries and the links between terms could be analysed, and synonyms for the same concept could be detected. RadLex was mainly created for the use in radiology reports, to aid structured reporting and the preparation of educational material (Lanlotz, 2006) [1]. In standard medical vocabularies such as MeSH (Medical Subject Headings) and UMLS (Unified Medical Language System) specific terms of radiology are often underrepresented, therefore RadLex was considered to be the best option for this task. The results show a surprising similarity between the usage behaviour in the two systems, but several subtle differences can also be noted. The average number of terms per query is 2.21 for GoldMiner and 2.07 for radTF, the used axes of RadLex (anatomy, pathology, findings, …) have almost the same distribution with clinical findings being the most frequent and the anatomical entity the second; also, combinations of RadLex axes are extremely similar between the two systems. Differences include a longer length of the sessions in radTF than in GoldMiner (3.4 and 1.9 queries per session on average). Several frequent search terms overlap but some strong differences exist in the details. In radTF the term "normal" is frequent, whereas in GoldMiner it is not. This makes intuitive sense, as in the literature normal cases are rarely described whereas in clinical work the comparison with normal cases is often a first step. The general similarity in many points is likely due to the fact that users of the two systems are influenced by their daily behaviour in using standard web search engines and follow this behaviour in their professional search. This means that many results and insights gained from standard web search can likely be transferred to more specialized search systems. Still, specialized log files can be used to find out more on reformulations and detailed strategies of users to find the right content. Copyright © 2015 Elsevier Inc. All rights reserved.
Web document ranking via active learning and kernel principal component analysis
NASA Astrophysics Data System (ADS)
Cai, Fei; Chen, Honghui; Shu, Zhen
2015-09-01
Web document ranking arises in many information retrieval (IR) applications, such as the search engine, recommendation system and online advertising. A challenging issue is how to select the representative query-document pairs and informative features as well for better learning and exploring new ranking models to produce an acceptable ranking list of candidate documents of each query. In this study, we propose an active sampling (AS) plus kernel principal component analysis (KPCA) based ranking model, viz. AS-KPCA Regression, to study the document ranking for a retrieval system, i.e. how to choose the representative query-document pairs and features for learning. More precisely, we fill those documents gradually into the training set by AS such that each of which will incur the highest expected DCG loss if unselected. Then, the KPCA is performed via projecting the selected query-document pairs onto p-principal components in the feature space to complete the regression. Hence, we can cut down the computational overhead and depress the impact incurred by noise simultaneously. To the best of our knowledge, we are the first to perform the document ranking via dimension reductions in two dimensions, namely, the number of documents and features simultaneously. Our experiments demonstrate that the performance of our approach is better than that of the baseline methods on the public LETOR 4.0 datasets. Our approach brings an improvement against RankBoost as well as other baselines near 20% in terms of MAP metric and less improvements using P@K and NDCG@K, respectively. Moreover, our approach is particularly suitable for document ranking on the noisy dataset in practice.
Historical Analysis of C-130E Resources
1977-07-01
DOCUMENTATION SEARCH AND COLLECTION FLOW 24 DIAGRAM 6. DATA ANALYSIS FLOW DIAGRAM TASK IV 29 7. C-130 INVENTORY BY CONFIGURATION 40 8. C-130 MISSION...RESORCE RESOURCES....EOUCE .’." * ESURE :;. 400 ................... - ., -...... ’.. ... ".......... ... ’... ’" SYSTEM DESIGN PERFORMANCE 4 , OPERATIO ...Force inventory , the control of the human and material resources crosses several functions and commands. Seldom are the initial estimates verified
78 FR 28258 - mPower\\TM\\ Design-Specific Review Standard
Federal Register 2010, 2011, 2012, 2013, 2014
2013-05-14
... Public Documents'' and then select ``Begin Web- based ADAMS Search.'' For problems with ADAMS, please... Classification ML12272A013 3.2.2 System Quality Group ML12272A015 Classification. 3.3.1 Severe Wind Loading... ML12324A156 3.3.2 Extreme Wind Loads ML12324A166 (Tornado and Hurricane Loads). 3.4.1 Internal Flood...
ERIC Educational Resources Information Center
Aagaard, James S.; And Others
This two-volume document specifies a protocol that was developed using the Reference Model for Open Systems Interconnection (OSI), which provides a framework for communications within a heterogeneous network environment. The protocol implements the features necessary for bibliographic searching, record maintenance, and mail transfer between…
ERIC Educational Resources Information Center
Veaner, Allen B.
Project BALLOTS is a large-scale library automation development project of the Stanford University Libraries which has demonstrated the feasibility of conducting on-line interactive searches of complex bibliographic files, with a large number of users working simultaneously in the same or different files. This report documents the continuing…
Finding the Sweet Spot for Catastrophic Incident Search and Rescue
2009-09-01
with SAR assets. Another high-tech SAR structure is found in the Cosmicheskaya Sistema Poiska Avariynyh Sudov56 (COSPAS-SARSAT) system. This model...shows clear 56 COSPAS is an acronym for the Russian, “Cosmicheskaya Sistema Poiska Avariynyh...under each of the preparedness categories. HSPD 5 is the companion document to HSPD 8, which discusses management of domestic incidents (Bush, 2003
G-Bean: an ontology-graph based web tool for biomedical literature retrieval
2014-01-01
Background Currently, most people use NCBI's PubMed to search the MEDLINE database, an important bibliographical information source for life science and biomedical information. However, PubMed has some drawbacks that make it difficult to find relevant publications pertaining to users' individual intentions, especially for non-expert users. To ameliorate the disadvantages of PubMed, we developed G-Bean, a graph based biomedical search engine, to search biomedical articles in MEDLINE database more efficiently. Methods G-Bean addresses PubMed's limitations with three innovations: (1) Parallel document index creation: a multithreaded index creation strategy is employed to generate the document index for G-Bean in parallel; (2) Ontology-graph based query expansion: an ontology graph is constructed by merging four major UMLS (Version 2013AA) vocabularies, MeSH, SNOMEDCT, CSP and AOD, to cover all concepts in National Library of Medicine (NLM) database; a Personalized PageRank algorithm is used to compute concept relevance in this ontology graph and the Term Frequency - Inverse Document Frequency (TF-IDF) weighting scheme is used to re-rank the concepts. The top 500 ranked concepts are selected for expanding the initial query to retrieve more accurate and relevant information; (3) Retrieval and re-ranking of documents based on user's search intention: after the user selects any article from the existing search results, G-Bean analyzes user's selections to determine his/her true search intention and then uses more relevant and more specific terms to retrieve additional related articles. The new articles are presented to the user in the order of their relevance to the already selected articles. Results Performance evaluation with 106 OHSUMED benchmark queries shows that G-Bean returns more relevant results than PubMed does when using these queries to search the MEDLINE database. PubMed could not even return any search result for some OHSUMED queries because it failed to form the appropriate Boolean query statement automatically from the natural language query strings. G-Bean is available at http://bioinformatics.clemson.edu/G-Bean/index.php. Conclusions G-Bean addresses PubMed's limitations with ontology-graph based query expansion, automatic document indexing, and user search intention discovery. It shows significant advantages in finding relevant articles from the MEDLINE database to meet the information need of the user. PMID:25474588
G-Bean: an ontology-graph based web tool for biomedical literature retrieval.
Wang, James Z; Zhang, Yuanyuan; Dong, Liang; Li, Lin; Srimani, Pradip K; Yu, Philip S
2014-01-01
Currently, most people use NCBI's PubMed to search the MEDLINE database, an important bibliographical information source for life science and biomedical information. However, PubMed has some drawbacks that make it difficult to find relevant publications pertaining to users' individual intentions, especially for non-expert users. To ameliorate the disadvantages of PubMed, we developed G-Bean, a graph based biomedical search engine, to search biomedical articles in MEDLINE database more efficiently. G-Bean addresses PubMed's limitations with three innovations: (1) Parallel document index creation: a multithreaded index creation strategy is employed to generate the document index for G-Bean in parallel; (2) Ontology-graph based query expansion: an ontology graph is constructed by merging four major UMLS (Version 2013AA) vocabularies, MeSH, SNOMEDCT, CSP and AOD, to cover all concepts in National Library of Medicine (NLM) database; a Personalized PageRank algorithm is used to compute concept relevance in this ontology graph and the Term Frequency - Inverse Document Frequency (TF-IDF) weighting scheme is used to re-rank the concepts. The top 500 ranked concepts are selected for expanding the initial query to retrieve more accurate and relevant information; (3) Retrieval and re-ranking of documents based on user's search intention: after the user selects any article from the existing search results, G-Bean analyzes user's selections to determine his/her true search intention and then uses more relevant and more specific terms to retrieve additional related articles. The new articles are presented to the user in the order of their relevance to the already selected articles. Performance evaluation with 106 OHSUMED benchmark queries shows that G-Bean returns more relevant results than PubMed does when using these queries to search the MEDLINE database. PubMed could not even return any search result for some OHSUMED queries because it failed to form the appropriate Boolean query statement automatically from the natural language query strings. G-Bean is available at http://bioinformatics.clemson.edu/G-Bean/index.php. G-Bean addresses PubMed's limitations with ontology-graph based query expansion, automatic document indexing, and user search intention discovery. It shows significant advantages in finding relevant articles from the MEDLINE database to meet the information need of the user.
Automated Patent Searching in the EPO: From Online Searching to Document Delivery.
ERIC Educational Resources Information Center
Nuyts, Annemie; Jonckheere, Charles
The European Patent Office (EPO) has recently implemented the last part of its ambitious automation project aimed at creating an automated search environment for approximately 1200 EPO patent search examiners. The examiners now have at their disposal an integrated set of tools offering a full range of functionalities from online searching, via…
Young, Kenneth J.
2014-01-01
Objective The purpose of this study was to evaluate publicly available information of chiropractic technique systems that advocate radiography for subluxation detection to identify links between chiropractic technique systems and to describe claims made of the health effects of the osseous misalignment component of the chiropractic subluxation and radiographic paradigms. Methods The Internet and publicly available documents were searched for information representing chiropractic technique systems that advocate radiography for subluxation detection. Key phrases including chiropractic, x-ray, radiography, and technique were identified from a Google search between April 2013 and March 2014. Phrases in Web sites and public documents were examined for any information about origins and potential links between these techniques, including the type of connection to BJ Palmer, who was the first chiropractor to advocate radiography for subluxation detection. Quotes were gathered to identify claims of health effects from osseous misalignment (subluxation) and paradigms of radiography. Techniques were grouped by region of the spine and how they could be traced back to B.J Palmer. A genealogy model and summary table of information on each technique were created. Patterns in year of origination and radiographic paradigms were noted, and percentages were calculated on elements of the techniques’ characteristics in comparison to the entire group. Results Twenty-three techniques were identified on the Internet: 6 full spine, 17 upper cervical, and 2 techniques generating other lineage. Most of the upper cervical techniques (14/16) traced their origins to a time when the Palmer School was teaching upper cervical technique, and all the full spine techniques (6/6) originated before or after this phase. All the technique systems’ documents attributed broad health effects to their methods. Many (21/23) of the techniques used spinal realignment on radiographs as one of their outcome measures. Conclusion Chiropractic technique systems in this study (ie, those that advocate for radiography for subluxation misalignment detection) seem to be closely related by descent, their claims of a variety of health effects associated with chiropractic subluxation, and their radiographic paradigms. PMID:25431540
Code of Federal Regulations, 2012 CFR
2012-04-01
... of producing the documents. (2) Searches—(i) Manual searches. Search fees will be assessed at the rate of $25.30 per hour. Charges for search time less than a full hour will be in increments of quarter hours. (ii) Computer searches. The FOIA Officer will charge the actual direct costs of conducting...
Code of Federal Regulations, 2013 CFR
2013-04-01
... of producing the documents. (2) Searches—(i) Manual searches. Search fees will be assessed at the rate of $25.30 per hour. Charges for search time less than a full hour will be in increments of quarter hours. (ii) Computer searches. The FOIA Officer will charge the actual direct costs of conducting...
Code of Federal Regulations, 2014 CFR
2014-04-01
... of producing the documents. (2) Searches—(i) Manual searches. Search fees will be assessed at the rate of $25.30 per hour. Charges for search time less than a full hour will be in increments of quarter hours. (ii) Computer searches. The FOIA Officer will charge the actual direct costs of conducting...
Searching Databases without Query-Building Aids: Implications for Dyslexic Users
ERIC Educational Resources Information Center
Berget, Gerd; Sandnes, Frode Eika
2015-01-01
Introduction: Few studies document the information searching behaviour of users with cognitive impairments. This paper therefore addresses the effect of dyslexia on information searching in a database with no tolerance for spelling errors and no query-building aids. The purpose was to identify effective search interface design guidelines that…
THE ROLE OF SEARCHING SERVICES IN AN ACQUISITIONS PROGRAM.
ERIC Educational Resources Information Center
LUECK, ANTOINETTE L.; AND OTHERS
A USER PRESENTS HIS POINT OF VIEW ON LITERATURE SEARCHING THROUGH THE MAJOR SEARCHING SERVICES IN THE OVERALL PROGRAM OF ACQUISITIONS FOR THE ENGINEERING STAFF OF THE AIR FORCE AERO PROPULSION LABORATORY. THESE MAJOR SEARCHING SERVICES INCLUDE THE DEFENSE DOCUMENTATION CENTER (DDC), THE NATIONAL AERONAUTICS AND SPACE ADMINISTRATION (NASA), THE…
Text Mining in Biomedical Domain with Emphasis on Document Clustering
2017-01-01
Objectives With the exponential increase in the number of articles published every year in the biomedical domain, there is a need to build automated systems to extract unknown information from the articles published. Text mining techniques enable the extraction of unknown knowledge from unstructured documents. Methods This paper reviews text mining processes in detail and the software tools available to carry out text mining. It also reviews the roles and applications of text mining in the biomedical domain. Results Text mining processes, such as search and retrieval of documents, pre-processing of documents, natural language processing, methods for text clustering, and methods for text classification are described in detail. Conclusions Text mining techniques can facilitate the mining of vast amounts of knowledge on a given topic from published biomedical research articles and draw meaningful conclusions that are not possible otherwise. PMID:28875048
Deterministic binary vectors for efficient automated indexing of MEDLINE/PubMed abstracts.
Wahle, Manuel; Widdows, Dominic; Herskovic, Jorge R; Bernstam, Elmer V; Cohen, Trevor
2012-01-01
The need to maintain accessibility of the biomedical literature has led to development of methods to assist human indexers by recommending index terms for newly encountered articles. Given the rapid expansion of this literature, it is essential that these methods be scalable. Document vector representations are commonly used for automated indexing, and Random Indexing (RI) provides the means to generate them efficiently. However, RI is difficult to implement in real-world indexing systems, as (1) efficient nearest-neighbor search requires retaining all document vectors in RAM, and (2) it is necessary to maintain a store of randomly generated term vectors to index future documents. Motivated by these concerns, this paper documents the development and evaluation of a deterministic binary variant of RI. The increased capacity demonstrated by binary vectors has implications for information retrieval, and the elimination of the need to retain term vectors facilitates distributed implementations, enhancing the scalability of RI.
Deterministic Binary Vectors for Efficient Automated Indexing of MEDLINE/PubMed Abstracts
Wahle, Manuel; Widdows, Dominic; Herskovic, Jorge R.; Bernstam, Elmer V.; Cohen, Trevor
2012-01-01
The need to maintain accessibility of the biomedical literature has led to development of methods to assist human indexers by recommending index terms for newly encountered articles. Given the rapid expansion of this literature, it is essential that these methods be scalable. Document vector representations are commonly used for automated indexing, and Random Indexing (RI) provides the means to generate them efficiently. However, RI is difficult to implement in real-world indexing systems, as (1) efficient nearest-neighbor search requires retaining all document vectors in RAM, and (2) it is necessary to maintain a store of randomly generated term vectors to index future documents. Motivated by these concerns, this paper documents the development and evaluation of a deterministic binary variant of RI. The increased capacity demonstrated by binary vectors has implications for information retrieval, and the elimination of the need to retain term vectors facilitates distributed implementations, enhancing the scalability of RI. PMID:23304369
CDAPubMed: a browser extension to retrieve EHR-based biomedical literature.
Perez-Rey, David; Jimenez-Castellanos, Ana; Garcia-Remesal, Miguel; Crespo, Jose; Maojo, Victor
2012-04-05
Over the last few decades, the ever-increasing output of scientific publications has led to new challenges to keep up to date with the literature. In the biomedical area, this growth has introduced new requirements for professionals, e.g., physicians, who have to locate the exact papers that they need for their clinical and research work amongst a huge number of publications. Against this backdrop, novel information retrieval methods are even more necessary. While web search engines are widespread in many areas, facilitating access to all kinds of information, additional tools are required to automatically link information retrieved from these engines to specific biomedical applications. In the case of clinical environments, this also means considering aspects such as patient data security and confidentiality or structured contents, e.g., electronic health records (EHRs). In this scenario, we have developed a new tool to facilitate query building to retrieve scientific literature related to EHRs. We have developed CDAPubMed, an open-source web browser extension to integrate EHR features in biomedical literature retrieval approaches. Clinical users can use CDAPubMed to: (i) load patient clinical documents, i.e., EHRs based on the Health Level 7-Clinical Document Architecture Standard (HL7-CDA), (ii) identify relevant terms for scientific literature search in these documents, i.e., Medical Subject Headings (MeSH), automatically driven by the CDAPubMed configuration, which advanced users can optimize to adapt to each specific situation, and (iii) generate and launch literature search queries to a major search engine, i.e., PubMed, to retrieve citations related to the EHR under examination. CDAPubMed is a platform-independent tool designed to facilitate literature searching using keywords contained in specific EHRs. CDAPubMed is visually integrated, as an extension of a widespread web browser, within the standard PubMed interface. It has been tested on a public dataset of HL7-CDA documents, returning significantly fewer citations since queries are focused on characteristics identified within the EHR. For instance, compared with more than 200,000 citations retrieved by breast neoplasm, fewer than ten citations were retrieved when ten patient features were added using CDAPubMed. This is an open source tool that can be freely used for non-profit purposes and integrated with other existing systems.
CDAPubMed: a browser extension to retrieve EHR-based biomedical literature
2012-01-01
Background Over the last few decades, the ever-increasing output of scientific publications has led to new challenges to keep up to date with the literature. In the biomedical area, this growth has introduced new requirements for professionals, e.g., physicians, who have to locate the exact papers that they need for their clinical and research work amongst a huge number of publications. Against this backdrop, novel information retrieval methods are even more necessary. While web search engines are widespread in many areas, facilitating access to all kinds of information, additional tools are required to automatically link information retrieved from these engines to specific biomedical applications. In the case of clinical environments, this also means considering aspects such as patient data security and confidentiality or structured contents, e.g., electronic health records (EHRs). In this scenario, we have developed a new tool to facilitate query building to retrieve scientific literature related to EHRs. Results We have developed CDAPubMed, an open-source web browser extension to integrate EHR features in biomedical literature retrieval approaches. Clinical users can use CDAPubMed to: (i) load patient clinical documents, i.e., EHRs based on the Health Level 7-Clinical Document Architecture Standard (HL7-CDA), (ii) identify relevant terms for scientific literature search in these documents, i.e., Medical Subject Headings (MeSH), automatically driven by the CDAPubMed configuration, which advanced users can optimize to adapt to each specific situation, and (iii) generate and launch literature search queries to a major search engine, i.e., PubMed, to retrieve citations related to the EHR under examination. Conclusions CDAPubMed is a platform-independent tool designed to facilitate literature searching using keywords contained in specific EHRs. CDAPubMed is visually integrated, as an extension of a widespread web browser, within the standard PubMed interface. It has been tested on a public dataset of HL7-CDA documents, returning significantly fewer citations since queries are focused on characteristics identified within the EHR. For instance, compared with more than 200,000 citations retrieved by breast neoplasm, fewer than ten citations were retrieved when ten patient features were added using CDAPubMed. This is an open source tool that can be freely used for non-profit purposes and integrated with other existing systems. PMID:22480327
Data warehouse governance programs in healthcare settings: a literature review and a call to action.
Elliott, Thomas E; Holmes, John H; Davidson, Arthur J; La Chance, Pierre-Andre; Nelson, Andrew F; Steiner, John F
2013-01-01
Given the extensive data stored in healthcare data warehouses, data warehouse governance policies are needed to ensure data integrity and privacy. This review examines the current state of the data warehouse governance literature as it applies to healthcare data warehouses, identifies knowledge gaps, provides recommendations, and suggests approaches for further research. A comprehensive literature search using five data bases, journal article title-search, and citation searches was conducted between 1997 and 2012. Data warehouse governance documents from two healthcare systems in the USA were also reviewed. A modified version of nine components from the Data Governance Institute Framework for data warehouse governance guided the qualitative analysis. Fifteen articles were retrieved. Only three were related to healthcare settings, each of which addressed only one of the nine framework components. Of the remaining 12 articles, 10 addressed between one and seven framework components and the remainder addressed none. Each of the two data warehouse governance plans obtained from healthcare systems in the USA addressed a subset of the framework components, and between them they covered all nine. While published data warehouse governance policies are rare, the 15 articles and two healthcare organizational documents reviewed in this study may provide guidance to creating such policies. Additional research is needed in this area to ensure that data warehouse governance polices are feasible and effective. The gap between the development of data warehouses in healthcare settings and formal governance policies is substantial, as evidenced by the sparse literature in this domain.
Implementation of a thesaurus in an electronic photograph imaging system
NASA Astrophysics Data System (ADS)
Partlow, Denise
1995-11-01
A photograph imaging system presents a unique set of requirements for indexing and retrieving images, unlike a standard imaging system for written documents. This paper presents the requirements, technical design, and development results for a hierarchical ANSI standard thesaurus embedded into a photograph archival system. The thesaurus design incorporates storage reduction techniques, permits fast searches, and contains flexible indexing methods. It can be extended to many applications other than the retrieval of photographs. When photographic images are indexed into an electronic system, they are subject to a variety of indexing problems based on what the indexer `sees.' For instance, the indexer may categorize an image as a boat when others might refer to it as a ship, sailboat, or raft. The thesaurus will allow a user to locate images containing any synonym for boat, regardless of how the image was actually indexed. In addition to indexing problems, photos may need to be retrieved based on a broad category, for instance, flowers. The thesaurus allows a search for `flowers' to locate all images containing a rose, hibiscus, or daisy, yet still allow a specific search for an image containing only a rose. The technical design and method of implementation for such a thesaurus is presented. The thesaurus is implemented using an SQL relational data base management system that supports blobs, binary large objects. The design incorporates unique compression methods for storing the thesaurus words. Words are indexed to photographs using the compressed word and allow for very rapid searches, eliminating lengthy string matches.
MO/DSD online information server and global information repository access
NASA Technical Reports Server (NTRS)
Nguyen, Diem; Ghaffarian, Kam; Hogie, Keith; Mackey, William
1994-01-01
Often in the past, standards and new technology information have been available only in hardcopy form, with reproduction and mailing costs proving rather significant. In light of NASA's current budget constraints and in the interest of efficient communications, the Mission Operations and Data Systems Directorate (MO&DSD) New Technology and Data Standards Office recognizes the need for an online information server (OLIS). This server would allow: (1) dissemination of standards and new technology information throughout the Directorate more quickly and economically; (2) online browsing and retrieval of documents that have been published for and by MO&DSD; and (3) searching for current and past study activities on related topics within NASA before issuing a task. This paper explores a variety of available information servers and searching tools, their current capabilities and limitations, and the application of these tools to MO&DSD. Most importantly, the discussion focuses on the way this concept could be easily applied toward improving dissemination of standards and new technologies and improving documentation processes.
A Three-fold Outlook of the Ultra-Efficient Engine Technology Program Office (UEET)
NASA Technical Reports Server (NTRS)
Graham, La Quilia E.
2004-01-01
The Ultra-Efficient Engine Technology (UEET) Office at NASA Glenn Research Center is a part of the Aeronautics Directorate. Its vision is to develop and hand off revolutionary turbine engine propulsion technologies that will enable future generation vehicles over a wide range of flight speeds. There are seven different technology area projects of UEET. During my tenure at NASA Glenn Research Center, my assignment was to assist three different areas of UEET, simultaneously. I worked with Kathy Zona in Education Outreach, Lynn Boukalik in Knowledge Management, and Denise Busch with Financial Management. All of my tasks were related to the business side of UEET. As an intern with Education Outreach I created a word search to partner with an exhibit of a Turbine Engine developed out of the UEET office. This exhibit is a portable model that is presented to students of varying ages. The word search complies with National Standards for Education which are part of every science, engineering, and technology teachers curriculum. I also updated a Conference Planning/Workshop Excel Spreadsheet for the UEET Office. I collected and inputted facility overviews from various venues, both on and off site to determine where to hold upcoming conferences. I then documented which facilities were compliant with the Federal Emergency Management Agency's (FEMA) Hotel and Motel Fire Safety Act of 1990. The second area in which I worked was Knowledge Management. a large knowledge management system online which has extensive documentation that continually needs reviewing, updating, and archiving. Knowledge management is the ability to bring individual or team knowledge to an organizational level so that the information can be stored, shared, reviewed, archived. Livelink and a secure server are the Knowledge Management systems that UEET utilizes, Through these systems, I was able to obtain the documents needed for archiving. My assignment was to obtain intellectual property including reports, presentations, or any other documents related to the project. My next task was to document the author, date of creation, and all other properties of each document. To archive these documents I worked extensively with Microsoft Excel. different financial systems of accounting such as the SAP business accounting system. I also learned the best ways to present financial data and shadowed my mentor as she presented financial data to both UEET's project management and the Resources Analysis and Management Office (RAMO). I analyzed the June 2004 financial data of UEET and used Microsoft Excel to input the results of the data. This process made it easier to present the full cost of the project in the month of June. In addition I assisted in the End of the Year 2003 Reconciliation of Purchases of UEET.
Program document for Energy Systems Optimization Program 2 (ESOP2). Volume 1: Engineering manual
NASA Technical Reports Server (NTRS)
Hamil, R. G.; Ferden, S. L.
1977-01-01
The Energy Systems Optimization Program, which is used to provide analyses of Modular Integrated Utility Systems (MIUS), is discussed. Modifications to the input format to allow modular inputs in specified blocks of data are described. An optimization feature which enables the program to search automatically for the minimum value of one parameter while varying the value of other parameters is reported. New program option flags for prime mover analyses and solar energy for space heating and domestic hot water are also covered.
The role of sensory perception in the development and targeting of tobacco products.
Carpenter, Carrie M; Wayne, Geoffrey Ferris; Connolly, Gregory N
2007-01-01
To examine tobacco industry research on smoking-related sensory effects, including differences in sensory perception across smoker groups, and to determine whether this research informed targeted product development and impacted the development of commercial tobacco products. We searched previously secret internal tobacco industry documents available online through document databases housed at Tobacco Documents Online, the British American Tobacco Document Archive and the Legacy Tobacco Documents Library. We identified relevant documents using a snowball sampling method to first search the databases using an initial set of key words and to then establish further search terms. Sensory research is a priority within the tobacco industry directly impacting commercial markets both in the United States and internationally. Sensory factors contribute to smoker satisfaction and product acceptance, and play an important role in controlling puffing behavior. Cigarette manufacturers have capitalized on distinct sensory preferences across gender, age and ethnic groups by tailoring products for specific populations. Regulation of tobacco products is needed to address product changes that are used to reinforce or contribute to tobacco dependence; for instance, the incorporation of additives that target attributes such as smoothness, harshness and aftertaste. Greater understanding of the role of sensory effects on smoking behavior may also help to inform the development of tobacco treatment options that support long-term tobacco abstinence.
Federal Register 2010, 2011, 2012, 2013, 2014
2010-05-03
...'' field when using either the Web-based search (advanced search) engine or the ADAMS FIND tool in Citrix... should enter ``05200011'' in the ``Docket Number'' field in the web-based search (advanced search) engine... ML100740441. To search for documents in ADAMS using Vogtle Units 3 and 4 COL application docket numbers, 52...
Marketing to America's youth: evidence from corporate documents
Cummings, K; Morley, C; Horan, J; Steger, C; Leavell, N
2002-01-01
Design: The data for this study come from tobacco industry documents collected from the tobacco industry's document websites, presently linked at http://www.tobaccoarchives.com. The websites were searched using "request for production" (RFP) codes, specified keyword searches, and serendipitous terms identified in document citations found with RFP and keyword searches. Results: Industry documents show that the cigarette manufacturers carefully monitored the smoking habits of teenagers over the past several decades. Candid quotes from industry executives refer to youth as a source of sales and as fundamental to the survival of the tobacco industry. The documents reveal that the features of cigarette brands (that is, use of filters, low tar, bland taste, etc), packaging (that is, size, colour and design), and advertising (that is, media placements and themes and imagery) were developed specifically to appeal to new smokers (that is, teenagers). Evidence also indicates that relevant youth oriented marketing documents may have been destroyed and that the language used in some of the more recent documents may have been sanitised to cover up efforts to market to youth. Conclusions: The tobacco industry's internal documents reveal an undeniable interest in marketing cigarettes to underage smokers. The industry's marketing approaches run counter to and predicate methods for tobacco prevention: (1) keep the price of the product high; (2) keep product placements and advertising away from schools and other areas with a high volume of youth traffic; (3) make cigarette advertising (that is, themes and visual images) unappealing to youth; (4) make product packaging unappealing to youth; and (5) design the product so it is not easy to inhale. PMID:11893810
NASA Technical Reports Server (NTRS)
1972-01-01
A review of the literature used in conducting the manned space flight nuclear system safety study is presented. The objectives of the presentation are to identify and evaluate for potential application to study the existing related literature and to provide the information required to include the related literature in the NASA Aerospace Safety Research and Data Institute. More than 15,000 documents were evaluated and identification forms were prepared for 850 reports.
A Framework for the Systematic Collection of Open Source Intelligence
DOE Office of Scientific and Technical Information (OSTI.GOV)
Pouchard, Line Catherine; Trien, Joseph P; Dobson, Jonathan D
2009-01-01
Following legislative directions, the Intelligence Community has been mandated to make greater use of Open Source Intelligence (OSINT). Efforts are underway to increase the use of OSINT but there are many obstacles. One of these obstacles is the lack of tools helping to manage the volume of available data and ascertain its credibility. We propose a unique system for selecting, collecting and storing Open Source data from the Web and the Open Source Center. Some data management tasks are automated, document source is retained, and metadata containing geographical coordinates are added to the documents. Analysts are thus empowered to search,more » view, store, and analyze Web data within a single tool. We present ORCAT I and ORCAT II, two implementations of the system.« less
77 FR 36583 - NRC Form 5, Occupational Dose Record for a Monitoring Period
Federal Register 2010, 2011, 2012, 2013, 2014
2012-06-19
... methods: Federal Rulemaking Web site: Go to http://www.regulations.gov and search for Docket ID NRC-2012... following methods: Federal Rulemaking Web Site: Go to http://www.regulations.gov and search for Docket ID... begin the search, select ``ADAMS Public Documents'' and then select ``Begin Web- based ADAMS Search...
Kozlowski, L; Dreschel, N; Stellman, S; Wilkenfeld, J; Weiss, E; Goldberg, M
2005-01-01
Background: The Barclay cigarette (Brown & Williamson) was introduced in 1980 in the USA in the most expensive launch in history. In the USA and around the world, Barclay was later determined to have a grooved filter design that was compromised by human smokers in the normal act of smoking, but that was measured as ultra-low tar using the standard tar testing protocol. Objectives: To evaluate whether Brown & Williamson knew of the compensatability of Barclay during the design process and before it was released; to evaluate initial responses of competing tobacco companies to Barclay, before complaints were made to the Federal Trade Commission in 1981. Methods: Internet databases of industry documents (Tobacco Documents Online, Legacy Tobacco Documents Library, Brown & Williamson Litigation discovery website, Guildford and major company websites) were searched using key words, key dates, and targeted searches. Documents related specifically to the development, evaluation and release of the Barclay cigarette and related to the responses by competing tobacco companies were examined. Results: Documents indicate the manufacturer was aware of Barclay design problems and was planning, before release, to respond to criticism. Competing companies quickly detected the filter groove stratagem and considered developing their own similar filter, but eventually backed off. Conclusion: The design problems with Barclay were readily understood by cigarette manufacturers, including the maker of Barclay, before official governmental evaluations occurred. Testing involving measured exposures to human smokers may in the end be crucial to identifying problems with novel cigarette designs. PMID:15735303
Mutatina, Boniface; Basaza, Robert; Obuku, Ekwaro; Lavis, John N; Sewankambo, Nelson
2017-02-06
Health policymakers in low- and middle-income countries continue to face difficulties in accessing and using research evidence for decision-making. This study aimed to identify and provide a refined categorisation of the policy documents necessary for building the content of a one-stop shop for documents relevant to health policy and systems in Uganda. The on-line resource is to facilitate timely access to well-packaged evidence for decision-making. We conducted a scoping review of Uganda-specific, health policy, and systems-relevant documents produced between 2000 and 2014. Our methods borrowed heavily from the 2005 Arksey and O'Malley approach for scoping reviews and involved five steps, which that include identification of the research question; identification of relevant documents; screening and selection of the documents; charting of the data; and collating, summarising and reporting results. We searched for the documents from websites of relevant government institutions, non-governmental organisations, health professional councils and associations, religious medical bureaus and research networks. We presented the review findings as numerical analyses of the volume and nature of documents and trends over time in the form of tables and charts. We identified a total of 265 documents including policies, strategies, plans, guidelines, rapid response summaries, evidence briefs for policy, and dialogue reports. The top three clusters of national priority areas addressed in the documents were governance, coordination, monitoring and evaluation (28%); disease prevention, mitigation, and control (23%); and health education, promotion, environmental health and nutrition (15%). The least addressed were curative, palliative care, rehabilitative services and health infrastructure, each addressed in three documents (1%), and early childhood development in one document. The volume of documents increased over the past 15 years; however, the distribution of the different document types over time has not been uniform. The review findings are necessary for mobilising and packaging the local policy-relevant documents in Uganda in a one-stop shop; where policymakers could easily access them to address pressing questions about the health system and interventions. The different types of available documents and the national priority areas covered provide a good basis for building and organising the content in a meaningful way for the resource.
Basic firefly algorithm for document clustering
NASA Astrophysics Data System (ADS)
Mohammed, Athraa Jasim; Yusof, Yuhanis; Husni, Husniza
2015-12-01
The Document clustering plays significant role in Information Retrieval (IR) where it organizes documents prior to the retrieval process. To date, various clustering algorithms have been proposed and this includes the K-means and Particle Swarm Optimization. Even though these algorithms have been widely applied in many disciplines due to its simplicity, such an approach tends to be trapped in a local minimum during its search for an optimal solution. To address the shortcoming, this paper proposes a Basic Firefly (Basic FA) algorithm to cluster text documents. The algorithm employs the Average Distance to Document Centroid (ADDC) as the objective function of the search. Experiments utilizing the proposed algorithm were conducted on the 20Newsgroups benchmark dataset. Results demonstrate that the Basic FA generates a more robust and compact clusters than the ones produced by K-means and Particle Swarm Optimization (PSO).
[Profile of a systematic search. Search areas, databases and reports].
Korsbek, Lisa; Bendix, Ane Friis; Kidholm, Kristian
2006-04-03
Systematic literature search is a fundamental in evidence-based medicine. But systematic literature search is not yet a very well used way of retrieving evidence-based information. This article profiles a systematic literature search for evidence-based literature. It goes through the most central databases and gives an example of how to document the literature search. The article also sums up the literature search in all reviews in Ugeskrift for Laeger in the year 2004.
Intelligent search and retrieval of a large multimedia knowledgebase for the Hubble Space Telescope
NASA Technical Reports Server (NTRS)
Clapis, Paul J.; Byers, William S.
1990-01-01
A document-retrieval assistant (DRA) in a microcomputer format is described which incorporates hypertext and natural language capabilities. Hypertext is used to introduce an intelligent search capability, and the natural-language interface permits access to specific data without the use of keywords. The DRA can be used to access and 'browse' the large multimedia database that is composed of project documentation from the HST.
1977-06-01
Abstract (Umit =0 worfs) A total of 325 archeological sites were documented during surveys of Cochiti Reservoir. Detailed summaries of environmental...ted sites within a regional context. during the course of this project. This and continuing re- search problems wer selected which focused upon search...for sites newly documented during survey. Survey of the permanent pool was conducted by Richard C.. Chapman, supervisory archeologist James
Naval Sea Systems Command On Watch 2010
2010-01-01
surface targets, such as zodiacs and fast patrol boats found in the littoral environment. As for future capabilities and goals for the program, An...Report Documentation Page Form ApprovedOMB No. 0704-0188 Public reporting burden for the collection of information is estimated to average 1 hour...per response, including the time for reviewing instructions, searching existing data sources, gathering and maintaining the data needed, and completing
Performance indicators for maternity care in a circumpolar context: a scoping review
Rich, Rebecca; D'Hont, Thomsen; Linton, Janice; Murphy, Kellie E.; Veillard, Jeremy; Chatwood, Susan
2016-01-01
Background In circumpolar regions, harsh climates and scattered populations have prompted the centralization of care and reduction of local maternity services. The resulting practice of routine evacuation for birth from smaller towns to larger urban centres points to a potential conflict between the necessity to ensure patient safety and the importance of delivering services that are responsive to the health needs and values of populations served. Objective To identify recommended performance/quality indicators for use in circumpolar maternity care systems. Methods We searched Scopus, Ebscohost databases (including Academic Search Complete and CINAHL), the Global Health Database, High North Research Documents, and online grey literature. Articles were included if they focused on maternal health indicators in the population of interest (Indigenous women, women receiving care in circumpolar or remote regions). Articles were excluded if they were not related to pregnancy, birth or the immediate post-partum or neonatal periods. Two reviewers independently reviewed articles for inclusion and extracted relevant data. Results Twenty-six documents were included. Twelve were government documents, seven were review articles or indicator compilations, four were indicator sets recommended by academics or non-governmental organizations and three were research papers. We extracted and categorized 81 unique health indicators. The majority of indicators reflected health systems processes and outcomes during the antenatal and intra-partum periods. Only two governmental indicator sets explicitly considered the needs of Indigenous peoples. Conclusions This review demonstrates that, although most circumpolar health systems engage in performance reporting for maternity care, efforts to capture local priorities and values are limited in most regions. Future work in this area should involve northern stakeholders in the process of indicator selection and development. PMID:27938636
Performance indicators for maternity care in a circumpolar context: a scoping review.
Rich, Rebecca; D'Hont, Thomsen; Linton, Janice; Murphy, Kellie E; Veillard, Jeremy; Chatwood, Susan
2016-01-01
In circumpolar regions, harsh climates and scattered populations have prompted the centralization of care and reduction of local maternity services. The resulting practice of routine evacuation for birth from smaller towns to larger urban centres points to a potential conflict between the necessity to ensure patient safety and the importance of delivering services that are responsive to the health needs and values of populations served. To identify recommended performance/quality indicators for use in circumpolar maternity care systems. We searched Scopus, Ebscohost databases (including Academic Search Complete and CINAHL), the Global Health Database, High North Research Documents, and online grey literature. Articles were included if they focused on maternal health indicators in the population of interest (Indigenous women, women receiving care in circumpolar or remote regions). Articles were excluded if they were not related to pregnancy, birth or the immediate post-partum or neonatal periods. Two reviewers independently reviewed articles for inclusion and extracted relevant data. Twenty-six documents were included. Twelve were government documents, seven were review articles or indicator compilations, four were indicator sets recommended by academics or non-governmental organizations and three were research papers. We extracted and categorized 81 unique health indicators. The majority of indicators reflected health systems processes and outcomes during the antenatal and intra-partum periods. Only two governmental indicator sets explicitly considered the needs of Indigenous peoples. This review demonstrates that, although most circumpolar health systems engage in performance reporting for maternity care, efforts to capture local priorities and values are limited in most regions. Future work in this area should involve northern stakeholders in the process of indicator selection and development.
Performance indicators for maternity care in a circumpolar context: a scoping review.
Rich, Rebecca; D'Hont, Thomsen; Linton, Janice; Murphy, Kellie E; Veillard, Jeremy; Chatwood, Susan
2016-01-01
Background In circumpolar regions, harsh climates and scattered populations have prompted the centralization of care and reduction of local maternity services. The resulting practice of routine evacuation for birth from smaller towns to larger urban centres points to a potential conflict between the necessity to ensure patient safety and the importance of delivering services that are responsive to the health needs and values of populations served. Objective To identify recommended performance/quality indicators for use in circumpolar maternity care systems. Methods We searched Scopus, Ebscohost databases (including Academic Search Complete and CINAHL), the Global Health Database, High North Research Documents, and online grey literature. Articles were included if they focused on maternal health indicators in the population of interest (Indigenous women, women receiving care in circumpolar or remote regions). Articles were excluded if they were not related to pregnancy, birth or the immediate post-partum or neonatal periods. Two reviewers independently reviewed articles for inclusion and extracted relevant data. Results Twenty-six documents were included. Twelve were government documents, seven were review articles or indicator compilations, four were indicator sets recommended by academics or non-governmental organizations and three were research papers. We extracted and categorized 81 unique health indicators. The majority of indicators reflected health systems processes and outcomes during the antenatal and intra-partum periods. Only two governmental indicator sets explicitly considered the needs of Indigenous peoples. Conclusions This review demonstrates that, although most circumpolar health systems engage in performance reporting for maternity care, efforts to capture local priorities and values are limited in most regions. Future work in this area should involve northern stakeholders in the process of indicator selection and development.
NASA Astrophysics Data System (ADS)
Karami, Mojtaba; Rangzan, Kazem; Saberi, Azim
2013-10-01
With emergence of air-borne and space-borne hyperspectral sensors, spectroscopic measurements are gaining more importance in remote sensing. Therefore, the number of available spectral reference data is constantly increasing. This rapid increase often exhibits a poor data management, which leads to ultimate isolation of data on disk storages. Spectral data without precise description of the target, methods, environment, and sampling geometry cannot be used by other researchers. Moreover, existing spectral data (in case it accompanied with good documentation) become virtually invisible or unreachable for researchers. Providing documentation and a data-sharing framework for spectral data, in which researchers are able to search for or share spectral data and documentation, would definitely improve the data lifetime. Relational Database Management Systems (RDBMS) are main candidates for spectral data management and their efficiency is proven by many studies and applications to date. In this study, a new approach to spectral data administration is presented based on spatial identity of spectral samples. This method benefits from scalability and performance of RDBMS for storage of spectral data, but uses GIS servers to provide users with interactive maps as an interface to the system. The spectral files, photographs and descriptive data are considered as belongings of a geospatial object. A spectral processing unit is responsible for evaluation of metadata quality and performing routine spectral processing tasks for newly-added data. As a result, by using internet browser software the users would be able to visually examine availability of data and/or search for data based on descriptive attributes associated to it. The proposed system is scalable and besides giving the users good sense of what data are available in the database, it facilitates participation of spectral reference data in producing geoinformation.
The International Patent Situation
ERIC Educational Resources Information Center
Helliwell, B. F. M.
1974-01-01
Highlights the differences in patent laws in different countries to illustrate the importance of searching foreign patents, indicates how patent searches should be tackled and what assistance is available from patent offices, searching organizations and commercial patent documentation services, and considers the probable effect of the Europatent…
Federal Register 2010, 2011, 2012, 2013, 2014
2010-07-08
... either the Web-based search (advanced search) engine or the ADAMS find tool in Citrix. Within 30 days.... To search for other related documents in ADAMS using the Watts Bar Nuclear Plant Unit 2 OL...
Jácome, Alberto G; Fdez-Riverola, Florentino; Lourenço, Anália
2016-07-01
Text mining and semantic analysis approaches can be applied to the construction of biomedical domain-specific search engines and provide an attractive alternative to create personalized and enhanced search experiences. Therefore, this work introduces the new open-source BIOMedical Search Engine Framework for the fast and lightweight development of domain-specific search engines. The rationale behind this framework is to incorporate core features typically available in search engine frameworks with flexible and extensible technologies to retrieve biomedical documents, annotate meaningful domain concepts, and develop highly customized Web search interfaces. The BIOMedical Search Engine Framework integrates taggers for major biomedical concepts, such as diseases, drugs, genes, proteins, compounds and organisms, and enables the use of domain-specific controlled vocabulary. Technologies from the Typesafe Reactive Platform, the AngularJS JavaScript framework and the Bootstrap HTML/CSS framework support the customization of the domain-oriented search application. Moreover, the RESTful API of the BIOMedical Search Engine Framework allows the integration of the search engine into existing systems or a complete web interface personalization. The construction of the Smart Drug Search is described as proof-of-concept of the BIOMedical Search Engine Framework. This public search engine catalogs scientific literature about antimicrobial resistance, microbial virulence and topics alike. The keyword-based queries of the users are transformed into concepts and search results are presented and ranked accordingly. The semantic graph view portraits all the concepts found in the results, and the researcher may look into the relevance of different concepts, the strength of direct relations, and non-trivial, indirect relations. The number of occurrences of the concept shows its importance to the query, and the frequency of concept co-occurrence is indicative of biological relations meaningful to that particular scope of research. Conversely, indirect concept associations, i.e. concepts related by other intermediary concepts, can be useful to integrate information from different studies and look into non-trivial relations. The BIOMedical Search Engine Framework supports the development of domain-specific search engines. The key strengths of the framework are modularity and extensibilityin terms of software design, the use of open-source consolidated Web technologies, and the ability to integrate any number of biomedical text mining tools and information resources. Currently, the Smart Drug Search keeps over 1,186,000 documents, containing more than 11,854,000 annotations for 77,200 different concepts. The Smart Drug Search is publicly accessible at http://sing.ei.uvigo.es/sds/. The BIOMedical Search Engine Framework is freely available for non-commercial use at https://github.com/agjacome/biomsef. Copyright © 2016 Elsevier Ireland Ltd. All rights reserved.
Automatic Identification of Topic Tags from Texts Based on Expansion-Extraction Approach
ERIC Educational Resources Information Center
Yang, Seungwon
2013-01-01
Identifying topics of a textual document is useful for many purposes. We can organize the documents by topics in digital libraries. Then, we could browse and search for the documents with specific topics. By examining the topics of a document, we can quickly understand what the document is about. To augment the traditional manual way of topic…
Search extension transforms Wiki into a relational system: a case for flavonoid metabolite database.
Arita, Masanori; Suwa, Kazuhiro
2008-09-17
In computer science, database systems are based on the relational model founded by Edgar Codd in 1970. On the other hand, in the area of biology the word 'database' often refers to loosely formatted, very large text files. Although such bio-databases may describe conflicts or ambiguities (e.g. a protein pair do and do not interact, or unknown parameters) in a positive sense, the flexibility of the data format sacrifices a systematic query mechanism equivalent to the widely used SQL. To overcome this disadvantage, we propose embeddable string-search commands on a Wiki-based system and designed a half-formatted database. As proof of principle, a database of flavonoid with 6902 molecular structures from over 1687 plant species was implemented on MediaWiki, the background system of Wikipedia. Registered users can describe any information in an arbitrary format. Structured part is subject to text-string searches to realize relational operations. The system was written in PHP language as the extension of MediaWiki. All modifications are open-source and publicly available. This scheme benefits from both the free-formatted Wiki style and the concise and structured relational-database style. MediaWiki supports multi-user environments for document management, and the cost for database maintenance is alleviated.
Search extension transforms Wiki into a relational system: A case for flavonoid metabolite database
Arita, Masanori; Suwa, Kazuhiro
2008-01-01
Background In computer science, database systems are based on the relational model founded by Edgar Codd in 1970. On the other hand, in the area of biology the word 'database' often refers to loosely formatted, very large text files. Although such bio-databases may describe conflicts or ambiguities (e.g. a protein pair do and do not interact, or unknown parameters) in a positive sense, the flexibility of the data format sacrifices a systematic query mechanism equivalent to the widely used SQL. Results To overcome this disadvantage, we propose embeddable string-search commands on a Wiki-based system and designed a half-formatted database. As proof of principle, a database of flavonoid with 6902 molecular structures from over 1687 plant species was implemented on MediaWiki, the background system of Wikipedia. Registered users can describe any information in an arbitrary format. Structured part is subject to text-string searches to realize relational operations. The system was written in PHP language as the extension of MediaWiki. All modifications are open-source and publicly available. Conclusion This scheme benefits from both the free-formatted Wiki style and the concise and structured relational-database style. MediaWiki supports multi-user environments for document management, and the cost for database maintenance is alleviated. PMID:18822113
Transformational derivation of programs using the Focus system
NASA Technical Reports Server (NTRS)
Reddy, Uday S.
1988-01-01
A program derivation support system called Focus is being constructed. It will formally derive programs using the paradigm of program transformation. The following issues are discussed: (1) the integration of validation and program derivation activities in the Focus system; (2) its tree-based user interface; (3) the control of search spaces in program derivation; and (4) the structure and organization of program derivation records. The inference procedures of the system are based on the integration of functional and logic programming principles. This brings about a synthesis of paradigms that were heretofore considered far apart, such as logical and executable specifications and constructive and transformational approaches to program derivation. A great emphasis has been placed, in the design of Focus, on achieving small search spaces during program derivation. The program manipulation operations such as expansion, simplification and rewriting were designed with this objective. The role of operations that are expensive in search spaces, such as folding, has been reduced. Program derivations are documented in Focus in a way that the high level descriptions of derivations are expressed only using program level information. All the meta-level information, together with dependencies between derivations of program components, is automatically recorded by the system at a lower level of description for its own use in replay.
Generating Hierarchical Document Indices from Common Denominators in Large Document Collections.
ERIC Educational Resources Information Center
O'Kane, Kevin C.
1996-01-01
Describes an algorithm for computer generation of hierarchical indexes for document collections. The resulting index, when presented with a graphical interface, provides users with a view of a document collection that permits general browsing and informal search activities via an access method that requires no keyboard entry or prior knowledge of…
2008-11-01
T or more words, where T is a threshold that is empirically set to 300 in the experiment. The second rule aims to remove pornographic documents...Some blog documents are embedded with pornographic words to attract search traffic. We identify a list of pornographic words. Given a blog document, all...document, this document is considered pornographic spam, and is discarded. The third rule removes documents written in foreign languages. We count the
The Impact of Subject Indexes on Semantic Indeterminacy in Enterprise Document Retrieval
ERIC Educational Resources Information Center
Schymik, Gregory
2012-01-01
Ample evidence exists to support the conclusion that enterprise search is failing its users. This failure is costing corporate America billions of dollars every year. Most enterprise search engines are built using web search engines as their foundations. These search engines are optimized for web use and are inadequate when used inside the…
Drug Testing and Searches in Public Schools: A Legal Analysis.
ERIC Educational Resources Information Center
Minnesota House of Representatives, St. Paul. Research Dept.
This document examines the Fourth Amendment as the source of search and seizure law; drug testing of school employees; and drug testing searches of students. The United States Supreme Court case that established the two-part test to determine the legality of a student search is discussed, three separate student drug testing programs that have been…
Searching for Sterile Neutrinos with MINOS
DOE Office of Scientific and Technical Information (OSTI.GOV)
Timmons, Ashley
2016-01-01
This document presents the latest results for a 3+1 sterile neutrino search using themore » $$10.56 \\times 10^{20}$$ protons-on-target data set taken from 2005 - 2012. By searching for oscillations driven by a large mass splitting, MINOS is sensitive to the existence of sterile neutrinos through any energy dependent deviations using a charged current sample, as well as looking at any relative deficit between neutral current events between the far and near detectors. This document will discuss the novel analysis that enabled a search for sterile neutrinos setting a limit in the previously unexplored regions in the parameter space $$\\{\\Delta m^{2}_{41}, \\sin^2\\theta_{24}\\}$$. The results presented can be compared to the parameter space suggested by LSND and MiniBooNE and complements other previous experimental searches for sterile neutrinos in the electron neutrino appearance channel.« less
78 FR 5838 - NRC Enforcement Policy
Federal Register 2010, 2011, 2012, 2013, 2014
2013-01-28
... submit comments by any of the following methods: Federal Rulemaking Web site: Go to http://www... of the following methods: Federal Rulemaking Web site: Go to http://www.regulations.gov and search... the search, select ``ADAMS Public Documents'' and then select ``Begin Web-based ADAMS Search.'' For...
78 FR 69710 - Luminant Generation Company, LLC
Federal Register 2010, 2011, 2012, 2013, 2014
2013-11-20
... methods: Federal Rulemaking Web site: Go to http://www.regulations.gov and search for Docket ID NRC-2008... . To begin the search, select ``ADAMS Public Documents'' and then select ``Begin Web- based ADAMS Search.'' For problems with ADAMS, please contact the NRC's Public [[Page 69711
Human Interface to Netcentricity
2006-06-01
experiencing. This is a radically different approach than using a federated search engine to bring back all relevant documents. The search engine...not be any closer to answering their question. More importantly, if they only have access to a 22 federated search , the program does not have the
Chemical-text hybrid search engines.
Zhou, Yingyao; Zhou, Bin; Jiang, Shumei; King, Frederick J
2010-01-01
As the amount of chemical literature increases, it is critical that researchers be enabled to accurately locate documents related to a particular aspect of a given compound. Existing solutions, based on text and chemical search engines alone, suffer from the inclusion of "false negative" and "false positive" results, and cannot accommodate diverse repertoire of formats currently available for chemical documents. To address these concerns, we developed an approach called Entity-Canonical Keyword Indexing (ECKI), which converts a chemical entity embedded in a data source into its canonical keyword representation prior to being indexed by text search engines. We implemented ECKI using Microsoft Office SharePoint Server Search, and the resultant hybrid search engine not only supported complex mixed chemical and keyword queries but also was applied to both intranet and Internet environments. We envision that the adoption of ECKI will empower researchers to pose more complex search questions that were not readily attainable previously and to obtain answers at much improved speed and accuracy.
Marketing to America's youth: evidence from corporate documents.
Cummings, K M; Morley, C P; Horan, J K; Steger, C; Leavell, N-R
2002-03-01
To evaluate the claim that the tobacco industry does not market its products to youth. The data for this study come from tobacco industry documents collected from the tobacco industry's document websites, presently linked at http://www.tobaccoarchives.com. The websites were searched using "request for production" (RFP) codes, specified keyword searches, and serendipitous terms identified in document citations found with RFP and keyword searches. Industry documents show that the cigarette manufacturers carefully monitored the smoking habits of teenagers over the past several decades. Candid quotes from industry executives refer to youth as a source of sales and as fundamental to the survival of the tobacco industry. The documents reveal that the features of cigarette brands (that is, use of filters, low tar, bland taste, etc), packaging (that is, size, colour and design), and advertising (that is, media placements and themes and imagery) were developed specifically to appeal to new smokers (that is, teenagers). Evidence also indicates that relevant youth oriented marketing documents may have been destroyed and that the language used in some of the more recent documents may have been sanitised to cover up efforts to market to youth. The tobacco industry's internal documents reveal an undeniable interest in marketing cigarettes to underage smokers. The industry's marketing approaches run counter to and predicate methods for tobacco prevention: (1) keep the price of the product high; (2) keep product placements and advertising away from schools and other areas with a high volume of youth traffic; (3) make cigarette advertising (that is, themes and visual images) unappealing to youth; (4) make product packaging unappealing to youth; and (5) design the product so it is not easy to inhale.
Search Search Home SH Reference Manual E19 Documentation Program Management Training/Drills Other River and Lake Summary (RVD) SH Reference Manual, E-19 Docs, Program Management, Training/Drills, Other
Searching Lexis and Westlaw: Part III.
ERIC Educational Resources Information Center
Franklin, Carl
1986-01-01
This last installment in a three-part series covers several important areas in the searching of legal information: online (group) training and customer service, documentation (search manuals and other aids), account representatives, microcomputer software, and pricing. Advantages and drawbacks of both the LEXIS and WESTLAW databases are noted.…
National Centers for Environmental Prediction
Organization Search Enter text Search Navigation Bar End Cap Search EMC Go Branches Global Climate and Weather Modeling Mesoscale Modeling Marine Modeling and Analysis Teams Climate Data Assimilation Ensembles and Post Model Configuration Collaborators Documentation and Code FAQ Operational Change Log Parallel Experiment
Data Warehouse Governance Programs in Healthcare Settings: A Literature Review and a Call to Action
Elliott, Thomas E.; Holmes, John H.; Davidson, Arthur J.; La Chance, Pierre-Andre; Nelson, Andrew F.; Steiner, John F.
2013-01-01
Purpose: Given the extensive data stored in healthcare data warehouses, data warehouse governance policies are needed to ensure data integrity and privacy. This review examines the current state of the data warehouse governance literature as it applies to healthcare data warehouses, identifies knowledge gaps, provides recommendations, and suggests approaches for further research. Methods: A comprehensive literature search using five data bases, journal article title-search, and citation searches was conducted between 1997 and 2012. Data warehouse governance documents from two healthcare systems in the USA were also reviewed. A modified version of nine components from the Data Governance Institute Framework for data warehouse governance guided the qualitative analysis. Results: Fifteen articles were retrieved. Only three were related to healthcare settings, each of which addressed only one of the nine framework components. Of the remaining 12 articles, 10 addressed between one and seven framework components and the remainder addressed none. Each of the two data warehouse governance plans obtained from healthcare systems in the USA addressed a subset of the framework components, and between them they covered all nine. Conclusions: While published data warehouse governance policies are rare, the 15 articles and two healthcare organizational documents reviewed in this study may provide guidance to creating such policies. Additional research is needed in this area to ensure that data warehouse governance polices are feasible and effective. The gap between the development of data warehouses in healthcare settings and formal governance policies is substantial, as evidenced by the sparse literature in this domain. PMID:25848561
Managing Content in a Matter of Minutes
NASA Technical Reports Server (NTRS)
2004-01-01
NASA software created to help scientists expeditiously search and organize their research documents is now aiding compliance personnel, law enforcement investigators, and the general public in their efforts to search, store, manage, and retrieve documents more efficiently. Developed at Ames Research Center, NETMARK software was designed to manipulate vast amounts of unstructured and semi-structured NASA documents. NETMARK is both a relational and object-oriented technology built on an Oracle enterprise-wide database. To ensure easy user access, Ames constructed NETMARK as a Web-enabled platform utilizing the latest in Internet technology. One of the significant benefits of the program was its ability to store and manage mission-critical data.
Recommending images of user interests from the biomedical literature
NASA Astrophysics Data System (ADS)
Clukey, Steven; Xu, Songhua
2013-03-01
Every year hundreds of thousands of biomedical images are published in journals and conferences. Consequently, finding images relevant to one's interests becomes an ever daunting task. This vast amount of literature creates a need for intelligent and easy-to-use tools that can help researchers effectively navigate through the content corpus and conveniently locate materials of their interests. Traditionally, literature search tools allow users to query content using topic keywords. However, manual query composition is often time and energy consuming. A better system would be one that can automatically deliver relevant content to a researcher without having the end user manually manifest one's search intent and interests via search queries. Such a computer-aided assistance for information access can be provided by a system that first determines a researcher's interests automatically and then recommends images relevant to the person's interests accordingly. The technology can greatly improve a researcher's ability to stay up to date in their fields of study by allowing them to efficiently browse images and documents matching their needs and interests among the vast amount of the biomedical literature. A prototype system implementation of the technology can be accessed via http://www.smartdataware.com.
Nakamura, R; Sasaki, M; Oikawa, H; Harada, S; Tamakawa, Y
2000-03-01
To use an intranet technique to develop an information system that simultaneously supports both diagnostic reports and radiotherapy planning images. Using a file server as the gateway a radiation oncology LAN was connected to an already operative RIS LAN. Dose-distribution images were saved in tagged-image-file format by way of a screen dump to the file server. X-ray simulator images and portal images were saved in encapsulated postscript format in the file server and automatically converted to portable document format. The files on the file server were automatically registered to the Web server by the search engine and were available for searching and browsing using the Web browser. It took less than a minute to register planning images. For clients, searching and browsing the file took less than 3 seconds. Over 150,000 reports and 4,000 images from a six-month period were accessible. Because the intranet technique was used, construction and maintenance was completed without specialty. Prompt access to essential information about radiotherapy has been made possible by this system. It promotes public access to radiotherapy planning that may improve the quality of treatment.
First Draft of the act Programming Language
2014-01-22
International Conference on, pages 1–15, 2013. [4] Yang Zhao, Jie Liu, and Edward A. Lee. A programming model for time - synchronized distributed real- time ...including the time for reviewing instructions, searching existing data sources, gathering and maintaining the data needed, and completing and reviewing...report is to document the first complete attempt at the design of a high-level programming language for timed systems called act. We define the lexical
ERIC Educational Resources Information Center
Moore, Kristina M., Ed.; And Others
This document contains 27 essays and 5 policy statements regarding creation of a comprehensive and unified federal employment and training policy for youth and adults. The following essays are included: "In Search of the American Way" (Wills); "Rethink the Demand Side" (Smith); "The Federal Government as a Change Agent" (Nathan); "Making Sense of…
Medical Surveillance Monthly Report. Volume 23, Number 8
2016-08-01
to be among the five most frequent diagnoses of digestive disorders associated with all medical encounters, with male hospitalizations, and with...male outpatient encounters.4–6 Medical accession standards specify that applicants for mili- tary service do not meet entrance standards if they have...Records of both inpatient and out- patient health care documented in the data- bases of the Defense Medical Surveillance System (DMSS) were searched to
XML syntax for clinical laboratory procedure manuals.
Saadawi, Gilan; Harrison, James H
2003-01-01
We have developed a document type description (DTD) in Extensable Markup Language (XML) for clinical laboratory procedures. Our XML syntax can adequately structure a variety of procedure types across different laboratories and is compatible with current procedure standards. The combination of this format with an XML content management system and appropriate style sheets will allow efficient procedure maintenance, distributed access, customized display and effective searching across a large body of test information.
SPECTRa-T: machine-based data extraction and semantic searching of chemistry e-theses.
Downing, Jim; Harvey, Matt J; Morgan, Peter B; Murray-Rust, Peter; Rzepa, Henry S; Stewart, Diana C; Tonge, Alan P; Townsend, Joe A
2010-02-22
The SPECTRa-T project has developed text-mining tools to extract named chemical entities (NCEs), such as chemical names and terms, and chemical objects (COs), e.g., experimental spectral assignments and physical chemistry properties, from electronic theses (e-theses). Although NCEs were readily identified within the two major document formats studied, only the use of structured documents enabled identification of chemical objects and their association with the relevant chemical entity (e.g., systematic chemical name). A corpus of theses was analyzed and it is shown that a high degree of semantic information can be extracted from structured documents. This integrated information has been deposited in a persistent Resource Description Framework (RDF) triple-store that allows users to conduct semantic searches. The strength and weaknesses of several document formats are reviewed.
Policies and Programs for Prevention and Control of Diabetes in Iran: A Document Analysis.
Faraji, Obeidollah; Etemad, Koorosh; Akbari Sari, Ali; Ravaghi, Hamid
2015-04-19
Trend analysis in 2005 to 2011 showed high growth in diabetes prevalence in Iran. Considering the high prevalence of diabetes in the country and likely to increase its prevalence in the future, the analysis of diabetes-related policies and programs is very important and effective in the prevention and control of diabetes. Therefore, the aim of the study was an analysis of policies and programs related to prevention and control of diabetes in Iran in 2014. This study was a policy analysis using deductive thematic content analysis of key documents. The health policy triangle framework was used in the data analysis. PubMed and ScienceDirect databases were searched to find relevant studies and documents. Also, hand searching was conducted among references of the identified studies. MAXQDA 10 software was used to organize and analyze data. The main reasons to take into consideration diabetes in Iran can be World Health Organization (WHO) report in 1989, and high prevalence of diabetes in the country. The major challenges in implementing the diabetes program include difficulty in referral levels of the program, lack of coordination between the private sector and the public sector and the limitations of reporting system in the specialized levels of the program. Besides strengthening referral system, the government should allocate more funds to the program and more importance to the educational programs for the public. Also, Non-Governmental Organizations (NGOs) and the private sector should involve in the formulation and implementation of the prevention and control programs of diabetes in the future.
Clustering and Recurring Anomaly Identification: Recurring Anomaly Detection System (ReADS)
NASA Technical Reports Server (NTRS)
McIntosh, Dawn
2006-01-01
This viewgraph presentation reviews the Recurring Anomaly Detection System (ReADS). The Recurring Anomaly Detection System is a tool to analyze text reports, such as aviation reports and maintenance records: (1) Text clustering algorithms group large quantities of reports and documents; Reduces human error and fatigue (2) Identifies interconnected reports; Automates the discovery of possible recurring anomalies; (3) Provides a visualization of the clusters and recurring anomalies We have illustrated our techniques on data from Shuttle and ISS discrepancy reports, as well as ASRS data. ReADS has been integrated with a secure online search
Social tagging in the life sciences: characterizing a new metadata resource for bioinformatics.
Good, Benjamin M; Tennis, Joseph T; Wilkinson, Mark D
2009-09-25
Academic social tagging systems, such as Connotea and CiteULike, provide researchers with a means to organize personal collections of online references with keywords (tags) and to share these collections with others. One of the side-effects of the operation of these systems is the generation of large, publicly accessible metadata repositories describing the resources in the collections. In light of the well-known expansion of information in the life sciences and the need for metadata to enhance its value, these repositories present a potentially valuable new resource for application developers. Here we characterize the current contents of two scientifically relevant metadata repositories created through social tagging. This investigation helps to establish how such socially constructed metadata might be used as it stands currently and to suggest ways that new social tagging systems might be designed that would yield better aggregate products. We assessed the metadata that users of CiteULike and Connotea associated with citations in PubMed with the following metrics: coverage of the document space, density of metadata (tags) per document, rates of inter-annotator agreement, and rates of agreement with MeSH indexing. CiteULike and Connotea were very similar on all of the measurements. In comparison to PubMed, document coverage and per-document metadata density were much lower for the social tagging systems. Inter-annotator agreement within the social tagging systems and the agreement between the aggregated social tagging metadata and MeSH indexing was low though the latter could be increased through voting. The most promising uses of metadata from current academic social tagging repositories will be those that find ways to utilize the novel relationships between users, tags, and documents exposed through these systems. For more traditional kinds of indexing-based applications (such as keyword-based search) to benefit substantially from socially generated metadata in the life sciences, more documents need to be tagged and more tags are needed for each document. These issues may be addressed both by finding ways to attract more users to current systems and by creating new user interfaces that encourage more collectively useful individual tagging behaviour.
Improving Communication About Serious Illness in Primary Care: A Review.
Lakin, Joshua R; Block, Susan D; Billings, J Andrew; Koritsanszky, Luca A; Cunningham, Rebecca; Wichmann, Lisa; Harvey, Doreen; Lamey, Jan; Bernacki, Rachelle E
2016-09-01
The Institute of Medicine recently called for systematic improvements in clinician-led conversations about goals, values, and care preferences for patients with serious and life-threatening illnesses. Studies suggest that these conversations are associated with improved outcomes for patients and their families, enhanced clinician satisfaction, and lower health care costs; however, the role of primary care clinicians in driving conversations about goals and priorities in serious illness is not well defined. To present a review of a structured search of the evidence base about communication in serious illness in primary care. MEDLINE was searched, via PubMed, on January 19, 2016, finding 911 articles; 126 articles were reviewed and selected titles were added from bibliography searches. Review of the literature informed 2 major topic areas: the role of primary care in communication about serious illness and clinician barriers and system failures that interfere with effective communication. Literature regarding the role that primary care plays in communication focused primarily on the ambiguity about whether primary care clinicians or specialists are responsible for initiating conversations, the benefits of primary care clinicians and specialists conducting conversations, and the quantity and quality of discussions. Timely and effective communication about serious illness in primary care is hampered by key clinician barriers, which include deficits in knowledge, skills, and attitudes; discomfort with prognostication; and lack of clarity about the appropriate timing and initiation of conversations. Finally, system failures in coordination, documentation, feedback, and quality improvement contribute to lack of conversations. Clinician and system barriers will challenge primary care clinicians and institutions to meet the needs of patients with serious illness. Ensuring that conversations about goals and values occur at the appropriate time for seriously ill patients will require improved training, validation, and dissemination of patient selection tools, systems for conducting and revisiting conversations, accessible documentation, and incentives for measurement, feedback, and continuous improvement.
... on the relevance score as determined by the search engine. Generally, the first document in the first results ... Spanish . snippet Brief result summary generated by the search engine that provides a preview of the relevant content ...
Code of Federal Regulations, 2013 CFR
2013-04-01
... use request, fees will be limited to reasonable standard charges for document search, review, and... reasonable standard charges for search time after the first 2 hours and duplication after the first 100 pages... of the Peace Corps records is as follows: (1) Manual search and review of records: Salary rate of...
31 CFR 1.7 - Fees for services.
Code of Federal Regulations, 2011 CFR
2011-07-01
... direct costs of searching for, reviewing, and duplicating the records sought. Commercial use requesters are not entitled to two hours of free search time or 100 free pages of duplication of documents... searching for and reviewing records even if there is ultimately no disclosure of records, or no records are...
31 CFR 1.7 - Fees for services.
Code of Federal Regulations, 2014 CFR
2014-07-01
... direct costs of searching for, reviewing, and duplicating the records sought. Commercial use requesters are not entitled to two hours of free search time or 100 free pages of duplication of documents... searching for and reviewing records even if there is ultimately no disclosure of records, or no records are...
31 CFR 1.7 - Fees for services.
Code of Federal Regulations, 2012 CFR
2012-07-01
... direct costs of searching for, reviewing, and duplicating the records sought. Commercial use requesters are not entitled to two hours of free search time or 100 free pages of duplication of documents... searching for and reviewing records even if there is ultimately no disclosure of records, or no records are...
76 FR 6181 - Information Collection Available for Public Comments and Recommendations
Federal Register 2010, 2011, 2012, 2013, 2014
2011-02-03
....regulations.gov/search/index.jsp . Specifically address whether this information collection is necessary for... version of this document is available on the World Wide Web at http://www.regulations.gov/search/index.jsp... visit http://www.regulations.gov/search/index.jsp . By Order of the Maritime Administrator. Dated...
Interactive Information Organization: Techniques and Evaluation
2001-05-01
information search and access. Locating interesting information on the World Wide Web is the main task of on-line search engines . Such engines accept a...likelihood of being relevant to the user’s request. The majority of today’s Web search engines follow this scenario. The ordering of documents in the
24 CFR Appendix B to Part 3500 - Illustrations of Requirements of RESPA
Code of Federal Regulations, 2014 CFR
2014-04-01
... title searches and related settlement services in connection with their transaction. B and A have an... title search and examination work, determines insurability of title, prepares documents containing... title search to determine insurability of the title, and the issuance of a title commitment where...
77 FR 26321 - Virginia Electric and Power Company
Federal Register 2010, 2011, 2012, 2013, 2014
2012-05-03
... NUCLEAR REGULATORY COMMISSION [Docket Nos. 50-338 and 50-339; NRC-2012-0051; License Nos. NPF-4...: Federal Rulemaking Web Site: Go to http://www.regulations.gov and search for Docket ID NRC-2012-0051... search, select ``ADAMS Public Documents'' and then select ``Begin Web- based ADAMS Search.'' For problems...
78 FR 70964 - Luminant Generation Company, LLC
Federal Register 2010, 2011, 2012, 2013, 2014
2013-11-27
... search for Docket ID NRC-2008-0594. Address questions about NRC dockets to Carol Gallagher; telephone... at http://www.nrc.gov/reading-rm/adams.html . To begin the search, select ``ADAMS Public Documents'' and then select ``Begin Web- based ADAMS Search.'' For problems with ADAMS, please contact the NRC's...
National Centers for Environmental Prediction
Organization Search Enter text Search Navigation Bar End Cap Search EMC Go Branches Global Climate and Weather Modeling Mesoscale Modeling Marine Modeling and Analysis Teams Climate Data Assimilation Ensembles and Post Configuration Collaborators Documentation and Code FAQ Operational Change Log Parallel Experiment Change Log
National Centers for Environmental Prediction
Organization Search Enter text Search Navigation Bar End Cap Search EMC Go Branches Global Climate and Weather Modeling Mesoscale Modeling Marine Modeling and Analysis Teams Climate Data Assimilation Ensembles and Post Collaborators Documentation and Code FAQ Operational Change Log Parallel Experiment Change Log Contacts
77 FR 31169 - Airworthiness Directives; Piper Aircraft, Inc. Airplanes
Federal Register 2010, 2011, 2012, 2013, 2014
2012-05-25
... detailed search for all applicable airworthiness related documents that apply to any airplane that has an incorrectly marked data plate and take necessary corrective actions based on the search findings. We are... affected model airplanes. The NPRM also proposed to require a detailed search for all applicable...
DOE Office of Scientific and Technical Information (OSTI.GOV)
Paul L. Wichlacz; Robert C. Starr; Brennon Orr
2003-09-01
This document summarizes previous descriptions of geochemical system conceptual models for the vadose zone and groundwater zone (aquifer) beneath the Idaho National Engineering and Environmental Laboratory (INEEL). The primary focus is on groundwater because contaminants derived from wastes disposed at INEEL are present in groundwater, groundwater provides a pathway for potential migration to receptors, and because geochemical characteristics in and processes in the aquifer can substantially affect the movement, attenuation, and toxicity of contaminants. The secondary emphasis is perched water bodies in the vadose zone. Perched water eventually reaches the regional groundwater system, and thus processes that affect contaminants inmore » the perched water bodies are important relative to the migration of contaminants into groundwater. Similarly, processes that affect solutes during transport from nearsurface disposal facilities downward through the vadose zone to the aquifer are relevant. Sediments in the vadose zone can affect both water and solute transport by restricting the downward migration of water sufficiently that a perched water body forms, and by retarding solute migration via ion exchange. Geochemical conceptual models have been prepared by a variety of researchers for different purposes. They have been published in documents prepared by INEEL contractors, the United States Geological Survey (USGS), academic researchers, and others. The documents themselves are INEEL and USGS reports, and articles in technical journals. The documents reviewed were selected from citation lists generated by searching the INEEL Technical Library, the INEEL Environmental Restoration Optical Imaging System, and the ISI Web of Science databases. The citation lists were generated using the keywords ground water, groundwater, chemistry, geochemistry, contaminant, INEL, INEEL, and Idaho. In addition, a list of USGS documents that pertain to the INEEL was obtained and manually searched. The documents that appeared to be the most pertinent were selected from further review. These documents are tabulated in the citation list. This report summarizes existing geochemical conceptual models, but does not attempt to generate a new conceptual model or select the ''right'' model. This document is organized as follows. Geochemical models are described in general in Section 2. Geochemical processes that control the transport and fate of contaminants introduced into groundwater are described in Section 3. The natural geochemistry of the Eastern Snake River Plain Aquifer (SRPA) is described in Section 4. The effect of waste disposal on the INEEL subsurface is described in Section 5. The geochemical behavior of the major contaminants is described in Section 6. Section 7 describes the site-specific geochemical models developed for various INEEL facilities.« less
Modeling Guru: Knowledge Base for NASA Modelers
NASA Astrophysics Data System (ADS)
Seablom, M. S.; Wojcik, G. S.; van Aartsen, B. H.
2009-05-01
Modeling Guru is an on-line knowledge-sharing resource for anyone involved with or interested in NASA's scientific models or High End Computing (HEC) systems. Developed and maintained by the NASA's Software Integration and Visualization Office (SIVO) and the NASA Center for Computational Sciences (NCCS), Modeling Guru's combined forums and knowledge base for research and collaboration is becoming a repository for the accumulated expertise of NASA's scientific modeling and HEC communities. All NASA modelers and associates are encouraged to participate and provide knowledge about the models and systems so that other users may benefit from their experience. Modeling Guru is divided into a hierarchy of communities, each with its own set forums and knowledge base documents. Current modeling communities include those for space science, land and atmospheric dynamics, atmospheric chemistry, and oceanography. In addition, there are communities focused on NCCS systems, HEC tools and libraries, and programming and scripting languages. Anyone may view most of the content on Modeling Guru (available at http://modelingguru.nasa.gov/), but you must log in to post messages and subscribe to community postings. The site offers a full range of "Web 2.0" features, including discussion forums, "wiki" document generation, document uploading, RSS feeds, search tools, blogs, email notification, and "breadcrumb" links. A discussion (a.k.a. forum "thread") is used to post comments, solicit feedback, or ask questions. If marked as a question, SIVO will monitor the thread, and normally respond within a day. Discussions can include embedded images, tables, and formatting through the use of the Rich Text Editor. Also, the user can add "Tags" to their thread to facilitate later searches. The "knowledge base" is comprised of documents that are used to capture and share expertise with others. The default "wiki" document lets users edit within the browser so others can easily collaborate on the same document, even allowing the author to select those who may edit and approve the document. To maintain knowledge integrity, all documents are moderated before they are visible to the public. Modeling Guru, running on Clearspace by Jive Software, has been an active resource to the NASA modeling and HEC communities for more than a year and currently has more than 100 active users. SIVO will soon install live instant messaging support, as well as a user-customizable homepage with social-networking features. In addition, SIVO plans to implement a large dataset/file storage capability so that users can quickly and easily exchange datasets and files with one another. Continued active community participation combined with periodic software updates and improved features will ensure that Modeling Guru remains a vibrant, effective, easy-to-use tool for the NASA scientific community.
Guidance Documents for Inert Ingredients
These guidance documents provide information on various inert ingredient issues, including the general process for submitting petitions or requests, adding trade names to our database, and doing searches related to inert ingredients.
Negotiating the Digital Library: Document Delivery.
ERIC Educational Resources Information Center
Jacobs, Neil; Morris, Anne
1999-01-01
The eLib-funded FIDDO (Focused Investigation of Document Delivery Options) project provides library managers/others with information to support policy decisions. Senior libraries were interviewed about the future of document delivery and interviews were analyzed with the support of NUD*IST (Nonnumerical Unstructured Data by Indexing, Searching and…
37 CFR 205.22 - Production of documents and testimony.
Code of Federal Regulations, 2012 CFR
2012-07-01
... 37 Patents, Trademarks, and Copyrights 1 2012-07-01 2012-07-01 false Production of documents and testimony. 205.22 Section 205.22 Patents, Trademarks, and Copyrights COPYRIGHT OFFICE, LIBRARY OF CONGRESS... fees associated with various document requests, searches, copies, and expedited handling are listed in...
37 CFR 205.22 - Production of documents and testimony.
Code of Federal Regulations, 2013 CFR
2013-07-01
... 37 Patents, Trademarks, and Copyrights 1 2013-07-01 2013-07-01 false Production of documents and testimony. 205.22 Section 205.22 Patents, Trademarks, and Copyrights COPYRIGHT OFFICE, LIBRARY OF CONGRESS... fees associated with various document requests, searches, copies, and expedited handling are listed in...
37 CFR 205.22 - Production of documents and testimony.
Code of Federal Regulations, 2011 CFR
2011-07-01
... 37 Patents, Trademarks, and Copyrights 1 2011-07-01 2011-07-01 false Production of documents and testimony. 205.22 Section 205.22 Patents, Trademarks, and Copyrights COPYRIGHT OFFICE, LIBRARY OF CONGRESS... fees associated with various document requests, searches, copies, and expedited handling are listed in...
Open Clients for Distributed Databases
NASA Astrophysics Data System (ADS)
Chayes, D. N.; Arko, R. A.
2001-12-01
We are actively developing a collection of open source example clients that demonstrate use of our "back end" data management infrastructure. The data management system is reported elsewhere at this meeting (Arko and Chayes: A Scaleable Database Infrastructure). In addition to their primary goal of being examples for others to build upon, some of these clients may have limited utility in them selves. More information about the clients and the data infrastructure is available on line at http://data.ldeo.columbia.edu. The available examples to be demonstrated include several web-based clients including those developed for the Community Review System of the Digital Library for Earth System Education, a real-time watch standers log book, an offline interface to use log book entries, a simple client to search on multibeam metadata and others are Internet enabled and generally web-based front ends that support searches against one or more relational databases using industry standard SQL queries. In addition to the web based clients, simple SQL searches from within Excel and similar applications will be demonstrated. By defining, documenting and publishing a clear interface to the fully searchable databases, it becomes relatively easy to construct client interfaces that are optimized for specific applications in comparison to building a monolithic data and user interface system.
Learning Design at White Sands Test Facility
NASA Technical Reports Server (NTRS)
Grotewiel, Shane
2010-01-01
During the Fall of 2010, I spent my time at NASA White Sands Test Facility in Las Cruces, NM as an Undergraduate Student Research Program (USRP) Intern. During that time, I was given three projects to work on: Large Altitude Simulation System (LASS) basket strainer, log books, and the design of a case for touch screen monitors used for simulations. I spent most of my time on the LASS basket strainer. The LASS system has a water feed line with a basket strainer that filters out rust. In 2009, there were three misfires which cost approximately $27,000 and about 8% of the allotted time. The strainer was getting a large change in pressure that would result in a shutdown of the system. I have designed a new basket that will eliminate the large pressure change and it can be used with the old basket strainer housing. The LASS system has three steam generators (modules). Documents pertaining to these modules are stored electronically, and the majority of the documents are not able to be searched with keywords, so they have to be gone through one by one. I have come up with an idea on how to organize these files so that the Propulsion Department may efficiently search through the documents needed. Propulsion also has a LASS simulator that incorporates two touch screen monitors. Currently these monitors are in six foot by two foot metal cabinet on wheels. During simulation these monitors are used in the block house and need to be taken out of the block house when not in use. I have designed different options for hand held cases for storing and transporting the monitors in and out of the block house. The three projects previously mentioned demonstrate my contributions to the Propulsion Department and have taught me real world experience that is essential in becoming a productive engineer.
US Geoscience Information Network, Web Services for Geoscience Information Discovery and Access
NASA Astrophysics Data System (ADS)
Richard, S.; Allison, L.; Clark, R.; Coleman, C.; Chen, G.
2012-04-01
The US Geoscience information network has developed metadata profiles for interoperable catalog services based on ISO19139 and the OGC CSW 2.0.2. Currently data services are being deployed for the US Dept. of Energy-funded National Geothermal Data System. These services utilize OGC Web Map Services, Web Feature Services, and THREDDS-served NetCDF for gridded datasets. Services and underlying datasets (along with a wide variety of other information and non information resources are registered in the catalog system. Metadata for registration is produced by various workflows, including harvest from OGC capabilities documents, Drupal-based web applications, transformation from tabular compilations. Catalog search is implemented using the ESRI Geoportal open-source server. We are pursuing various client applications to demonstrated discovery and utilization of the data services. Currently operational applications allow catalog search and data acquisition from map services in an ESRI ArcMap extension, a catalog browse and search application built on openlayers and Django. We are developing use cases and requirements for other applications to utilize geothermal data services for resource exploration and evaluation.
NASA Astrophysics Data System (ADS)
Yang, Liang-Chih; Lu, Hsi-Peng
This paper depicts a longitudinal investigation of knowledge management system development from industrial perspectives. Snapshots on three surveys (2002, 2006, and 2010) of Taiwanese companies were conducted and compared, which is to explore the perceived understandings and requirements for the applications of a knowledge management system.From the surveys, it was found that the most useful applications were document management, knowledge search and retrieval, and knowledge repository and map. The emerging applications were expert management, document security, and knowledge automation such as auto-classification, auto-abstract and auto-keyword generation. The most wanted services along with KMS were consulting service, success story-sharing, and modularization while deploying knowledge management system in the enterprises. The trends and transformation of a KM system were also collected and analyzed. We suggest that a company should use different knowledge management approach according to its corporate main business function. Combing intellectual capital theories proposed by other researchers, we categorize knowledge management focus as staff-centric, system-centric, and customer-centric knowledge from industrial perspectives.
Breeding novel solutions in the brain: a model of Darwinian neurodynamics.
Szilágyi, András; Zachar, István; Fedor, Anna; de Vladar, Harold P; Szathmáry, Eörs
2016-01-01
Background : The fact that surplus connections and neurons are pruned during development is well established. We complement this selectionist picture by a proof-of-principle model of evolutionary search in the brain, that accounts for new variations in theory space. We present a model for Darwinian evolutionary search for candidate solutions in the brain. Methods : We combine known components of the brain - recurrent neural networks (acting as attractors), the action selection loop and implicit working memory - to provide the appropriate Darwinian architecture. We employ a population of attractor networks with palimpsest memory. The action selection loop is employed with winners-share-all dynamics to select for candidate solutions that are transiently stored in implicit working memory. Results : We document two processes: selection of stored solutions and evolutionary search for novel solutions. During the replication of candidate solutions attractor networks occasionally produce recombinant patterns, increasing variation on which selection can act. Combinatorial search acts on multiplying units (activity patterns) with hereditary variation and novel variants appear due to (i) noisy recall of patterns from the attractor networks, (ii) noise during transmission of candidate solutions as messages between networks, and, (iii) spontaneously generated, untrained patterns in spurious attractors. Conclusions : Attractor dynamics of recurrent neural networks can be used to model Darwinian search. The proposed architecture can be used for fast search among stored solutions (by selection) and for evolutionary search when novel candidate solutions are generated in successive iterations. Since all the suggested components are present in advanced nervous systems, we hypothesize that the brain could implement a truly evolutionary combinatorial search system, capable of generating novel variants.
Ilic, Nina; Savic, Snezana; Siegel, Evan; Atkinson, Kerry; Tasic, Ljiljana
2012-12-01
Recent development of a wide range of regulatory standards applicable to production and use of tissues, cells, and other biologics (or biologicals), as advanced therapies, indicates considerable interest in the regulation of these products. The objective of this study was to analyze and compare high-tier documents within the Australian, European, and U.S. biologic drug regulatory environments using qualitative methodology. Cohort 1 of the selected 18 high-tier regulatory documents from the European Medicines Agency (EMA), the U.S. Food and Drug Administration (FDA), and the Therapeutic Goods Administration (TGA) regulatory frameworks were subject to a manual documentary analysis. These documents were consistent with the legal requirements for manufacturing and use of biologic drugs in humans and fall into six different categories. Manual analysis included a terminology search. The occurrence, frequency, and interchangeable use of different terms and phrases were recorded in the manual documentary analysis. Despite obvious differences, manual documentary analysis revealed certain consistency in use of terminology across analyzed frameworks. Phrase search frequencies have shown less uniformity than the search of terms. Overall, the EMA framework's documents referred to "medicinal products" and "marketing authorization(s)," the FDA documents discussed "drug(s)" or "biologic(s)," and the TGA documents referred to "biological(s)." Although high-tier documents often use different terminology they share concepts and themes. Documents originating from the same source have more conjunction in their terminology although they belong to different frameworks (i.e., Good Clinical Practice requirements based on the Declaration of Helsinki, 1964). Automated (software-based) documentary analysis should be obtained for the conceptual and relational analysis.
Savic, Snezana; Siegel, Evan; Atkinson, Kerry; Tasic, Ljiljana
2012-01-01
Recent development of a wide range of regulatory standards applicable to production and use of tissues, cells, and other biologics (or biologicals), as advanced therapies, indicates considerable interest in the regulation of these products. The objective of this study was to analyze and compare high-tier documents within the Australian, European, and U.S. biologic drug regulatory environments using qualitative methodology. Cohort 1 of the selected 18 high-tier regulatory documents from the European Medicines Agency (EMA), the U.S. Food and Drug Administration (FDA), and the Therapeutic Goods Administration (TGA) regulatory frameworks were subject to a manual documentary analysis. These documents were consistent with the legal requirements for manufacturing and use of biologic drugs in humans and fall into six different categories. Manual analysis included a terminology search. The occurrence, frequency, and interchangeable use of different terms and phrases were recorded in the manual documentary analysis. Despite obvious differences, manual documentary analysis revealed certain consistency in use of terminology across analyzed frameworks. Phrase search frequencies have shown less uniformity than the search of terms. Overall, the EMA framework's documents referred to “medicinal products” and “marketing authorization(s),” the FDA documents discussed “drug(s)” or “biologic(s),” and the TGA documents referred to “biological(s).” Although high-tier documents often use different terminology they share concepts and themes. Documents originating from the same source have more conjunction in their terminology although they belong to different frameworks (i.e., Good Clinical Practice requirements based on the Declaration of Helsinki, 1964). Automated (software-based) documentary analysis should be obtained for the conceptual and relational analysis. PMID:23283551
2015-01-01
Background PubMed is the largest biomedical bibliographic information source on the Internet. PubMed has been considered one of the most important and reliable sources of up-to-date health care evidence. Previous studies examined the effects of domain expertise/knowledge on search performance using PubMed. However, very little is known about PubMed users’ knowledge of information retrieval (IR) functions and their usage in query formulation. Objective The purpose of this study was to shed light on how experienced/nonexperienced PubMed users perform their search queries by analyzing a full-day query log. Our hypotheses were that (1) experienced PubMed users who use system functions quickly retrieve relevant documents and (2) nonexperienced PubMed users who do not use them have longer search sessions than experienced users. Methods To test these hypotheses, we analyzed PubMed query log data containing nearly 3 million queries. User sessions were divided into two categories: experienced and nonexperienced. We compared experienced and nonexperienced users per number of sessions, and experienced and nonexperienced user sessions per session length, with a focus on how fast they completed their sessions. Results To test our hypotheses, we measured how successful information retrieval was (at retrieving relevant documents), represented as the decrease rates of experienced and nonexperienced users from a session length of 1 to 2, 3, 4, and 5. The decrease rate (from a session length of 1 to 2) of the experienced users was significantly larger than that of the nonexperienced groups. Conclusions Experienced PubMed users retrieve relevant documents more quickly than nonexperienced PubMed users in terms of session length. PMID:26139516
37 CFR 205.22 - Production of documents and testimony.
Code of Federal Regulations, 2014 CFR
2014-07-01
... 37 Patents, Trademarks, and Copyrights 1 2014-07-01 2014-07-01 false Production of documents and testimony. 205.22 Section 205.22 Patents, Trademarks, and Copyrights U.S. COPYRIGHT OFFICE, LIBRARY OF... fees associated with various document requests, searches, copies, and expedited handling are listed in...
EMERSE: The Electronic Medical Record Search Engine
Hanauer, David A.
2006-01-01
EMERSE (The Electronic Medical Record Search Engine) is an intuitive, powerful search engine for free-text documents in the electronic medical record. It offers multiple options for creating complex search queries yet has an interface that is easy enough to be used by those with minimal computer experience. EMERSE is ideal for retrospective chart reviews and data abstraction and may have potential for clinical care as well.
Health system strategies supporting transition to adult care.
Hepburn, Charlotte Moore; Cohen, Eyal; Bhawra, Jasmin; Weiser, Natalie; Hayeems, Robin Z; Guttmann, Astrid
2015-06-01
The transition from paediatric to adult care is associated with poor clinical outcomes, increased costs and low patient and family satisfaction. However, little is known about health system strategies to streamline and safeguard care for youth transitioning to adult services. Moreover, the needs of children and youth are often excluded from broader health system reform discussions, leaving this population especially vulnerable to system 'disintegration'. (1) To explore the international policy profile of paediatric-to-adult care transitions, and (2) to document policy objectives, initiatives and outcomes for jurisdictions publicly committed to addressing transition issues. An international policy scoping review of all publicly available government documents detailing transition-related strategies was completed using a web-based search. Our analysis included a comparable cohort of nine wealthy Organisation for Economic Co-operation and Development (OECD) jurisdictions with Beveridge-style healthcare systems (deemed those most likely to benefit from system-level transition strategies). Few jurisdictions address transition of care issues in either health or broader social policy documents. While many jurisdictions refer to standardised practice guidelines, a few report the intention to use powerful policy levers (including physician remuneration and non-physician investments) to facilitate the uptake of best practice. Most jurisdictions do not address the policy infrastructure required to support successful transitions, and rigorous evaluations of transition strategies are rare. Despite the well-documented risks and costs associated with a poor transition from paediatric to adult care, little policy attention has been paid to this issue. We recommend that healthcare providers engage health system planners in the design and evaluation of system-level, policy-sensitive transition strategies. Published by the BMJ Publishing Group Limited. For permission to use (where not already granted under a licence) please go to http://group.bmj.com/group/rights-licensing/permissions.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Hummel, K.E.
1987-12-01
Expert systems are artificial intelligence programs that solve problems requiring large amounts of heuristic knowledge, based on years of experience and tradition. Production systems are domain-independent tools that support the development of rule-based expert systems. This document describes a general purpose production system known as HERB. This system was developed to support the programming of expert systems using hierarchically structured rule bases. HERB encourages the partitioning of rules into multiple rule bases and supports the use of multiple conflict resolution strategies. Multiple rule bases can also be placed on a system stack and simultaneously searched during each interpreter cycle. Bothmore » backward and forward chaining rules are supported by HERB. The condition portion of each rule can contain both patterns, which are matched with facts in a data base, and LISP expressions, which are explicitly evaluated in the LISP environment. Properties of objects can also be stored in the HERB data base and referenced within the scope of each rule. This document serves both as an introduction to the principles of LISP-based production systems and as a user's manual for the HERB system. 6 refs., 17 figs.« less
A robust omnifont open-vocabulary Arabic OCR system using pseudo-2D-HMM
NASA Astrophysics Data System (ADS)
Rashwan, Abdullah M.; Rashwan, Mohsen A.; Abdel-Hameed, Ahmed; Abdou, Sherif; Khalil, A. H.
2012-01-01
Recognizing old documents is highly desirable since the demand for quickly searching millions of archived documents has recently increased. Using Hidden Markov Models (HMMs) has been proven to be a good solution to tackle the main problems of recognizing typewritten Arabic characters. These attempts however achieved a remarkable success for omnifont OCR under very favorable conditions, they didn't achieve the same performance in practical conditions, i.e. noisy documents. In this paper we present an omnifont, large-vocabulary Arabic OCR system using Pseudo Two Dimensional Hidden Markov Model (P2DHMM), which is a generalization of the HMM. P2DHMM offers a more efficient way to model the Arabic characters, such model offer both minimal dependency on the font size/style (omnifont), and high level of robustness against noise. The evaluation results of this system are very promising compared to a baseline HMM system and best OCRs available in the market (Sakhr and NovoDynamics). The recognition accuracy of the P2DHMM classifier is measured against the classic HMM classifier, the average word accuracy rates for P2DHMM and HMM classifiers are 79% and 66% respectively. The overall system accuracy is measured against Sakhr and NovoDynamics OCR systems, the average word accuracy rates for P2DHMM, NovoDynamics, and Sakhr are 74%, 71%, and 61% respectively.
Graphics-based intelligent search and abstracting using Data Modeling
NASA Astrophysics Data System (ADS)
Jaenisch, Holger M.; Handley, James W.; Case, Carl T.; Songy, Claude G.
2002-11-01
This paper presents an autonomous text and context-mining algorithm that converts text documents into point clouds for visual search cues. This algorithm is applied to the task of data-mining a scriptural database comprised of the Old and New Testaments from the Bible and the Book of Mormon, Doctrine and Covenants, and the Pearl of Great Price. Results are generated which graphically show the scripture that represents the average concept of the database and the mining of the documents down to the verse level.
The Search for an Advanced Fighter: A History from the XF-108 to the Advanced Tactical Fighter
1986-04-01
V, ,tt AIR COMMAND AND- STAFF COLLEGE STUDENT REPORT THE SEARCH FOR AN ADVANCED FIGHTER, A HISTORY FROM THE XF-108 TO THE &ELECTE j - MAJOR ROBERT P...expressed in this document are those of the author. They are- J not intended and should not be thought to represent official ideas, attitudes, or policies of...the general public. A loan copy of the document may be obtained from the Air University Interlibrary Loan Service (AULILDEX, Maxwell AFB, Alabama
NASA Technical Reports Server (NTRS)
Warren, Wayne H., Jr.
1990-01-01
A list of 1186 ultraviolet-excess objects (designated KUV) was compiled as a result of a search conducted with the 105-cm Schmidt telescope of the Kiso station of the Tokyo Astronomical Observatory. This document describes the machine readable version of the KUV survey list and presents a sample listing showing the logical records as they are recorded in the machine readable catalog. The KUV data include equatorial coordinates, magnitudes, color indices, and identifications for previously cataloged objects.
16 CFR 803.3 - Statement of reasons for noncompliance.
Code of Federal Regulations, 2011 CFR
2011-01-01
... made to obtain such information and documents, including the names of persons who searched for required... and all facts relied on in support thereof, including the identity of each document, its author...
16 CFR 803.3 - Statement of reasons for noncompliance.
Code of Federal Regulations, 2010 CFR
2010-01-01
... made to obtain such information and documents, including the names of persons who searched for required... and all facts relied on in support thereof, including the identity of each document, its author...
77 FR 33786 - NRC Enforcement Policy Revision
Federal Register 2010, 2011, 2012, 2013, 2014
2012-06-07
... methods: Federal Rulemaking Web site: Go to http://www.regulations.gov and search for Docket ID NRC-2011... search, select ``ADAMS Public Documents'' and then select ``Begin Web- based ADAMS Search.'' For problems... either 2.3.2.a. or b. must be met for the disposition of a violation as an NCV.'' The following new...
16 CFR 4.8 - Costs for obtaining Commission records.
Code of Federal Regulations, 2013 CFR
2013-01-01
... section: (1) The term search includes all time spent looking, manually or by automated means, for material... use requesters will be charged for the direct costs to search for, review, and duplicate documents. A... public. (3) Other requesters. Other requesters will be charged for the direct costs to search for and...
16 CFR 4.8 - Costs for obtaining Commission records.
Code of Federal Regulations, 2014 CFR
2014-01-01
... section: (1) The term search includes all time spent looking, manually or by automated means, for material... use requesters will be charged for the direct costs to search for, review, and duplicate documents. A... public. (3) Other requesters. Other requesters will be charged for the direct costs to search for and...
16 CFR 4.8 - Costs for obtaining Commission records.
Code of Federal Regulations, 2012 CFR
2012-01-01
... section: (1) The term search includes all time spent looking, manually or by automated means, for material... use requesters will be charged for the direct costs to search for, review, and duplicate documents. A... public. (3) Other requesters. Other requesters will be charged for the direct costs to search for and...
75 FR 78340 - Information Collection Available for Public Comments and Recommendations
Federal Register 2010, 2011, 2012, 2013, 2014
2010-12-15
.../search/index.jsp . Specifically address whether this information collection is necessary for proper... version of this document is available on the World Wide Web at http://www.regulations.gov/search/index.jsp... visit http://www.regulations.gov/search/index.jsp . (Authority: 49 CFR 1.66) By Order of the Maritime...
76 FR 28845 - Information Collection Available for Public Comments and Recommendations
Federal Register 2010, 2011, 2012, 2013, 2014
2011-05-18
... ./search/index.jsp. Specifically address whether this information collection is necessary for proper... version of this document is available on the World Wide Web at http://www.regulations.gov/search/index,jsp... visit http://www.regulations.gov/search/index.jsp . Authority: 49 CFR 1.66. Dated: May 9, 2011. By Order...
78 FR 7818 - Duane Arnold Energy Center; Application for Amendment to Facility Operating License
Federal Register 2010, 2011, 2012, 2013, 2014
2013-02-04
... methods: Federal Rulemaking Web site: Go to http://www.regulations.gov and search for Docket ID NRC-2013... search, select ``ADAMS Public Documents'' and then select ``Begin Web- based ADAMS Search.'' For problems... INFORMATION CONTACT: Karl D. Feintuch, Project Manager, Office of Nuclear Reactor Regulation, U.S. Nuclear...
77 FR 67837 - Callaway Plant, Unit 1; Application for Amendment to Facility Operating License
Federal Register 2010, 2011, 2012, 2013, 2014
2012-11-14
... methods: Federal Rulemaking Web site: Go to http://www.regulations.gov and search for Docket ID NRC-2012... search, select ``ADAMS Public Documents'' and then select ``Begin Web- based ADAMS Search.'' For problems... INFORMATION CONTACT: Carl F. Lyon, Project Manager, Office of Nuclear Reactor Regulation, U.S. Nuclear...
Internet Search Engines - Fluctuations in Document Accessibility.
ERIC Educational Resources Information Center
Mettrop, Wouter; Nieuwenhuysen, Paul
2001-01-01
Reports an empirical investigation of the consistency of retrieval through Internet search engines. Evaluates 13 engines: AltaVista, EuroFerret, Excite, HotBot, InfoSeek, Lycos, MSN, NorthernLight, Snap, WebCrawler, and three national Dutch engines: Ilse, Search.nl and Vindex. The focus is on a characteristic related to size: the degree of…
Finding Information on the World Wide Web: The Retrieval Effectiveness of Search Engines.
ERIC Educational Resources Information Center
Pathak, Praveen; Gordon, Michael
1999-01-01
Describes a study that examined the effectiveness of eight search engines for the World Wide Web. Calculated traditional information-retrieval measures of recall and precision at varying numbers of retrieved documents to use as the bases for statistical comparisons of retrieval effectiveness. Also examined the overlap between search engines.…
Bea-Muñoz, M; Medina-Sánchez, M; Flórez-García, M T
2015-01-01
Patients with spinal cord injuries and their carers have access to leaflets on Internet that they can use as educational material to complement traditional forms of education. The aim of this study is to evaluate the readability of informative documents in Spanish, obtained from Internet and aimed at patients with spinal cord injuries. A search was made with the Google search engine using the following key words: recommendation, advice, guide, manual, self-care, education and information, adding spinal cord injury, paraplegia and tetraplegia to each of the terms. We analyzed the first 50 results of each search. The readability of the leaflets was studied with the Flesch-Szigriszt index and the INFLESZ scale, both available on the INFLESZ program. Also indicated were year of publication, country and number of authors of the documents obtained. We obtained 16 documents, developed between 2001 and 2011. Readability oscillated between 43.34 (some-what difficult) and 62 (normal), with an average value of 51.56 (somewhat difficult). Only 4 pamphlets (25%) showed a Flesch-Szigriszt index of ≥ 55 (normal). There was no difference in readability by year, authors or country of publication. The readability of 75% of the documents studied was "somewhat difficult" according to the INFLESZ scale. These results coincide with previous studies, in both Spanish and English. If the readability of this type of documents is improved, it will be easier to achieve their educational goal.
Development of a Search Strategy for an Evidence Based Retrieval Service
Ho, Gah Juan; Liew, Su May; Ng, Chirk Jenn; Hisham Shunmugam, Ranita; Glasziou, Paul
2016-01-01
Background Physicians are often encouraged to locate answers for their clinical queries via an evidence-based literature search approach. The methods used are often not clearly specified. Inappropriate search strategies, time constraint and contradictory information complicate evidence retrieval. Aims Our study aimed to develop a search strategy to answer clinical queries among physicians in a primary care setting Methods Six clinical questions of different medical conditions seen in primary care were formulated. A series of experimental searches to answer each question was conducted on 3 commonly advocated medical databases. We compared search results from a PICO (patients, intervention, comparison, outcome) framework for questions using different combinations of PICO elements. We also compared outcomes from doing searches using text words, Medical Subject Headings (MeSH), or a combination of both. All searches were documented using screenshots and saved search strategies. Results Answers to all 6 questions using the PICO framework were found. A higher number of systematic reviews were obtained using a 2 PICO element search compared to a 4 element search. A more optimal choice of search is a combination of both text words and MeSH terms. Despite searching using the Systematic Review filter, many non-systematic reviews or narrative reviews were found in PubMed. There was poor overlap between outcomes of searches using different databases. The duration of search and screening for the 6 questions ranged from 1 to 4 hours. Conclusion This strategy has been shown to be feasible and can provide evidence to doctors’ clinical questions. It has the potential to be incorporated into an interventional study to determine the impact of an online evidence retrieval system. PMID:27935993
2011-03-28
particular topic of interest. Paper -based documents require the availability of a physical instance of a document, involving the transport of documents...repository of documents via the World Wide Web and search engines offer support in locating documents that are likely to contain relevant information. The... Web , with news agencies, newspapers, various organizations, and individuals as sources. Clearly the analysis, interpretation, and integration of
NASA Astrophysics Data System (ADS)
Ozer, N.
2006-12-01
Researches aimed at enriching the number of available documentary sources on earthquakes have an important role in seismology. To this end, this paper documents the history of prominent earthquakes associated with the NW-SE trending Sultandag-Aksehir Fault and Aksehir-Afyon graben system in Western-Central Anatolia since the historical times through 1766. This work also combines the earthquake data for both historical and instrumental periods, previously listed in various catalogues and resources, for the studied area. Documents from the Ottoman archives and libraries as well as the Ottoman and Turkish newspapers were scrutinized, and eight previously unreported earthquakes in the latter half of the nineteenth century and four new earthquakes in the period 1900-1931 were revealed. For the period from 1766 to 1931, the total number of known earthquakes for the area under investigation increased from eighteen to thirty thanks to the document search. Furthermore, the existing information on eleven previously reported earthquakes is updated for the period from 1862 to 1946. Earthquakes from 1946 to 1964 are compiled from the catalogues for data completeness.
Usability Evaluation of an Unstructured Clinical Document Query Tool for Researchers.
Hultman, Gretchen; McEwan, Reed; Pakhomov, Serguei; Lindemann, Elizabeth; Skube, Steven; Melton, Genevieve B
2018-01-01
Natural Language Processing - Patient Information Extraction for Researchers (NLP-PIER) was developed for clinical researchers for self-service Natural Language Processing (NLP) queries with clinical notes. This study was to conduct a user-centered analysis with clinical researchers to gain insight into NLP-PIER's usability and to gain an understanding of the needs of clinical researchers when using an application for searching clinical notes. Clinical researcher participants (n=11) completed tasks using the system's two existing search interfaces and completed a set of surveys and an exit interview. Quantitative data including time on task, task completion rate, and survey responses were collected. Interviews were analyzed qualitatively. Survey scores, time on task and task completion proportions varied widely. Qualitative analysis indicated that participants found the system to be useful and usable in specific projects. This study identified several usability challenges and our findings will guide the improvement of NLP-PIER 's interfaces.
Historical literature review on waste classification and categorization
DOE Office of Scientific and Technical Information (OSTI.GOV)
Croff, A.G.; Richmond, A.A.; Williams, J.P.
1995-03-01
The Staff of the Waste Management Document Library (WMDL), in cooperation with Allen Croff have been requested to provide information support for a historical search concerning waste categorization/classification. This bibliography has been compiled under the sponsorship of Oak Ridge National Laboratory`s Chemical Technology Division to help in Allen`s ongoing committee work with the NRC/NRCP. After examining the search, Allen Croff saw the value of the search being published. Permission was sought from the database providers to allow limited publication (i.e. 20--50 copies) of the search for internal distribution at the Oak Ridge National Laboratory and for Allen Croff`s associated committee.more » Citations from the database providers who did not grant legal permission for their material to be published have been omitted from the literature review. Some of the longer citations have been included in an abbreviated form in the search to allow the format of the published document to be shortened from approximately 1,400 pages. The bibliography contains 372 citations.« less
Nano Mapper: an Internet knowledge mapping system for nanotechnology development
NASA Astrophysics Data System (ADS)
Li, Xin; Hu, Daning; Dang, Yan; Chen, Hsinchun; Roco, Mihail C.; Larson, Catherine A.; Chan, Joyce
2009-04-01
Nanotechnology research has experienced rapid growth in recent years. Advances in information technology enable efficient investigation of publications, their contents, and relationships for large sets of nanotechnology-related documents in order to assess the status of the field. This paper presents the development of a new knowledge mapping system, called Nano Mapper (http://nanomapper.eller.arizona.edu), which integrates the analysis of nanotechnology patents and research grants into a Web-based platform. The Nano Mapper system currently contains nanotechnology-related patents for 1976-2006 from the United States Patent and Trademark Office (USPTO), European Patent Office (EPO), and Japan Patent Office (JPO), as well as grant documents from the U.S. National Science Foundation (NSF) for the same time period. The system provides complex search functionalities, and makes available a set of analysis and visualization tools (statistics, trend graphs, citation networks, and content maps) that can be applied to different levels of analytical units (countries, institutions, technical fields) and for different time intervals. The paper shows important nanotechnology patenting activities at USPTO for 2005-2006 identified through the Nano Mapper system.
Nano Mapper: an Internet knowledge mapping system for nanotechnology development
Hu, Daning; Dang, Yan; Chen, Hsinchun; Roco, Mihail C.; Larson, Catherine A.; Chan, Joyce
2008-01-01
Nanotechnology research has experienced rapid growth in recent years. Advances in information technology enable efficient investigation of publications, their contents, and relationships for large sets of nanotechnology-related documents in order to assess the status of the field. This paper presents the development of a new knowledge mapping system, called Nano Mapper (http://nanomapper.eller.arizona.edu), which integrates the analysis of nanotechnology patents and research grants into a Web-based platform. The Nano Mapper system currently contains nanotechnology-related patents for 1976–2006 from the United States Patent and Trademark Office (USPTO), European Patent Office (EPO), and Japan Patent Office (JPO), as well as grant documents from the U.S. National Science Foundation (NSF) for the same time period. The system provides complex search functionalities, and makes available a set of analysis and visualization tools (statistics, trend graphs, citation networks, and content maps) that can be applied to different levels of analytical units (countries, institutions, technical fields) and for different time intervals. The paper shows important nanotechnology patenting activities at USPTO for 2005–2006 identified through the Nano Mapper system. PMID:21170121
Astronomical Software Directory Service
NASA Astrophysics Data System (ADS)
Hanisch, Robert J.; Payne, Harry; Hayes, Jeffrey
1997-01-01
With the support of NASA's Astrophysics Data Program (NRA 92-OSSA-15), we have developed the Astronomical Software Directory Service (ASDS): a distributed, searchable, WWW-based database of software packages and their related documentation. ASDS provides integrated access to 56 astronomical software packages, with more than 16,000 URLs indexed for full-text searching. Users are performing about 400 searches per month. A new aspect of our service is the inclusion of telescope and instrumentation manuals, which prompted us to change the name to the Astronomical Software and Documentation Service. ASDS was originally conceived to serve two purposes: to provide a useful Internet service in an area of expertise of the investigators (astronomical software), and as a research project to investigate various architectures for searching through a set of documents distributed across the Internet. Two of the co-investigators were then installing and maintaining astronomical software as their primary job responsibility. We felt that a service which incorporated our experience in this area would be more useful than a straightforward listing of software packages. The original concept was for a service based on the client/server model, which would function as a directory/referral service rather than as an archive. For performing the searches, we began our investigation with a decision to evaluate the Isite software from the Center for Networked Information Discovery and Retrieval (CNIDR). This software was intended as a replacement for Wide-Area Information Service (WAIS), a client/server technology for performing full-text searches through a set of documents. Isite had some additional features that we considered attractive, and we enjoyed the cooperation of the Isite developers, who were happy to have ASDS as a demonstration project. We ended up staying with the software throughout the project, making modifications to take advantage of new features as they came along, as well as influencing the software development. The Web interface to the search engine is provided by a gateway program written in C++ by a consultant to the project (A. Warnock).
A Full-Text-Based Search Engine for Finding Highly Matched Documents Across Multiple Categories
NASA Technical Reports Server (NTRS)
Nguyen, Hung D.; Steele, Gynelle C.
2016-01-01
This report demonstrates the full-text-based search engine that works on any Web-based mobile application. The engine has the capability to search databases across multiple categories based on a user's queries and identify the most relevant or similar. The search results presented here were found using an Android (Google Co.) mobile device; however, it is also compatible with other mobile phones.
EMERSE: The Electronic Medical Record Search Engine
Hanauer, David A.
2006-01-01
EMERSE (The Electronic Medical Record Search Engine) is an intuitive, powerful search engine for free-text documents in the electronic medical record. It offers multiple options for creating complex search queries yet has an interface that is easy enough to be used by those with minimal computer experience. EMERSE is ideal for retrospective chart reviews and data abstraction and may have potential for clinical care as well. PMID:17238560
Glanville, Julie; Eyers, John; Jones, Andrew M; Shemilt, Ian; Wang, Grace; Johansen, Marit; Fiander, Michelle; Rothstein, Hannah
2017-09-01
This article reviews the available evidence and guidance on methods to identify reports of quasi-experimental (QE) studies to inform systematic reviews of health care, public health, international development, education, crime and justice, and social welfare. Research, guidance, and examples of search strategies were identified by searching a range of databases, key guidance documents, selected reviews, conference proceedings, and personal communication. Current practice and research evidence were summarized. Four thousand nine hundred twenty-four records were retrieved by database searches, and additional documents were obtained by other searches. QE studies are challenging to identify efficiently because they have no standardized nomenclature and may be indexed in various ways. Reliable search filters are not available. There is a lack of specific resources devoted to collecting QE studies and little evidence on where best to search. Searches to identify QE studies should search a range of resources and, until indexing improves, use strategies that focus on the topic rather than the study design. Better definitions, better indexing in databases, prospective registers, and reporting guidance are required to improve the retrieval of QE studies and promote systematic reviews of what works based on the evidence from such studies. Copyright © 2017 Elsevier Inc. All rights reserved.
A simple and effective figure caption detection system for old-style documents
NASA Astrophysics Data System (ADS)
Liu, Zongyi; Zhou, Hanning
2011-01-01
Identifying figure captions has wide applications in producing high quality e-books such as kindle books or ipad books. In this paper, we present a rule-based system to detect horizontal figure captions in old-style documents. Our algorithm consists of three steps: (i) segment images into regions of different types such as text and figures, (ii) search the best caption region candidate based on heuristic rules such as region alignments and distances, and (iii) expand caption regions identified in step (ii) with its neighboring text-regions in order to correct oversegmentation errors. We test our algorithm using 81 images collected from old-style books, with each image containing at least one figure area. We show that the approach is able to correctly detect figure captions from images with different layouts, and we also measure its performances in terms of both precision rate and recall rate.
Bedside patient data viewer using RFID and e-Ink technology.
Nikodijevic, Aleksandar; Pichler, Patrick; Forjan, Mathias; Sauermann, Stefan
2014-01-01
In the daily routine of hospitals, which work with paper based medical records, the staff has to find the appropriate patient file if it needs information about the patient. With the introduction of ELGA the Austrian hospitals have to use specific standards for their clinical documentation. These structured documents can be used to feed an e-Ink reader with information about every patient in a hospital. Combined with RFID and security measures, the clinical staff is supported during the patient file searching process. The developed experimental setup of the Bedside Patient Data Viewer demonstrates a prototype of such a system. An Amazon Kindle Paperwhite is used to display processed data, supplied by a Raspberry Pi with an attached RFID module for identification purposes. Results show that such a system can be implemented, however a lot of organizational and technical issues remain to be solved.
Federal Register 2010, 2011, 2012, 2013, 2014
2011-04-22
... NUCLEAR REGULATORY COMMISSION [Docket No. 70-3098; NRC-2011-0081] Shaw AREVA MOX Services, Mixed... following methods: Federal Rulemaking Web site: Go to http://www.regulations.gov and search for documents... publicly available documents related to this notice using the following methods: NRC's Public Document Room...
Federal Register 2010, 2011, 2012, 2013, 2014
2010-04-26
... participant (or its counsel or representative) to digitally sign documents and access the E-Submittal server... site: Go to http://www.regulations.gov and search for documents filed under Docket ID NRC-2010-0160... documents related to this notice see Section V, Further Information. SUPPLEMENTARY INFORMATION: I...
Lowe, Dianne; Ebi, Kristie L; Forsberg, Bertil
2011-12-01
With climate change, there has been an increase in the frequency, intensity and duration of heatwave events. In response to the devastating mortality and morbidity of recent heatwave events, many countries have introduced heatwave early warning systems (HEWS). HEWS are designed to reduce the avoidable human health consequences of heatwaves through timely notification of prevention measures to vulnerable populations. To identify the key characteristics of HEWS in European countries to help inform modification of current, and development of, new systems and plans. We searched the internet to identify HEWS policy or government documents for 33 European countries and requested information from relevant organizations. We translated the HEWS documents and extracted details on the trigger indicators, thresholds for action, notification strategies, message intermediaries, communication and dissemination strategies, prevention strategies recommended and specified target audiences. Twelve European countries have HEWS. Although there are many similarities among the HEWS, there also are differences in key characteristics that could inform improvements in heatwave early warning plans.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Not Available
1979-01-01
This supplemental title list of publicly available documents pertaining to Three Mile Island Unit 2, Docket 50-320, includes those preincident and postincident documents filed in the NRC Public Document Room between July 1, 1979 and October 31, 1979. It is divided into preincident and postincident listings, with both listings sub-divided into the categories used for filing and searching in the NRC Public Document Room.
22 CFR 212.35 - Schedule of fees and method of payment for services rendered.
Code of Federal Regulations, 2014 CFR
2014-04-01
...) Definitions. (1) Direct costs means those expenditures which the Agency actually incurs in searching for and... request. (2) Search includes all time spent looking for material that is responsive to a request, including page-by-page or line-by-line identification of material within documents. Line-by-line search will...
Searching for Significance in Unstructured Data: Text Mining with Leximancer
ERIC Educational Resources Information Center
Thomas, David A.
2014-01-01
Scholars in many knowledge domains rely on sophisticated information technologies to search for and retrieve records and publications pertinent to their research interests. But what is a scholar to do when a search identifies hundreds of documents, any of which might be vital or irrelevant to his or her work? The problem is further complicated by…
Pharmacy Research Online. A Guide for Faculty.
ERIC Educational Resources Information Center
Parkin, Derral; And Others
This document is a self-paced training packet developed for a pilot project at the University of Houston-University Park to teach pharmacy faculty members to do their own online searching. The training begins with general topics such as the kinds of searches that can be done effectively online, the selection of appropriate databases to search, and…
Art Research Online. A Guide for Faculty.
ERIC Educational Resources Information Center
Parkin, Derral; And Others
This document is a self-paced training packet developed for a pilot project at the University of Houston-University Park to teach art faculty members to do their own online searching. The training begins with general topics such as the kinds of searches that can be done most effectively online, the selection of appropriate databases to search, and…
ERIC Educational Resources Information Center
Indiana Career and Postsecondary Advancement Center, Bloomington.
This packet documents suggestions for conducting a job search, writing a resume, and interviewing. Suggestions for a job search include the following: knowing one's skills, knowing what one is looking for, and knowing where to look for jobs, including newspaper advertisements, employment centers, employment firms, networking, and blind attempts.…
Cameron, Delroy; Sheth, Amit P; Jaykumar, Nishita; Thirunarayan, Krishnaprasad; Anand, Gaurish; Smith, Gary A
2014-12-01
While contemporary semantic search systems offer to improve classical keyword-based search, they are not always adequate for complex domain specific information needs. The domain of prescription drug abuse, for example, requires knowledge of both ontological concepts and "intelligible constructs" not typically modeled in ontologies. These intelligible constructs convey essential information that include notions of intensity, frequency, interval, dosage and sentiments, which could be important to the holistic needs of the information seeker. In this paper, we present a hybrid approach to domain specific information retrieval that integrates ontology-driven query interpretation with synonym-based query expansion and domain specific rules, to facilitate search in social media on prescription drug abuse. Our framework is based on a context-free grammar (CFG) that defines the query language of constructs interpretable by the search system. The grammar provides two levels of semantic interpretation: 1) a top-level CFG that facilitates retrieval of diverse textual patterns, which belong to broad templates and 2) a low-level CFG that enables interpretation of specific expressions belonging to such textual patterns. These low-level expressions occur as concepts from four different categories of data: 1) ontological concepts, 2) concepts in lexicons (such as emotions and sentiments), 3) concepts in lexicons with only partial ontology representation, called lexico-ontology concepts (such as side effects and routes of administration (ROA)), and 4) domain specific expressions (such as date, time, interval, frequency and dosage) derived solely through rules. Our approach is embodied in a novel Semantic Web platform called PREDOSE, which provides search support for complex domain specific information needs in prescription drug abuse epidemiology. When applied to a corpus of over 1 million drug abuse-related web forum posts, our search framework proved effective in retrieving relevant documents when compared with three existing search systems.
Cameron, Delroy; Sheth, Amit P.; Jaykumar, Nishita; Thirunarayan, Krishnaprasad; Anand, Gaurish; Smith, Gary A.
2015-01-01
While contemporary semantic search systems offer to improve classical keyword-based search, they are not always adequate for complex domain specific information needs. The domain of prescription drug abuse, for example, requires knowledge of both ontological concepts and “intelligible constructs” not typically modeled in ontologies. These intelligible constructs convey essential information that include notions of intensity, frequency, interval, dosage and sentiments, which could be important to the holistic needs of the information seeker. In this paper, we present a hybrid approach to domain specific information retrieval that integrates ontology-driven query interpretation with synonym-based query expansion and domain specific rules, to facilitate search in social media on prescription drug abuse. Our framework is based on a context-free grammar (CFG) that defines the query language of constructs interpretable by the search system. The grammar provides two levels of semantic interpretation: 1) a top-level CFG that facilitates retrieval of diverse textual patterns, which belong to broad templates and 2) a low-level CFG that enables interpretation of specific expressions belonging to such textual patterns. These low-level expressions occur as concepts from four different categories of data: 1) ontological concepts, 2) concepts in lexicons (such as emotions and sentiments), 3) concepts in lexicons with only partial ontology representation, called lexico-ontology concepts (such as side effects and routes of administration (ROA)), and 4) domain specific expressions (such as date, time, interval, frequency and dosage) derived solely through rules. Our approach is embodied in a novel Semantic Web platform called PREDOSE, which provides search support for complex domain specific information needs in prescription drug abuse epidemiology. When applied to a corpus of over 1 million drug abuse-related web forum posts, our search framework proved effective in retrieving relevant documents when compared with three existing search systems. PMID:25814917
EVA Wiki - Transforming Knowledge Management for EVA Flight Controllers and Instructors
NASA Technical Reports Server (NTRS)
Johnston, Stephanie S.; Alpert, Brian K.; Montalvo, Edwin James; Welsh, Lawrence Daren; Wray, Scott; Mavridis, Costa
2016-01-01
The EVA Wiki was recently implemented as the primary knowledge database to retain critical knowledge and skills in the EVA Operations group at NASA's Johnson Space Center by ensuring that information is recorded in a common, easy to search repository. Prior to the EVA Wiki, information required for EVA flight controllers and instructors was scattered across different sources, including multiple file share directories, SharePoint, individual computers, and paper archives. Many documents were outdated, and data was often difficult to find and distribute. In 2011, a team recognized that these knowledge management problems could be solved by creating an EVA Wiki using MediaWiki, a free and open-source software developed by the Wikimedia Foundation. The EVA Wiki developed into an EVA-specific Wikipedia on an internal NASA server. While the technical implementation of the wiki had many challenges, one of the biggest hurdles came from a cultural shift. Like many enterprise organizations, the EVA Operations group was accustomed to hierarchical data structures and individually-owned documents. Instead of sorting files into various folders, the wiki searches content. Rather than having a single document owner, the wiki harmonized the efforts of many contributors and established an automated revision controlled system. As the group adapted to the wiki, the usefulness of this single portal for information became apparent. It transformed into a useful data mining tool for EVA flight controllers and instructors, as well as hundreds of others that support the EVA. Program managers, engineers, astronauts, flight directors, and flight controllers in differing disciplines now have an easier-to-use, searchable system to find EVA data. This paper presents the benefits the EVA Wiki has brought to NASA's EVA community, as well as the cultural challenges it had to overcome.
EVA Wiki - Transforming Knowledge Management for EVA Flight Controllers and Instructors
NASA Technical Reports Server (NTRS)
Johnston, Stephanie S.; Alpert, Brian K.; Montalvo, Edwin James; Welsh, Lawrence Daren; Wray, Scott; Mavridis, Costa
2016-01-01
The EVA Wiki was recently implemented as the primary knowledge database to retain critical knowledge and skills in the EVA Operations group at NASA's Johnson Space Center by ensuring that information is recorded in a common, easy to search repository. Prior to the EVA Wiki, information required for EVA flight controllers and instructors was scattered across different sources, including multiple file share directories, SharePoint, individual computers, and paper archives. Many documents were outdated, and data was often difficult to find and distribute. In 2011, a team recognized that these knowledge management problems could be solved by creating an EVA Wiki using MediaWiki, a free and open-source software developed by the Wikimedia Foundation. The EVA Wiki developed into an EVA-specific Wikipedia on an internal NASA server. While the technical implementation of the wiki had many challenges, one of the biggest hurdles came from a cultural shift. Like many enterprise organizations, the EVA Operations group was accustomed to hierarchical data structures and individually-owned documents. Instead of sorting files into various folders, the wiki searches content. Rather than having a single document owner, the wiki harmonized the efforts of many contributors and established an automated revision controlled system. As the group adapted to the wiki, the usefulness of this single portal for information became apparent. It transformed into a useful data mining tool for EVA flight controllers and instructors, as well as hundreds of others that support EVA. Program managers, engineers, astronauts, flight directors, and flight controllers in differing disciplines now have an easier-to-use, searchable system to find EVA data. This paper presents the benefits the EVA Wiki has brought to NASA's EVA community, as well as the cultural challenges it had to overcome.
From Dust to Planets: The Tale Told by Moderately Volatile Element Depletion (MOVED)
NASA Technical Reports Server (NTRS)
Yin, Qing-Zhu
2004-01-01
The pronounced depletion of moderately volatile elements (MOVE, that condense or evaporate at temperatures in the range 1350-650K) relative to the average solar composition is a characteristic feature in most primitive chondrites and bulk terrestrial planets. It differs from the composition of the Sun and from the materials further away from the Sun (CI chondrites). None of the remaining planets or even meteorites shows an enrichment of volatile elements that would balance the depletion in the inner Solar System. Whether this depletion occurred in solar nebular stage or in planetary formation stage has been the subject of long lasting debate. The search for mysterite initiated in 1973 continues today in search of lost planets. Here I show that the MOVED patterns demonstrate a clear connection between the rocky materials of the inner solar system and the interstellar dust. The inheritance of interstellar materials by the solar system is not only documented by the presence of presolar grains, various isotopic anomalies, but also expressed in the chemical element distribution in the inner solar system.
Font adaptive word indexing of modern printed documents.
Marinai, Simone; Marino, Emanuele; Soda, Giovanni
2006-08-01
We propose an approach for the word-level indexing of modern printed documents which are difficult to recognize using current OCR engines. By means of word-level indexing, it is possible to retrieve the position of words in a document, enabling queries involving proximity of terms. Web search engines implement this kind of indexing, allowing users to retrieve Web pages on the basis of their textual content. Nowadays, digital libraries hold collections of digitized documents that can be retrieved either by browsing the document images or relying on appropriate metadata assembled by domain experts. Word indexing tools would therefore increase the access to these collections. The proposed system is designed to index homogeneous document collections by automatically adapting to different languages and font styles without relying on OCR engines for character recognition. The approach is based on three main ideas: the use of Self Organizing Maps (SOM) to perform unsupervised character clustering, the definition of one suitable vector-based word representation whose size depends on the word aspect-ratio, and the run-time alignment of the query word with indexed words to deal with broken and touching characters. The most appropriate applications are for processing modern printed documents (17th to 19th centuries) where current OCR engines are less accurate. Our experimental analysis addresses six data sets containing documents ranging from books of the 17th century to contemporary journals.
Document image retrieval through word shape coding.
Lu, Shijian; Li, Linlin; Tan, Chew Lim
2008-11-01
This paper presents a document retrieval technique that is capable of searching document images without OCR (optical character recognition). The proposed technique retrieves document images by a new word shape coding scheme, which captures the document content through annotating each word image by a word shape code. In particular, we annotate word images by using a set of topological shape features including character ascenders/descenders, character holes, and character water reservoirs. With the annotated word shape codes, document images can be retrieved by either query keywords or a query document image. Experimental results show that the proposed document image retrieval technique is fast, efficient, and tolerant to various types of document degradation.
Search and retrieval of office files using dBASE 3
NASA Technical Reports Server (NTRS)
Breazeale, W. L.; Talley, C. R.
1986-01-01
Described is a method of automating the office files retrieval process using a commercially available software package (dBASE III). The resulting product is a menu-driven computer program which requires no computer skills to operate. One part of the document is written for the potential user who has minimal computer experience and uses sample menu screens to explain the program; while a second part is oriented towards the computer literate individual and includes rather detailed descriptions of the methodology and search routines. Although much of the programming techniques are explained, this document is not intended to be a tutorial on dBASE III. It is hoped that the document will serve as a stimulus for other applications of dBASE III.
Turning Search into Knowledge Management.
ERIC Educational Resources Information Center
Kaufman, David
2002-01-01
Discussion of knowledge management for electronic data focuses on creating a high quality similarity ranking algorithm. Topics include similarity ranking and unstructured data management; searching, categorization, and summarization of documents; query evaluation; considering sentences in addition to keywords; and vector models. (LRW)
CwicStart - a proof-of-concept client for the CEOSWGISS Integrated Catalog (CWIC)
NASA Astrophysics Data System (ADS)
Newman, D. J.; Mitchell, A. E.
2012-12-01
Keywords - Earth Science, data discovery, agile development, ruby on rails, catalog, OGC Audience - Earth Science application developers What is CwicStart CwicStart is a prototypical earth science data discovery web application designed, developed and hosted by NASA's Earth Observing System (EOS) Clearinghouse (ECHO). CwicStart searches the CEOS WGISS Integrated Catalog (CWIC) to provide users with dataset and granule level metadata from sources as diverse as NASA, NOAA, INPE and AOE. CwicStart demonstrates the ease of which it is possible to stand up a functioning client against the CWIC. From start to finish, CwicStart was designed, developed and deployed in one month. Built from OGC getCapabilities document of CWIC The CwicStart application takes the OGC getCapabilities (http://www.opengeospatial.org/standards/is) document describing CWIC, as it's starting point for providing a user interface suitable for interrogating CWIC. Consequently, it can allow the user to constrain their search by the following criteria, - Generic search terms - Spatial bounding box - Start date/time and end date/time - ISO-queryable key-value pairs User Interface inspired by Reverb ECHO's state-of-the-art earth science discovery tool, Reverb (http://reverb.echo.nasa.gov) was used as a guideline for the user interface components of CwicStart. It incorporates OpenLayers to provide point-and-click spatial constraint specification and calendar input for temporal constraints. Discovery involves two phases: dataset discovery and granule discovery with full pagination support for large results sets. CwicStart supports 'graceful degradation' of support for multiple browsers and accessibility requirements. Implemented in Ruby on Rails for Agile Development CwicStart is implemented in Ruby on Rails, a dynamic, rapid development language and environment that facilitates Agile development and is resilient to changing requirements. Using an Agile development methodology ECHO was able to stand up the first iteration of CwicStart in an iteration lasting only one week. Three subsequent week-long iterations delivered the current functionality. CwicStart can be found at the following location, https://testbed.echo.nasa.gov/cwic-start/ About CWIC The WGISS team provides an application, the CEOS WGISS Integrated Catalog (CWIC) with the following capabilities. - Provide an access point for major CEOS agency catalog systems. - Interface to user interface clients by using the GEO standards. - Send directory/collection searches to the International Directory Network. - Distribute inventory/product searches to the CEOS agency inventory systems using the agency systems native protocol. - Offered as the CEOS community catalog as part of the GEO common infrastructure. CWIC Partners - Committee on Earth Observing Satellites (CEOS) - International Directory Network (IDN) - U.S. National Aeronautics and Space (NASA) - Earth Observing System (EOS) Clearinghouse (ECHO) - U.S. National Oceanographic and Atmospheric Administration (NOAA) - Comprehensive Large Array Data Stewardship System (CLASS) - U.S. Geological Survey (USGS) - Landsat Catalog System - U.S. Geological Survey (USGS) - LSI Portal - National Institute for Space Research (INPE), Brazil - Academy of Opto-Electronics (AOE), Chinese Academy of Sciences (CAS)
Discovery in a World of Mashups
NASA Astrophysics Data System (ADS)
King, T. A.; Ritschel, B.; Hourcle, J. A.; Moon, I. S.
2014-12-01
When the first digital information was stored electronically, discovery of what existed was through file names and the organization of the file system. With the advent of networks, digital information was shared on a wider scale, but discovery remained based on file and folder names. With a growing number of information sources, named based discovery quickly became ineffective. The keyword based search engine was one of the first types of a mashup in the world of Web 1.0. Embedded links from one document to another with prescribed relationships between files and the world of Web 2.0 was formed. Search engines like Google used the links to improve search results and a worldwide mashup was formed. While a vast improvement, the need for semantic (meaning rich) discovery was clear, especially for the discovery of scientific data. In response, every science discipline defined schemas to describe their type of data. Some core schemas where shared, but most schemas are custom tailored even though they share many common concepts. As with the networking of information sources, science increasingly relies on data from multiple disciplines. So there is a need to bring together multiple sources of semantically rich information. We explore how harvesting, conceptual mapping, facet based search engines, search term promotion, and style sheets can be combined to create the next generation of mashups in the emerging world of Web 3.0. We use NASA's Planetary Data System and NASA's Heliophysics Data Environment to illustrate how to create a multi-discipline mash-up.
Automatic indexing of scanned documents: a layout-based approach
NASA Astrophysics Data System (ADS)
Esser, Daniel; Schuster, Daniel; Muthmann, Klemens; Berger, Michael; Schill, Alexander
2012-01-01
Archiving official written documents such as invoices, reminders and account statements in business and private area gets more and more important. Creating appropriate index entries for document archives like sender's name, creation date or document number is a tedious manual work. We present a novel approach to handle automatic indexing of documents based on generic positional extraction of index terms. For this purpose we apply the knowledge of document templates stored in a common full text search index to find index positions that were successfully extracted in the past.
Soliciting and Responding to Patients' Questions about Diabetes Through Online Sources.
Crangle, Colleen E; Bradley, Colin; Carlin, Paul F; Esterhay, Robert J; Harper, Roy; Kearney, Patricia M; Lorig, Kate; McCarthy, Vera J C; McTear, Michael F; Tuttle, Mark S; Wallace, Jonathan G; Savage, Eileen
2017-03-01
When patients cannot get answers from health professionals or retain the information given, increasingly they search online for answers, with limited success. Researchers from the United States, Ireland, and the United Kingdom explored this problem for patients with type 2 diabetes mellitus (T2DM). In 2014, patients attending an outpatient clinic (UK) were asked to submit questions about diabetes. Ten questions judged representative of different types of patient concerns were selected by the researchers and submitted to search engines within trusted and vetted websites in the United States, Ireland, and the United Kingdom. Two researchers independently assessed if answers could be found in the three top-ranked documents returned at each website. The 2014 search was repeated in June, 2016, examining the two top-ranked documents returned. One hundred and sixty-four questions were collected from 120 patients during 12 outpatient clinics. Most patients had T2DM (95%). Most questions were about diabetes (N = 155) with the remainder related to clinic operation (N = 9). Of the questions on diabetes, 152 were about T2DM. The 2014 assessment found no adequate answers to the questions in 90 documents (10 questions, 3 websites, 3 top documents). In the 2016 assessment, 1 document out of 60 (10 questions, 3 websites, 2 top documents) provided an adequate answer relating to 1 of the 10 questions. Available online sources of information do not provide answers to questions from patients with diabetes. Our results highlight the urgent need to develop novel ways of providing answers to patient questions about T2DM.
New Tools to Document and Manage Data/Metadata: Example NGEE Arctic and ARM
NASA Astrophysics Data System (ADS)
Crow, M. C.; Devarakonda, R.; Killeffer, T.; Hook, L.; Boden, T.; Wullschleger, S.
2017-12-01
Tools used for documenting, archiving, cataloging, and searching data are critical pieces of informatics. This poster describes tools being used in several projects at Oak Ridge National Laboratory (ORNL), with a focus on the U.S. Department of Energy's Next Generation Ecosystem Experiment in the Arctic (NGEE Arctic) and Atmospheric Radiation Measurements (ARM) project, and their usage at different stages of the data lifecycle. The Online Metadata Editor (OME) is used for the documentation and archival stages while a Data Search tool supports indexing, cataloging, and searching. The NGEE Arctic OME Tool [1] provides a method by which researchers can upload their data and provide original metadata with each upload while adhering to standard metadata formats. The tool is built upon a Java SPRING framework to parse user input into, and from, XML output. Many aspects of the tool require use of a relational database including encrypted user-login, auto-fill functionality for predefined sites and plots, and file reference storage and sorting. The Data Search Tool conveniently displays each data record in a thumbnail containing the title, source, and date range, and features a quick view of the metadata associated with that record, as well as a direct link to the data. The search box incorporates autocomplete capabilities for search terms and sorted keyword filters are available on the side of the page, including a map for geo-searching. These tools are supported by the Mercury [2] consortium (funded by DOE, NASA, USGS, and ARM) and developed and managed at Oak Ridge National Laboratory. Mercury is a set of tools for collecting, searching, and retrieving metadata and data. Mercury collects metadata from contributing project servers, then indexes the metadata to make it searchable using Apache Solr, and provides access to retrieve it from the web page. Metadata standards that Mercury supports include: XML, Z39.50, FGDC, Dublin-Core, Darwin-Core, EML, and ISO-19115.
MINER - A Mobile Imager of Neutrons for Emergency Responders
DOE Office of Scientific and Technical Information (OSTI.GOV)
Goldsmith, John E. M.; Brennan, James S.; Gerling, Mark D
2014-10-01
We have developed a mobile fast neutron imaging platform to enhance the capabilities of emergency responders in the localization and characterization of special nuclear material. This mobile imager of neutrons for emergency responders (MINER) is based on the Neutron Scatter Camera, a large segmented imaging system that was optimized for large-area search applications. Due to the reduced size and power requirements of a man-portable system, MINER has been engineered to fit a much smaller form factor, and to be operated from either a battery or AC power. We chose a design that enabled omnidirectional (4π) imaging, with only a ~twofoldmore » decrease in sensitivity compared to the much larger neutron scatter cameras. The system was designed to optimize its performance for neutron imaging and spectroscopy, but it does also function as a Compton camera for gamma imaging. This document outlines the project activities, broadly characterized as system development, laboratory measurements, and deployments, and presents sample results in these areas. Additional information can be found in the documents that reside in WebPMIS.« less
The Marriage of Fax and Online.
ERIC Educational Resources Information Center
Basch, Reva
1995-01-01
Discusses the use of fax transmissions. Highlights include searching by fax, including online service, print and electronic publishing, and database producers; customer service, including documentation updates, new product announcements, and marketing materials; document delivery; problems; and fax messaging. (four references) (LRW)
Tobacco document research reporting
Carter, S
2005-01-01
Design: Interpretive analysis of published research. Sample: 173 papers indexed in Medline between 1995 and 2004 that cited tobacco industry documents. Analysis: Information about year published, journal and author, and a set of codes relating to methods reporting, were managed in N*Vivo. This coding formed the basis of an interpretation of tobacco document research reporting. Results: Two types of papers were identified. The first used tobacco documents as the primary data source (A-papers). The second was dedicated to another purpose but cited a small number of documents (B-papers). In B-papers documents were used either to provide a specific example or to support an expansive contention. A-papers contained information about purpose, sources, searching, analysis, and limitations that differed by author and journal and over time. A-papers had no clear methodological context, but used words from three major traditions—interpretive research, positivist research, and history—to describe analysis. Interpretation: A descriptive mainstream form of tobacco document reporting is proposed, initially typical but decreasing, and a continuum of positioning of the researcher, from conduit to constructor. Reporting practices, particularly from experienced researchers, appeared to evolve towards researcher as constructor, with later papers showing more complex purposes, diverse sources, and detail of searching and analysis. Tobacco document research could learn from existing research traditions: a model for planning and evaluating tobacco document research is presented. PMID:16319359
Federal Register 2010, 2011, 2012, 2013, 2014
2013-08-13
... http://www.regulations.gov and search for Docket ID NRC-2013-0182. Address questions about NRC dockets... NRC Library at http://www.nrc.gov/reading-rm/adams.html . To begin the search, select ``ADAMS Public Documents'' and then select ``Begin Web- based ADAMS Search.'' For problems with ADAMS, please contact the...
Code of Federal Regulations, 2011 CFR
2011-10-01
... Criminal History Check must I satisfy to determine an individual's suitability to serve in a covered... a Sponsor § 2552.27 What two search components of the National Service Criminal History Check must I... conduct and document a National Service Criminal History Check, which consists of the following two search...
2016-07-21
Todays internet has multiple webs. The surface web is what Google and other search engines index and pull based on links. Essentially, the surface...financial records, research and development), and personal data (medical records or legal documents). These are all deep web. Standard search engines dont
ERIC Educational Resources Information Center
Burke, Arvid J.; Burke, Mary A.
After a summary of background knowledge useful in searching for information, the authors cover extensively the sources available to the researcher interested in locating educational data or conducting a search of bibliographic materials. They list reference books, dictionaries, almanacs, yearbooks, subject matter summaries; and sources for…
Logic-Based Retrieval: Technology for Content-Oriented and Analytical Querying of Patent Data
NASA Astrophysics Data System (ADS)
Klampanos, Iraklis Angelos; Wu, Hengzhi; Roelleke, Thomas; Azzam, Hany
Patent searching is a complex retrieval task. An initial document search is only the starting point of a chain of searches and decisions that need to be made by patent searchers. Keyword-based retrieval is adequate for document searching, but it is not suitable for modelling comprehensive retrieval strategies. DB-like and logical approaches are the state-of-the-art techniques to model strategies, reasoning and decision making. In this paper we present the application of logical retrieval to patent searching. The two grand challenges are expressiveness and scalability, where high degree of expressiveness usually means a loss in scalability. In this paper we report how to maintain scalability while offering the expressiveness of logical retrieval required for solving patent search tasks. We present logical retrieval background, and how to model data-source selection and results' fusion. Moreover, we demonstrate the modelling of a retrieval strategy, a technique by which patent professionals are able to express, store and exchange their strategies and rationales when searching patents or when making decisions. An overview of the architecture and technical details complement the paper, while the evaluation reports preliminary results on how query processing times can be guaranteed, and how quality is affected by trading off responsiveness.
Miniature infrared data acquisition and telemetry system
NASA Technical Reports Server (NTRS)
Stokes, J. H.; Ward, S. M.
1985-01-01
The Miniature Infrared Data Acquisition and Telemetry (MIRDAT) Phase 1 study was performed to determine the technical and commercial feasibility of producing a miniaturized electro-optical telemetry system. This system acquires and transmits experimental data from aircraft scale models for realtime monitoring in wind tunnels. During the Phase 1 study, miniature prototype MIRDAT telemetry devices were constructed, successfully tested in the laboratory and delivered to the user for wind tunnel testing. A search was conducted for commercially available components and advanced hybrid techniques to further miniaturize the system during Phase 2 development. A design specification was generated from laboratory testing, user requirements and discussions with component manufacturers. Finally, a preliminary design of the proposed MIRDAT system was documented for Phase 2 development.