Evolution of Query Optimization Methods
NASA Astrophysics Data System (ADS)
Hameurlain, Abdelkader; Morvan, Franck
Query optimization is the most critical phase in query processing. In this paper, we try to describe synthetically the evolution of query optimization methods from uniprocessor relational database systems to data Grid systems through parallel, distributed and data integration systems. We point out a set of parameters to characterize and compare query optimization methods, mainly: (i) size of the search space, (ii) type of method (static or dynamic), (iii) modification types of execution plans (re-optimization or re-scheduling), (iv) level of modification (intra-operator and/or inter-operator), (v) type of event (estimation errors, delay, user preferences), and (vi) nature of decision-making (centralized or decentralized control).
Li, Jian; Yang, Yu-Guang; Chen, Xiu-Bo; Zhou, Yi-Hua; Shi, Wei-Min
2016-08-19
A novel quantum private database query protocol is proposed, based on passive round-robin differential phase-shift quantum key distribution. Compared with previous quantum private database query protocols, the present protocol has the following unique merits: (i) the user Alice can obtain one and only one key bit so that both the efficiency and security of the present protocol can be ensured, and (ii) it does not require to change the length difference of the two arms in a Mach-Zehnder interferometer and just chooses two pulses passively to interfere with so that it is much simpler and more practical. The present protocol is also proved to be secure in terms of the user security and database security.
Li, Jian; Yang, Yu-Guang; Chen, Xiu-Bo; Zhou, Yi-Hua; Shi, Wei-Min
2016-01-01
A novel quantum private database query protocol is proposed, based on passive round-robin differential phase-shift quantum key distribution. Compared with previous quantum private database query protocols, the present protocol has the following unique merits: (i) the user Alice can obtain one and only one key bit so that both the efficiency and security of the present protocol can be ensured, and (ii) it does not require to change the length difference of the two arms in a Mach-Zehnder interferometer and just chooses two pulses passively to interfere with so that it is much simpler and more practical. The present protocol is also proved to be secure in terms of the user security and database security. PMID:27539654
Robust Requirements Tracing via Internet Search Technology: Improving an IV and V Technique. Phase 2
NASA Technical Reports Server (NTRS)
Hayes, Jane; Dekhtyar, Alex
2004-01-01
There are three major objectives to this phase of the work. (1) Improvement of Information Retrieval (IR) methods for Independent Verification and Validation (IV&V) requirements tracing. Information Retrieval methods are typically developed for very large (order of millions - tens of millions and more documents) document collections and therefore, most successfully used methods somewhat sacrifice precision and recall in order to achieve efficiency. At the same time typical IR systems treat all user queries as independent of each other and assume that relevance of documents to queries is subjective for each user. The IV&V requirements tracing problem has a much smaller data set to operate on, even for large software development projects; the set of queries is predetermined by the high-level specification document and individual requirements considered as query input to IR methods are not necessarily independent from each other. Namely, knowledge about the links for one requirement may be helpful in determining the links of another requirement. Finally, while the final decision on the exact form of the traceability matrix still belongs to the IV&V analyst, his/her decisions are much less arbitrary than those of an Internet search engine user. All this suggests that the information available to us in the framework of the IV&V tracing problem can be successfully leveraged to enhance standard IR techniques, which in turn would lead to increased recall and precision. We developed several new methods during Phase II; (2) IV&V requirements tracing IR toolkit. Based on the methods developed in Phase I and their improvements developed in Phase II, we built a toolkit of IR methods for IV&V requirements tracing. The toolkit has been integrated, at the data level, with SAIC's SuperTracePlus (STP) tool; (3) Toolkit testing. We tested the methods included in the IV&V requirements tracing IR toolkit on a number of projects.
Accessing the public MIMIC-II intensive care relational database for clinical research.
Scott, Daniel J; Lee, Joon; Silva, Ikaro; Park, Shinhyuk; Moody, George B; Celi, Leo A; Mark, Roger G
2013-01-10
The Multiparameter Intelligent Monitoring in Intensive Care II (MIMIC-II) database is a free, public resource for intensive care research. The database was officially released in 2006, and has attracted a growing number of researchers in academia and industry. We present the two major software tools that facilitate accessing the relational database: the web-based QueryBuilder and a downloadable virtual machine (VM) image. QueryBuilder and the MIMIC-II VM have been developed successfully and are freely available to MIMIC-II users. Simple example SQL queries and the resulting data are presented. Clinical studies pertaining to acute kidney injury and prediction of fluid requirements in the intensive care unit are shown as typical examples of research performed with MIMIC-II. In addition, MIMIC-II has also provided data for annual PhysioNet/Computing in Cardiology Challenges, including the 2012 Challenge "Predicting mortality of ICU Patients". QueryBuilder is a web-based tool that provides easy access to MIMIC-II. For more computationally intensive queries, one can locally install a complete copy of MIMIC-II in a VM. Both publicly available tools provide the MIMIC-II research community with convenient querying interfaces and complement the value of the MIMIC-II relational database.
Accessing the public MIMIC-II intensive care relational database for clinical research
2013-01-01
Background The Multiparameter Intelligent Monitoring in Intensive Care II (MIMIC-II) database is a free, public resource for intensive care research. The database was officially released in 2006, and has attracted a growing number of researchers in academia and industry. We present the two major software tools that facilitate accessing the relational database: the web-based QueryBuilder and a downloadable virtual machine (VM) image. Results QueryBuilder and the MIMIC-II VM have been developed successfully and are freely available to MIMIC-II users. Simple example SQL queries and the resulting data are presented. Clinical studies pertaining to acute kidney injury and prediction of fluid requirements in the intensive care unit are shown as typical examples of research performed with MIMIC-II. In addition, MIMIC-II has also provided data for annual PhysioNet/Computing in Cardiology Challenges, including the 2012 Challenge “Predicting mortality of ICU Patients”. Conclusions QueryBuilder is a web-based tool that provides easy access to MIMIC-II. For more computationally intensive queries, one can locally install a complete copy of MIMIC-II in a VM. Both publicly available tools provide the MIMIC-II research community with convenient querying interfaces and complement the value of the MIMIC-II relational database. PMID:23302652
A study of medical and health queries to web search engines.
Spink, Amanda; Yang, Yin; Jansen, Jim; Nykanen, Pirrko; Lorence, Daniel P; Ozmutlu, Seda; Ozmutlu, H Cenk
2004-03-01
This paper reports findings from an analysis of medical or health queries to different web search engines. We report results: (i). comparing samples of 10000 web queries taken randomly from 1.2 million query logs from the AlltheWeb.com and Excite.com commercial web search engines in 2001 for medical or health queries, (ii). comparing the 2001 findings from Excite and AlltheWeb.com users with results from a previous analysis of medical and health related queries from the Excite Web search engine for 1997 and 1999, and (iii). medical or health advice-seeking queries beginning with the word 'should'. Findings suggest: (i). a small percentage of web queries are medical or health related, (ii). the top five categories of medical or health queries were: general health, weight issues, reproductive health and puberty, pregnancy/obstetrics, and human relationships, and (iii). over time, the medical and health queries may have declined as a proportion of all web queries, as the use of specialized medical/health websites and e-commerce-related queries has increased. Findings provide insights into medical and health-related web querying and suggests some implications for the use of the general web search engines when seeking medical/health information.
Distributed query plan generation using multiobjective genetic algorithm.
Panicker, Shina; Kumar, T V Vijay
2014-01-01
A distributed query processing strategy, which is a key performance determinant in accessing distributed databases, aims to minimize the total query processing cost. One way to achieve this is by generating efficient distributed query plans that involve fewer sites for processing a query. In the case of distributed relational databases, the number of possible query plans increases exponentially with respect to the number of relations accessed by the query and the number of sites where these relations reside. Consequently, computing optimal distributed query plans becomes a complex problem. This distributed query plan generation (DQPG) problem has already been addressed using single objective genetic algorithm, where the objective is to minimize the total query processing cost comprising the local processing cost (LPC) and the site-to-site communication cost (CC). In this paper, this DQPG problem is formulated and solved as a biobjective optimization problem with the two objectives being minimize total LPC and minimize total CC. These objectives are simultaneously optimized using a multiobjective genetic algorithm NSGA-II. Experimental comparison of the proposed NSGA-II based DQPG algorithm with the single objective genetic algorithm shows that the former performs comparatively better and converges quickly towards optimal solutions for an observed crossover and mutation probability.
Distributed Query Plan Generation Using Multiobjective Genetic Algorithm
Panicker, Shina; Vijay Kumar, T. V.
2014-01-01
A distributed query processing strategy, which is a key performance determinant in accessing distributed databases, aims to minimize the total query processing cost. One way to achieve this is by generating efficient distributed query plans that involve fewer sites for processing a query. In the case of distributed relational databases, the number of possible query plans increases exponentially with respect to the number of relations accessed by the query and the number of sites where these relations reside. Consequently, computing optimal distributed query plans becomes a complex problem. This distributed query plan generation (DQPG) problem has already been addressed using single objective genetic algorithm, where the objective is to minimize the total query processing cost comprising the local processing cost (LPC) and the site-to-site communication cost (CC). In this paper, this DQPG problem is formulated and solved as a biobjective optimization problem with the two objectives being minimize total LPC and minimize total CC. These objectives are simultaneously optimized using a multiobjective genetic algorithm NSGA-II. Experimental comparison of the proposed NSGA-II based DQPG algorithm with the single objective genetic algorithm shows that the former performs comparatively better and converges quickly towards optimal solutions for an observed crossover and mutation probability. PMID:24963513
Wang, Shuai; Xing, Huijie; Zhao, Mengjing; Lu, Danyi; Li, Zhijie; Dong, Dong; Wu, Baojian
2016-01-01
Mechanistic understanding of the metabolism-transport interplay assumes great importance in pharmaceutical fields because the knowledge can help to interpret drug/xenobiotic metabolism and disposition studies as well as the drug-drug interactions in vivo. About 10 years ago, it started to recognize that cellular phase II metabolism is strongly influenced by the excretion (efflux transport) of generated metabolites, a kinetic phenomenon termed "phase II metabolism-transport interplay". This interplay is believed to have significant effects on the pharmacokinetics (bioavailability) of drugs/chemicals undergoing phase II metabolism. In this article, we review the studies investigating the phase II metabolism-transport interplay using cell models, perfused rat intestine, and intact rats. The potential confounding factors in exploring such interplay is also summarized. Moreover, the mechanism underlying the phase II metabolism-transport interplay is discussed. Various studies with engineered cells and rodents have demonstrated that there is an interaction (interplay) between phase II enzymes and efflux transporters. This type of interplay mainly refers to the dependence of phase II (conjugative) metabolism on the activities of efflux transporters. In general, inhibiting efflux transporters or decreasing their expression causes the reductions in metabolite excretion, apparent excretion clearance (CLapp) and total metabolism (fmet), as well as an increase in the intracellular level of metabolite (Ci). The deconjugation mediated by hydrolase (acting as a "bridge") is essential for the interplay to play out based on pharmacokinetic modeling/simulations, cell and animal studies. The hydrolases bridge the two processes (i.e., metabolite formation and excretion) and enable the interplay thereof (a bridging effect). Without the bridge, metabolite formation is independent on its downstream process excretion, thus impact of metabolite excretion on its formation is impossible. Deconjugation (mediated by hydrolases) plays an essential role in the conjugation-transport interplay. Copyright© Bentham Science Publishers; For any queries, please email at epub@benthamscience.org.
The Ned IIS project - forest ecosystem management
W. Potter; D. Nute; J. Wang; F. Maier; Michael Twery; H. Michael Rauscher; P. Knopp; S. Thomasma; M. Dass; H. Uchiyama
2002-01-01
For many years we have held to the notion that an Intelligent Information System (IIS) is composed of a unified knowledge base, database, and model base. The main idea behind this notion is the transparent processing of user queries. The system is responsible for "deciding" which information sources to access in order to fulfil a query regardless of whether...
An Intelligent Information System for forest management: NED/FVS integration
J. Wang; W.D. Potter; D. Nute; F. Maier; H. Michael Rauscher; M.J. Twery; S. Thomasma; P. Knopp
2002-01-01
An Intelligent Information System (IIS) is viewed as composed of a unified knowledge base, database, and model base. This allows an IIS to provide responses to user queries regardless of whether the query process involves a data retrieval, an inference, a computational method, a problem solving module, or some combination of these. NED-2 is a full-featured intelligent...
Query Transformations for Result Merging
2014-11-01
tors, term dependence, query expansion 1. INTRODUCTION Federated search deals with the problem of aggregating results from multiple search engines . The...invidual search engines are (i) typically focused on a particular domain or a particular corpus, (ii) employ diverse retrieval models, and (iii...determine which search engines are appropri- ate for addressing the information need (resource selection), and (ii) merging the results returned by
Jung, HaRim; Song, MoonBae; Youn, Hee Yong; Kim, Ung Mo
2015-09-18
A content-matched (CM) rangemonitoring query overmoving objects continually retrieves the moving objects (i) whose non-spatial attribute values are matched to given non-spatial query values; and (ii) that are currently located within a given spatial query range. In this paper, we propose a new query indexing structure, called the group-aware query region tree (GQR-tree) for efficient evaluation of CMrange monitoring queries. The primary role of the GQR-tree is to help the server leverage the computational capabilities of moving objects in order to improve the system performance in terms of the wireless communication cost and server workload. Through a series of comprehensive simulations, we verify the superiority of the GQR-tree method over the existing methods.
Nonchronological video synopsis and indexing.
Pritch, Yael; Rav-Acha, Alex; Peleg, Shmuel
2008-11-01
The amount of captured video is growing with the increased numbers of video cameras, especially the increase of millions of surveillance cameras that operate 24 hours a day. Since video browsing and retrieval is time consuming, most captured video is never watched or examined. Video synopsis is an effective tool for browsing and indexing of such a video. It provides a short video representation, while preserving the essential activities of the original video. The activity in the video is condensed into a shorter period by simultaneously showing multiple activities, even when they originally occurred at different times. The synopsis video is also an index into the original video by pointing to the original time of each activity. Video Synopsis can be applied to create a synopsis of an endless video streams, as generated by webcams and by surveillance cameras. It can address queries like "Show in one minute the synopsis of this camera broadcast during the past day''. This process includes two major phases: (i) An online conversion of the endless video stream into a database of objects and activities (rather than frames). (ii) A response phase, generating the video synopsis as a response to the user's query.
Jung, HaRim; Song, MoonBae; Youn, Hee Yong; Kim, Ung Mo
2015-01-01
A content-matched (CM) range monitoring query over moving objects continually retrieves the moving objects (i) whose non-spatial attribute values are matched to given non-spatial query values; and (ii) that are currently located within a given spatial query range. In this paper, we propose a new query indexing structure, called the group-aware query region tree (GQR-tree) for efficient evaluation of CM range monitoring queries. The primary role of the GQR-tree is to help the server leverage the computational capabilities of moving objects in order to improve the system performance in terms of the wireless communication cost and server workload. Through a series of comprehensive simulations, we verify the superiority of the GQR-tree method over the existing methods. PMID:26393613
Visually defining and querying consistent multi-granular clinical temporal abstractions.
Combi, Carlo; Oliboni, Barbara
2012-02-01
The main goal of this work is to propose a framework for the visual specification and query of consistent multi-granular clinical temporal abstractions. We focus on the issue of querying patient clinical information by visually defining and composing temporal abstractions, i.e., high level patterns derived from several time-stamped raw data. In particular, we focus on the visual specification of consistent temporal abstractions with different granularities and on the visual composition of different temporal abstractions for querying clinical databases. Temporal abstractions on clinical data provide a concise and high-level description of temporal raw data, and a suitable way to support decision making. Granularities define partitions on the time line and allow one to represent time and, thus, temporal clinical information at different levels of detail, according to the requirements coming from the represented clinical domain. The visual representation of temporal information has been considered since several years in clinical domains. Proposed visualization techniques must be easy and quick to understand, and could benefit from visual metaphors that do not lead to ambiguous interpretations. Recently, physical metaphors such as strips, springs, weights, and wires have been proposed and evaluated on clinical users for the specification of temporal clinical abstractions. Visual approaches to boolean queries have been considered in the last years and confirmed that the visual support to the specification of complex boolean queries is both an important and difficult research topic. We propose and describe a visual language for the definition of temporal abstractions based on a set of intuitive metaphors (striped wall, plastered wall, brick wall), allowing the clinician to use different granularities. A new algorithm, underlying the visual language, allows the physician to specify only consistent abstractions, i.e., abstractions not containing contradictory conditions on the component abstractions. Moreover, we propose a visual query language where different temporal abstractions can be composed to build complex queries: temporal abstractions are visually connected through the usual logical connectives AND, OR, and NOT. The proposed visual language allows one to simply define temporal abstractions by using intuitive metaphors, and to specify temporal intervals related to abstractions by using different temporal granularities. The physician can interact with the designed and implemented tool by point-and-click selections, and can visually compose queries involving several temporal abstractions. The evaluation of the proposed granularity-related metaphors consisted in two parts: (i) solving 30 interpretation exercises by choosing the correct interpretation of a given screenshot representing a possible scenario, and (ii) solving a complex exercise, by visually specifying through the interface a scenario described only in natural language. The exercises were done by 13 subjects. The percentage of correct answers to the interpretation exercises were slightly different with respect to the considered metaphors (54.4--striped wall, 73.3--plastered wall, 61--brick wall, and 61--no wall), but post hoc statistical analysis on means confirmed that differences were not statistically significant. The result of the user's satisfaction questionnaire related to the evaluation of the proposed granularity-related metaphors ratified that there are no preferences for one of them. The evaluation of the proposed logical notation consisted in two parts: (i) solving five interpretation exercises provided by a screenshot representing a possible scenario and by three different possible interpretations, of which only one was correct, and (ii) solving five exercises, by visually defining through the interface a scenario described only in natural language. Exercises had an increasing difficulty. The evaluation involved a total of 31 subjects. Results related to this evaluation phase confirmed us about the soundness of the proposed solution even in comparison with a well known proposal based on a tabular query form (the only significant difference is that our proposal requires more time for the training phase: 21 min versus 14 min). In this work we have considered the issue of visually composing and querying temporal clinical patient data. In this context we have proposed a visual framework for the specification of consistent temporal abstractions with different granularities and for the visual composition of different temporal abstractions to build (possibly) complex queries on clinical databases. A new algorithm has been proposed to check the consistency of the specified granular abstraction. From the evaluation of the proposed metaphors and interfaces and from the comparison of the visual query language with a well known visual method for boolean queries, the soundness of the overall system has been confirmed; moreover, pros and cons and possible improvements emerged from the comparison of different visual metaphors and solutions. Copyright © 2011 Elsevier B.V. All rights reserved.
Comment on "Secure quantum private information retrieval using phase-encoded queries"
NASA Astrophysics Data System (ADS)
Shi, Run-hua; Mu, Yi; Zhong, Hong; Zhang, Shun
2016-12-01
In this Comment, we reexamine the security of phase-encoded quantum private query (QPQ). We find that the current phase-encoded QPQ protocols, including their applications, are vulnerable to a probabilistic entangle-and-measure attack performed by the owner of the database. Furthermore, we discuss how to overcome this security loophole and present an improved cheat-sensitive QPQ protocol without losing the good features of the original protocol.
RCQ-GA: RDF Chain Query Optimization Using Genetic Algorithms
NASA Astrophysics Data System (ADS)
Hogenboom, Alexander; Milea, Viorel; Frasincar, Flavius; Kaymak, Uzay
The application of Semantic Web technologies in an Electronic Commerce environment implies a need for good support tools. Fast query engines are needed for efficient querying of large amounts of data, usually represented using RDF. We focus on optimizing a special class of SPARQL queries, the so-called RDF chain queries. For this purpose, we devise a genetic algorithm called RCQ-GA that determines the order in which joins need to be performed for an efficient evaluation of RDF chain queries. The approach is benchmarked against a two-phase optimization algorithm, previously proposed in literature. The more complex a query is, the more RCQ-GA outperforms the benchmark in solution quality, execution time needed, and consistency of solution quality. When the algorithms are constrained by a time limit, the overall performance of RCQ-GA compared to the benchmark further improves.
Personalized query suggestion based on user behavior
NASA Astrophysics Data System (ADS)
Chen, Wanyu; Hao, Zepeng; Shao, Taihua; Chen, Honghui
Query suggestions help users refine their queries after they input an initial query. Previous work mainly concentrated on similarity-based and context-based query suggestion approaches. However, models that focus on adapting to a specific user (personalization) can help to improve the probability of the user being satisfied. In this paper, we propose a personalized query suggestion model based on users’ search behavior (UB model), where we inject relevance between queries and users’ search behavior into a basic probabilistic model. For the relevance between queries, we consider their semantical similarity and co-occurrence which indicates the behavior information from other users in web search. Regarding the current user’s preference to a query, we combine the user’s short-term and long-term search behavior in a linear fashion and deal with the data sparse problem with Bayesian probabilistic matrix factorization (BPMF). In particular, we also investigate the impact of different personalization strategies (the combination of the user’s short-term and long-term search behavior) on the performance of query suggestion reranking. We quantify the improvement of our proposed UB model against a state-of-the-art baseline using the public AOL query logs and show that it beats the baseline in terms of metrics used in query suggestion reranking. The experimental results show that: (i) for personalized ranking, users’ behavioral information helps to improve query suggestion effectiveness; and (ii) given a query, merging information inferred from the short-term and long-term search behavior of a particular user can result in a better performance than both plain approaches.
APT: what it has enabled us to do
NASA Astrophysics Data System (ADS)
Blacker, Brett S.; Golombek, Daniel
2004-09-01
With the development and operations deployment of the Astronomer's Proposal Tool (APT), Hubble Space Telescope (HST) proposers have been provided with an integrated toolset for Phase I and Phase II. This toolset consists of editors for filling out proposal information, an Orbit Planner for determining observation feasibility, a Visit Planner for determining schedulability, diagnostic and reporting tools and an integrated Visual Target Tuner (VTT) for viewing exposure specifications. The VTT can also overlay HST"s field of view on user-selected Flexible Image Transport System (FITS) images, perform bright object checks and query the HST archive. In addition to these direct benefits for the HST user, STScI"s internal Phase I process has been able to take advantage of the APT products. APT has enabled a substantial streamlining of the process and software processing tools, which enabled a compression by three months of the Phase I to Phase II schedule, allowing to schedule observations earlier and thus further benefiting HST observers. Some of the improvements to our process include: creating a compact disk (CD) of Phase I products; being able to print all proposals on the day of the deadline; link the proposal in Portable Document Format (PDF) with a database, and being able to run all Phase I software on a single platform. In this paper we will discuss the operational results of using APT for HST's Cycles 12 and 13 Phase I process and will show the improvements for the users and the overall process that is allowing STScI to obtain scientific results with HST three months earlier than in previous years. We will also show how APT can be and is being used for multiple missions.
Hu, Hai; Brzeski, Henry; Hutchins, Joe; Ramaraj, Mohan; Qu, Long; Xiong, Richard; Kalathil, Surendran; Kato, Rand; Tenkillaya, Santhosh; Carney, Jerry; Redd, Rosann; Arkalgudvenkata, Sheshkumar; Shahzad, Kashif; Scott, Richard; Cheng, Hui; Meadow, Stephen; McMichael, John; Sheu, Shwu-Lin; Rosendale, David; Kvecher, Leonid; Ahern, Stephen; Yang, Song; Zhang, Yonghong; Jordan, Rick; Somiari, Stella B; Hooke, Jeffrey; Shriver, Craig D; Somiari, Richard I; Liebman, Michael N
2004-10-01
The Windber Research Institute is an integrated high-throughput research center employing clinical, genomic and proteomic platforms to produce terabyte levels of data. We use biomedical informatics technologies to integrate all of these operations. This report includes information on a multi-year, multi-phase hybrid data warehouse project currently under development in the Institute. The purpose of the warehouse is to host the terabyte-level of internal experimentally generated data as well as data from public sources. We have previously reported on the phase I development, which integrated limited internal data sources and selected public databases. Currently, we are completing phase II development, which integrates our internal automated data sources and develops visualization tools to query across these data types. This paper summarizes our clinical and experimental operations, the data warehouse development, and the challenges we have faced. In phase III we plan to federate additional manual internal and public data sources and then to develop and adapt more data analysis and mining tools. We expect that the final implementation of the data warehouse will greatly facilitate biomedical informatics research.
A target recognition method for maritime surveillance radars based on hybrid ensemble selection
NASA Astrophysics Data System (ADS)
Fan, Xueman; Hu, Shengliang; He, Jingbo
2017-11-01
In order to improve the generalisation ability of the maritime surveillance radar, a novel ensemble selection technique, termed Optimisation and Dynamic Selection (ODS), is proposed. During the optimisation phase, the non-dominated sorting genetic algorithm II for multi-objective optimisation is used to find the Pareto front, i.e. a set of ensembles of classifiers representing different tradeoffs between the classification error and diversity. During the dynamic selection phase, the meta-learning method is used to predict whether a candidate ensemble is competent enough to classify a query instance based on three different aspects, namely, feature space, decision space and the extent of consensus. The classification performance and time complexity of ODS are compared against nine other ensemble methods using a self-built full polarimetric high resolution range profile data-set. The experimental results clearly show the effectiveness of ODS. In addition, the influence of the selection of diversity measures is studied concurrently.
2013-01-01
Background Clinical Intelligence, as a research and engineering discipline, is dedicated to the development of tools for data analysis for the purposes of clinical research, surveillance, and effective health care management. Self-service ad hoc querying of clinical data is one desirable type of functionality. Since most of the data are currently stored in relational or similar form, ad hoc querying is problematic as it requires specialised technical skills and the knowledge of particular data schemas. Results A possible solution is semantic querying where the user formulates queries in terms of domain ontologies that are much easier to navigate and comprehend than data schemas. In this article, we are exploring the possibility of using SADI Semantic Web services for semantic querying of clinical data. We have developed a prototype of a semantic querying infrastructure for the surveillance of, and research on, hospital-acquired infections. Conclusions Our results suggest that SADI can support ad-hoc, self-service, semantic queries of relational data in a Clinical Intelligence context. The use of SADI compares favourably with approaches based on declarative semantic mappings from data schemas to ontologies, such as query rewriting and RDFizing by materialisation, because it can easily cope with situations when (i) some computation is required to turn relational data into RDF or OWL, e.g., to implement temporal reasoning, or (ii) integration with external data sources is necessary. PMID:23497556
A novel methodology for querying web images
NASA Astrophysics Data System (ADS)
Prabhakara, Rashmi; Lee, Ching Cheng
2005-01-01
Ever since the advent of Internet, there has been an immense growth in the amount of image data that is available on the World Wide Web. With such a magnitude of image availability, an efficient and effective image retrieval system is required to make use of this information. This research presents an effective image matching and indexing technique that improvises on existing integrated image retrieval methods. The proposed technique follows a two-phase approach, integrating query by topic and query by example specification methods. The first phase consists of topic-based image retrieval using an improved text information retrieval (IR) technique that makes use of the structured format of HTML documents. It consists of a focused crawler that not only provides for the user to enter the keyword for the topic-based search but also, the scope in which the user wants to find the images. The second phase uses the query by example specification to perform a low-level content-based image match for the retrieval of smaller and relatively closer results of the example image. Information related to the image feature is automatically extracted from the query image by the image processing system. A technique that is not computationally intensive based on color feature is used to perform content-based matching of images. The main goal is to develop a functional image search and indexing system and to demonstrate that better retrieval results can be achieved with this proposed hybrid search technique.
A novel methodology for querying web images
NASA Astrophysics Data System (ADS)
Prabhakara, Rashmi; Lee, Ching Cheng
2004-12-01
Ever since the advent of Internet, there has been an immense growth in the amount of image data that is available on the World Wide Web. With such a magnitude of image availability, an efficient and effective image retrieval system is required to make use of this information. This research presents an effective image matching and indexing technique that improvises on existing integrated image retrieval methods. The proposed technique follows a two-phase approach, integrating query by topic and query by example specification methods. The first phase consists of topic-based image retrieval using an improved text information retrieval (IR) technique that makes use of the structured format of HTML documents. It consists of a focused crawler that not only provides for the user to enter the keyword for the topic-based search but also, the scope in which the user wants to find the images. The second phase uses the query by example specification to perform a low-level content-based image match for the retrieval of smaller and relatively closer results of the example image. Information related to the image feature is automatically extracted from the query image by the image processing system. A technique that is not computationally intensive based on color feature is used to perform content-based matching of images. The main goal is to develop a functional image search and indexing system and to demonstrate that better retrieval results can be achieved with this proposed hybrid search technique.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Madduri, Kamesh; Wu, Kesheng
The Resource Description Framework (RDF) is a popular data model for representing linked data sets arising from the web, as well as large scienti c data repositories such as UniProt. RDF data intrinsically represents a labeled and directed multi-graph. SPARQL is a query language for RDF that expresses subgraph pattern- nding queries on this implicit multigraph in a SQL- like syntax. SPARQL queries generate complex intermediate join queries; to compute these joins e ciently, we propose a new strategy based on bitmap indexes. We store the RDF data in column-oriented structures as compressed bitmaps along with two dictionaries. This papermore » makes three new contributions. (i) We present an e cient parallel strategy for parsing the raw RDF data, building dictionaries of unique entities, and creating compressed bitmap indexes of the data. (ii) We utilize the constructed bitmap indexes to e ciently answer SPARQL queries, simplifying the join evaluations. (iii) To quantify the performance impact of using bitmap indexes, we compare our approach to the state-of-the-art triple-store RDF-3X. We nd that our bitmap index-based approach to answering queries is up to an order of magnitude faster for a variety of SPARQL queries, on gigascale RDF data sets.« less
Biomedical Requirements for High Productivity Computing Systems
2005-04-01
server at http://www.ncbi.nlm.nih.gov/BLAST/. There are many variants of BLAST, including: 1. BLASTN - Compares a DNA query to a DNA database. Searches ...database (3 reading frames from each strand of the DNA) searching . 13 4. TBLASTN - Compares a protein query to a DNA database, in the 6 possible...the molecular during this phase. After eliminating molecules that could not match the query , an atom-by-atom search for the molecules in conducted
Topoisomerase II Inhibitors and Poisons, and the Influence of Cell Cycle Checkpoints.
D Arcy, Nicholas; Gabrielli, Brian
2017-01-01
Interactions between the decatenation checkpoint and Topoisomerase II (TopoII) are vital for maintaining integrity of the genome. Agents that target this enzyme have been in clinical use in cancer therapy for over 30 years with great success. The types of compounds that have been developed to target TopoII are broadly divided into poisons and catalytic inhibitors. The TopoII poisons are in clinical use as anti-cancer therapies, although in common to most chemotherapeutic agents, they display considerable normal tissue toxicity. Inhibition of the TopoIIb isoform has been implicated in this cytotoxicity. Response to TopoII active agents is determined by several factors, but cell cycle checkpoints play a large role in sensitivity and resistance. The G2/M phase checkpoints are of particular importance in considering the effectiveness of these drugs and are reviewed in this article. Functionality of the ATM dependent decatenation checkpoint may represent a new avenue for selective cancer therapy. Here we review the function of TopoII, the anti-cancer mechanisms and limitations of current catalytic inhibitors and poisons, and their influence on cell cycle checkpoints. We will also assess potential new mechanisms for targeting this enzyme to limit normal tissue toxicity, and how the cell cycle checkpoint triggered by these drugs may provide an alternative and possibly better target for novel therapies. Copyright© Bentham Science Publishers; For any queries, please email at epub@benthamscience.org.
Conservation-Oriented Hbim. The Bimexplorer Web Tool
NASA Astrophysics Data System (ADS)
Quattrini, R.; Pierdicca, R.; Morbidoni, C.; Malinverni, E. S.
2017-05-01
The application of (H)BIM within the domain of Architectural Historical Heritage has huge potential that can be even exploited within the restoration domain. The work presents a novel approach to solve the widespread interoperability issue related to the data enrichment in BIM environment, by developing and testing a web tool based on a specific workflow experienced choosing as the case study a Romanic church in Portonovo, Ancona, Italy. Following the need to make the data, organized in a BIM environment, usable for the different actors involved in the restoration phase, we have created a pipeline that take advantage of BIM existing platforms and semantic-web technologies, enabling the end user to query a repository composed of semantically structured data. The pipeline of work consists in four major steps: i) modelling an ontology with the main information needs for the domain of interest, providing a data structure that can be leveraged to inform the data-enrichment phase and, later, to meaningfully query the data; ii) data enrichment, by creating a set of shared parameters reflecting the properties in our domain ontology; iii) structuring data in a machine-readable format (through a data conversion) to represent the domain (ontology) and analyse data of specific buildings respectively; iv) development of a demonstrative data exploration web application based on the faceted browsing paradigm and allowing to exploit both structured metadata and 3D visualization. The application can be configured by a domain expert to reflect a given domain ontology, and used by an operator to query and explore the data in a more efficient and reliable way. With the proposed solution the analysis of data can be reused together with the 3D model, providing the end-user with a non proprietary tool; in this way, the planned maintenance or the restoration project became more collaborative and interactive, optimizing the whole process of HBIM data collection.
Object-Oriented Query Language For Events Detection From Images Sequences
NASA Astrophysics Data System (ADS)
Ganea, Ion Eugen
2015-09-01
In this paper is presented a method to represent the events extracted from images sequences and the query language used for events detection. Using an object oriented model the spatial and temporal relationships between salient objects and also between events are stored and queried. This works aims to unify the storing and querying phases for video events processing. The object oriented language syntax used for events processing allow the instantiation of the indexes classes in order to improve the accuracy of the query results. The experiments were performed on images sequences provided from sport domain and it shows the reliability and the robustness of the proposed language. To extend the language will be added a specific syntax for constructing the templates for abnormal events and for detection of the incidents as the final goal of the research.
KBGIS-II: A knowledge-based geographic information system
NASA Technical Reports Server (NTRS)
Smith, Terence; Peuquet, Donna; Menon, Sudhakar; Agarwal, Pankaj
1986-01-01
The architecture and working of a recently implemented Knowledge-Based Geographic Information System (KBGIS-II), designed to satisfy several general criteria for the GIS, is described. The system has four major functions including query-answering, learning and editing. The main query finds constrained locations for spatial objects that are describable in a predicate-calculus based spatial object language. The main search procedures include a family of constraint-satisfaction procedures that use a spatial object knowledge base to search efficiently for complex spatial objects in large, multilayered spatial data bases. These data bases are represented in quadtree form. The search strategy is designed to reduce the computational cost of search in the average case. The learning capabilities of the system include the addition of new locations of complex spatial objects to the knowledge base as queries are answered, and the ability to learn inductively definitions of new spatial objects from examples. The new definitions are added to the knowledge base by the system. The system is performing all its designated tasks successfully. Future reports will relate performance characteristics of the system.
Web image retrieval using an effective topic and content-based technique
NASA Astrophysics Data System (ADS)
Lee, Ching-Cheng; Prabhakara, Rashmi
2005-03-01
There has been an exponential growth in the amount of image data that is available on the World Wide Web since the early development of Internet. With such a large amount of information and image available and its usefulness, an effective image retrieval system is thus greatly needed. In this paper, we present an effective approach with both image matching and indexing techniques that improvise on existing integrated image retrieval methods. This technique follows a two-phase approach, integrating query by topic and query by example specification methods. In the first phase, The topic-based image retrieval is performed by using an improved text information retrieval (IR) technique that makes use of the structured format of HTML documents. This technique consists of a focused crawler that not only provides for the user to enter the keyword for the topic-based search but also, the scope in which the user wants to find the images. In the second phase, we use query by example specification to perform a low-level content-based image match in order to retrieve smaller and relatively closer results of the example image. From this, information related to the image feature is automatically extracted from the query image. The main objective of our approach is to develop a functional image search and indexing technique and to demonstrate that better retrieval results can be achieved.
Wang, Amy Y; Lancaster, William J; Wyatt, Matthew C; Rasmussen, Luke V; Fort, Daniel G; Cimino, James J
2017-01-01
A major challenge in using electronic health record repositories for research is the difficulty matching subject eligibility criteria to query capabilities of the repositories. We propose categories for study criteria corresponding to the effort needed for querying those criteria: "easy" (supporting automated queries), mixed (initial automated querying with manual review), "hard" (fully manual record review), and "impossible" or "point of enrollment" (not typically in health repositories). We obtained a sample of 292 criteria from 20 studies from ClinicalTrials.gov. Six independent reviewers, three each from two academic research institutions, rated criteria according to our four types. We observed high interrater reliability both within and between institutions. The analysis demonstrated typical features of criteria that map with varying levels of difficulty to repositories. We propose using these features to improve enrollment workflow through more standardized study criteria, self-service repository queries, and analyst-mediated retrievals.
Wang, Amy Y.; Lancaster, William J.; Wyatt, Matthew C.; Rasmussen, Luke V.; Fort, Daniel G.; Cimino, James J.
2017-01-01
A major challenge in using electronic health record repositories for research is the difficulty matching subject eligibility criteria to query capabilities of the repositories. We propose categories for study criteria corresponding to the effort needed for querying those criteria: “easy” (supporting automated queries), mixed (initial automated querying with manual review), “hard” (fully manual record review), and “impossible” or “point of enrollment” (not typically in health repositories). We obtained a sample of 292 criteria from 20 studies from ClinicalTrials.gov. Six independent reviewers, three each from two academic research institutions, rated criteria according to our four types. We observed high interrater reliability both within and between institutions. The analysis demonstrated typical features of criteria that map with varying levels of difficulty to repositories. We propose using these features to improve enrollment workflow through more standardized study criteria, self-service repository queries, and analyst-mediated retrievals. PMID:29854246
A privacy preserving protocol for tracking participants in phase I clinical trials.
El Emam, Khaled; Farah, Hanna; Samet, Saeed; Essex, Aleksander; Jonker, Elizabeth; Kantarcioglu, Murat; Earle, Craig C
2015-10-01
Some phase 1 clinical trials offer strong financial incentives for healthy individuals to participate in their studies. There is evidence that some individuals enroll in multiple trials concurrently. This creates safety risks and introduces data quality problems into the trials. Our objective was to construct a privacy preserving protocol to track phase 1 participants to detect concurrent enrollment. A protocol using secure probabilistic querying against a database of trial participants that allows for screening during telephone interviews and on-site enrollment was developed. The match variables consisted of demographic information. The accuracy (sensitivity, precision, and negative predictive value) of the matching and its computational performance in seconds were measured under simulated environments. Accuracy was also compared to non-secure matching methods. The protocol performance scales linearly with the database size. At the largest database size of 20,000 participants, a query takes under 20s on a 64 cores machine. Sensitivity, precision, and negative predictive value of the queries were consistently at or above 0.9, and were very similar to non-secure versions of the protocol. The protocol provides a reasonable solution to the concurrent enrollment problems in phase 1 clinical trials, and is able to ensure that personal information about participants is kept secure. Copyright © 2015 The Authors. Published by Elsevier Inc. All rights reserved.
A general natural-language text processor for clinical radiology.
Friedman, C; Alderson, P O; Austin, J H; Cimino, J J; Johnson, S B
1994-01-01
OBJECTIVE: Development of a general natural-language processor that identifies clinical information in narrative reports and maps that information into a structured representation containing clinical terms. DESIGN: The natural-language processor provides three phases of processing, all of which are driven by different knowledge sources. The first phase performs the parsing. It identifies the structure of the text through use of a grammar that defines semantic patterns and a target form. The second phase, regularization, standardizes the terms in the initial target structure via a compositional mapping of multi-word phrases. The third phase, encoding, maps the terms to a controlled vocabulary. Radiology is the test domain for the processor and the target structure is a formal model for representing clinical information in that domain. MEASUREMENTS: The impression sections of 230 radiology reports were encoded by the processor. Results of an automated query of the resultant database for the occurrences of four diseases were compared with the analysis of a panel of three physicians to determine recall and precision. RESULTS: Without training specific to the four diseases, recall and precision of the system (combined effect of the processor and query generator) were 70% and 87%. Training of the query component increased recall to 85% without changing precision. PMID:7719797
A web-based data-querying tool based on ontology-driven methodology and flowchart-based model.
Ping, Xiao-Ou; Chung, Yufang; Tseng, Yi-Ju; Liang, Ja-Der; Yang, Pei-Ming; Huang, Guan-Tarn; Lai, Feipei
2013-10-08
Because of the increased adoption rate of electronic medical record (EMR) systems, more health care records have been increasingly accumulating in clinical data repositories. Therefore, querying the data stored in these repositories is crucial for retrieving the knowledge from such large volumes of clinical data. The aim of this study is to develop a Web-based approach for enriching the capabilities of the data-querying system along the three following considerations: (1) the interface design used for query formulation, (2) the representation of query results, and (3) the models used for formulating query criteria. The Guideline Interchange Format version 3.5 (GLIF3.5), an ontology-driven clinical guideline representation language, was used for formulating the query tasks based on the GLIF3.5 flowchart in the Protégé environment. The flowchart-based data-querying model (FBDQM) query execution engine was developed and implemented for executing queries and presenting the results through a visual and graphical interface. To examine a broad variety of patient data, the clinical data generator was implemented to automatically generate the clinical data in the repository, and the generated data, thereby, were employed to evaluate the system. The accuracy and time performance of the system for three medical query tasks relevant to liver cancer were evaluated based on the clinical data generator in the experiments with varying numbers of patients. In this study, a prototype system was developed to test the feasibility of applying a methodology for building a query execution engine using FBDQMs by formulating query tasks using the existing GLIF. The FBDQM-based query execution engine was used to successfully retrieve the clinical data based on the query tasks formatted using the GLIF3.5 in the experiments with varying numbers of patients. The accuracy of the three queries (ie, "degree of liver damage," "degree of liver damage when applying a mutually exclusive setting," and "treatments for liver cancer") was 100% for all four experiments (10 patients, 100 patients, 1000 patients, and 10,000 patients). Among the three measured query phases, (1) structured query language operations, (2) criteria verification, and (3) other, the first two had the longest execution time. The ontology-driven FBDQM-based approach enriched the capabilities of the data-querying system. The adoption of the GLIF3.5 increased the potential for interoperability, shareability, and reusability of the query tasks.
Goetz, Matthew B; Bowman, Candice; Hoang, Tuyen; Anaya, Henry; Osborn, Teresa; Gifford, Allen L; Asch, Steven M
2008-03-19
We describe how we used the framework of the U.S. Department of Veterans Affairs (VA) Quality Enhancement Research Initiative (QUERI) to develop a program to improve rates of diagnostic testing for the Human Immunodeficiency Virus (HIV). This venture was prompted by the observation by the CDC that 25% of HIV-infected patients do not know their diagnosis - a point of substantial importance to the VA, which is the largest provider of HIV care in the United States. Following the QUERI steps (or process), we evaluated: 1) whether undiagnosed HIV infection is a high-risk, high-volume clinical issue within the VA, 2) whether there are evidence-based recommendations for HIV testing, 3) whether there are gaps in the performance of VA HIV testing, and 4) the barriers and facilitators to improving current practice in the VA.Based on our findings, we developed and initiated a QUERI step 4/phase 1 pilot project using the precepts of the Chronic Care Model. Our improvement strategy relies upon electronic clinical reminders to provide decision support; audit/feedback as a clinical information system, and appropriate changes in delivery system design. These activities are complemented by academic detailing and social marketing interventions to achieve provider activation. Our preliminary formative evaluation indicates the need to ensure leadership and team buy-in, address facility-specific barriers, refine the reminder, and address factors that contribute to inter-clinic variances in HIV testing rates. Preliminary unadjusted data from the first seven months of our program show 3-5 fold increases in the proportion of at-risk patients who are offered HIV testing at the VA sites (stations) where the pilot project has been undertaken; no change was seen at control stations. This project demonstrates the early success of the application of the QUERI process to the development of a program to improve HIV testing rates. Preliminary unadjusted results show that the coordinated use of audit/feedback, provider activation, and organizational change can increase HIV testing rates for at-risk patients. We are refining our program prior to extending our work to a small-scale, multi-site evaluation (QUERI step 4/phase 2). We also plan to evaluate the durability/sustainability of the intervention effect, the costs of HIV testing, and the number of newly identified HIV-infected patients. Ultimately, we will evaluate this program in other geographically dispersed stations (QUERI step 4/phases 3 and 4).
Goetz, Matthew B; Bowman, Candice; Hoang, Tuyen; Anaya, Henry; Osborn, Teresa; Gifford, Allen L; Asch, Steven M
2008-01-01
Background We describe how we used the framework of the U.S. Department of Veterans Affairs (VA) Quality Enhancement Research Initiative (QUERI) to develop a program to improve rates of diagnostic testing for the Human Immunodeficiency Virus (HIV). This venture was prompted by the observation by the CDC that 25% of HIV-infected patients do not know their diagnosis – a point of substantial importance to the VA, which is the largest provider of HIV care in the United States. Methods Following the QUERI steps (or process), we evaluated: 1) whether undiagnosed HIV infection is a high-risk, high-volume clinical issue within the VA, 2) whether there are evidence-based recommendations for HIV testing, 3) whether there are gaps in the performance of VA HIV testing, and 4) the barriers and facilitators to improving current practice in the VA. Based on our findings, we developed and initiated a QUERI step 4/phase 1 pilot project using the precepts of the Chronic Care Model. Our improvement strategy relies upon electronic clinical reminders to provide decision support; audit/feedback as a clinical information system, and appropriate changes in delivery system design. These activities are complemented by academic detailing and social marketing interventions to achieve provider activation. Results Our preliminary formative evaluation indicates the need to ensure leadership and team buy-in, address facility-specific barriers, refine the reminder, and address factors that contribute to inter-clinic variances in HIV testing rates. Preliminary unadjusted data from the first seven months of our program show 3–5 fold increases in the proportion of at-risk patients who are offered HIV testing at the VA sites (stations) where the pilot project has been undertaken; no change was seen at control stations. Discussion This project demonstrates the early success of the application of the QUERI process to the development of a program to improve HIV testing rates. Preliminary unadjusted results show that the coordinated use of audit/feedback, provider activation, and organizational change can increase HIV testing rates for at-risk patients. We are refining our program prior to extending our work to a small-scale, multi-site evaluation (QUERI step 4/phase 2). We also plan to evaluate the durability/sustainability of the intervention effect, the costs of HIV testing, and the number of newly identified HIV-infected patients. Ultimately, we will evaluate this program in other geographically dispersed stations (QUERI step 4/phases 3 and 4). PMID:18353185
Unstructured medical image query using big data - An epilepsy case study.
Istephan, Sarmad; Siadat, Mohammad-Reza
2016-02-01
Big data technologies are critical to the medical field which requires new frameworks to leverage them. Such frameworks would benefit medical experts to test hypotheses by querying huge volumes of unstructured medical data to provide better patient care. The objective of this work is to implement and examine the feasibility of having such a framework to provide efficient querying of unstructured data in unlimited ways. The feasibility study was conducted specifically in the epilepsy field. The proposed framework evaluates a query in two phases. In phase 1, structured data is used to filter the clinical data warehouse. In phase 2, feature extraction modules are executed on the unstructured data in a distributed manner via Hadoop to complete the query. Three modules have been created, volume comparer, surface to volume conversion and average intensity. The framework allows for user-defined modules to be imported to provide unlimited ways to process the unstructured data hence potentially extending the application of this framework beyond epilepsy field. Two types of criteria were used to validate the feasibility of the proposed framework - the ability/accuracy of fulfilling an advanced medical query and the efficiency that Hadoop provides. For the first criterion, the framework executed an advanced medical query that spanned both structured and unstructured data with accurate results. For the second criterion, different architectures were explored to evaluate the performance of various Hadoop configurations and were compared to a traditional Single Server Architecture (SSA). The surface to volume conversion module performed up to 40 times faster than the SSA (using a 20 node Hadoop cluster) and the average intensity module performed up to 85 times faster than the SSA (using a 40 node Hadoop cluster). Furthermore, the 40 node Hadoop cluster executed the average intensity module on 10,000 models in 3h which was not even practical for the SSA. The current study is limited to epilepsy field and further research and more feature extraction modules are required to show its applicability in other medical domains. The proposed framework advances data-driven medicine by unleashing the content of unstructured medical data in an efficient and unlimited way to be harnessed by medical experts. Copyright © 2015 Elsevier Inc. All rights reserved.
Monitoring Influenza Epidemics in China with Search Query from Baidu
Lv, Benfu; Peng, Geng; Chunara, Rumi; Brownstein, John S.
2013-01-01
Several approaches have been proposed for near real-time detection and prediction of the spread of influenza. These include search query data for influenza-related terms, which has been explored as a tool for augmenting traditional surveillance methods. In this paper, we present a method that uses Internet search query data from Baidu to model and monitor influenza activity in China. The objectives of the study are to present a comprehensive technique for: (i) keyword selection, (ii) keyword filtering, (iii) index composition and (iv) modeling and detection of influenza activity in China. Sequential time-series for the selected composite keyword index is significantly correlated with Chinese influenza case data. In addition, one-month ahead prediction of influenza cases for the first eight months of 2012 has a mean absolute percent error less than 11%. To our knowledge, this is the first study on the use of search query data from Baidu in conjunction with this approach for estimation of influenza activity in China. PMID:23750192
Otte, Willem M; van Diessen, Eric; Bell, Gail S; Sander, Josemir W
2013-12-01
In old and modern times and across cultures, recurrent seizures have been attributed to the lunar phase. It is unclear whether this relationship should be classified as a myth or whether a true connection exists between moon phases and seizures. We analyzed the worldwide aggregated search queries related to epilepsy health-seeking behavior between 2005 and 2012. Epilepsy-related Internet searches increased in periods with a high moon illumination. The overall association was weak (r=0.11, 95% confidence interval: 0.07 to 0.14) but seems to be higher than most control search queries not related to epilepsy. Increased sleep deprivation during periods of full moon might explain this positive association and warrants further study into epilepsy-related health-seeking behavior on the Internet, the lunar phase, and its contribution to nocturnal luminance. © 2013.
Classification of ECG beats using deep belief network and active learning.
G, Sayantan; T, Kien P; V, Kadambari K
2018-04-12
A new semi-supervised approach based on deep learning and active learning for classification of electrocardiogram signals (ECG) is proposed. The objective of the proposed work is to model a scientific method for classification of cardiac irregularities using electrocardiogram beats. The model follows the Association for the Advancement of medical instrumentation (AAMI) standards and consists of three phases. In phase I, feature representation of ECG is learnt using Gaussian-Bernoulli deep belief network followed by a linear support vector machine (SVM) training in the consecutive phase. It yields three deep models which are based on AAMI-defined classes, namely N, V, S, and F. In the last phase, a query generator is introduced to interact with the expert to label few beats to improve accuracy and sensitivity. The proposed approach depicts significant improvement in accuracy with minimal queries posed to the expert and fast online training as tested on the MIT-BIH Arrhythmia Database and the MIT-BIH Supra-ventricular Arrhythmia Database (SVDB). With 100 queries labeled by the expert in phase III, the method achieves an accuracy of 99.5% in "S" versus all classifications (SVEB) and 99.4% accuracy in "V " versus all classifications (VEB) on MIT-BIH Arrhythmia Database. In a similar manner, it is attributed that an accuracy of 97.5% for SVEB and 98.6% for VEB on SVDB database is achieved respectively. Graphical Abstract Reply- Deep belief network augmented by active learning for efficient prediction of arrhythmia.
Pradeep, Jothimani; Ambroise, Stanley; Gunasekaran, Dhandapany
2017-01-01
Introduction Query (Q) fever is an important zoonosis and a cause of concern for humans, due to the potential bioterrorism threat posed by the causative agent, Coxiella burnetii. Because of the danger of contracting the illness, isolation attempts are seldom made. Serological and molecular diagnostic tests are the main option. Aim To study the prevalence of acute Q fever in Puducherry and surrounding districts of Tamil Nadu, India, employing a new commercial Real-Time Polymerase Chain Reaction (RT-PCR) kit and confirming it by the gold standard Immunofluorescence Assay (IFA). Materials and Methods Acute phase blood samples from 72 consecutive febrile patients and 24 healthy individuals were included in this prospective study. DNA was extracted from the buffy coats and preserved at -80°C. Detection of C. burnetii was carried out employing a commercial Real-Time PCR kit. Serum samples were tested for IgM (Phase I+II) and IgG (Phase I+II) by QM-120 and QG-120, Coxiella burnetii IFA Fuller Laboratories, California, USA. Sensitivity, Specificity, Positive Predictive Value (PPV) and Negative Predictive Value (NPV) were calculated keeping IFA as the reference. Results Presumptive diagnosis of acute Q fever was made in two febrile patients by the Genesig Easy kit (2.78%). In addition to these two PCR positive cases, one more patient was positive for both Phase II IgM and Phase II IgG antibodies by the gold standard IFA. All 24 healthy controls were negative for Q fever by both PCR and IFA. The sensitivity, specificity, NPV and PPV for Genesig Easy kit PCR were: 66.67%, 100%, 100% and 98.57 % respectively against IFA as the reference. Conclusion The true prevalence of Q fever in India and other developing countries is poorly understood, owing to the difficulties in the diagnosis of this infection. Since molecular diagnostic tests have good specificity and are mandated for confirmation of single acute samples, validation of commercial Q fever PCR kits is the need of the hour. Genesig Easy kit in our hands was found to be reliable with the moderate sensitivity and high specificity. Performing both PCR (with acute specimens) and IFA (with paired sera) would be ideal for Q fever diagnosis. PMID:29207703
Semantics Enabled Queries in EuroGEOSS: a Discovery Augmentation Approach
NASA Astrophysics Data System (ADS)
Santoro, M.; Mazzetti, P.; Fugazza, C.; Nativi, S.; Craglia, M.
2010-12-01
One of the main challenges in Earth Science Informatics is to build interoperability frameworks which allow users to discover, evaluate, and use information from different scientific domains. This needs to address multidisciplinary interoperability challenges concerning both technological and scientific aspects. From the technological point of view, it is necessary to provide a set of special interoperability arrangement in order to develop flexible frameworks that allow a variety of loosely-coupled services to interact with each other. From a scientific point of view, it is necessary to document clearly the theoretical and methodological assumptions underpinning applications in different scientific domains, and develop cross-domain ontologies to facilitate interdisciplinary dialogue and understanding. In this presentation we discuss a brokering approach that extends the traditional Service Oriented Architecture (SOA) adopted by most Spatial Data Infrastructures (SDIs) to provide the necessary special interoperability arrangements. In the EC-funded EuroGEOSS (A European approach to GEOSS) project, we distinguish among three possible functional brokering components: discovery, access and semantics brokers. This presentation focuses on the semantics broker, the Discovery Augmentation Component (DAC), which was specifically developed to address the three thematic areas covered by the EuroGEOSS project: biodiversity, forestry and drought. The EuroGEOSS DAC federates both semantics (e.g. SKOS repositories) and ISO-compliant geospatial catalog services. The DAC can be queried using common geospatial constraints (i.e. what, where, when, etc.). Two different augmented discovery styles are supported: a) automatic query expansion; b) user assisted query expansion. In the first case, the main discovery steps are: i. the query keywords (the what constraint) are “expanded” with related concepts/terms retrieved from the set of federated semantic services. A default expansion regards the multilinguality relationship; ii. The resulting queries are submitted to the federated catalog services; iii. The DAC performs a “smart” aggregation of the queries results and provides them back to the client. In the second case, the main discovery steps are: i. the user browses the federated semantic repositories and selects the concepts/terms-of-interest; ii. The DAC creates the set of geospatial queries based on the selected concepts/terms and submits them to the federated catalog services; iii. The DAC performs a “smart” aggregation of the queries results and provides them back to the client. A Graphical User Interface (GUI) was also developed for testing and interacting with the DAC. The entire brokering framework is deployed in the context of EuroGEOSS infrastructure and it is used in a couple of GEOSS AIP-3 use scenarios: the “e-Habitat Use Scenario” for the Biodiversity and Climate Change topic, and the “Comprehensive Drought Index Use Scenario” for Water/Drought topic
A Web-Based Data-Querying Tool Based on Ontology-Driven Methodology and Flowchart-Based Model
Ping, Xiao-Ou; Chung, Yufang; Liang, Ja-Der; Yang, Pei-Ming; Huang, Guan-Tarn; Lai, Feipei
2013-01-01
Background Because of the increased adoption rate of electronic medical record (EMR) systems, more health care records have been increasingly accumulating in clinical data repositories. Therefore, querying the data stored in these repositories is crucial for retrieving the knowledge from such large volumes of clinical data. Objective The aim of this study is to develop a Web-based approach for enriching the capabilities of the data-querying system along the three following considerations: (1) the interface design used for query formulation, (2) the representation of query results, and (3) the models used for formulating query criteria. Methods The Guideline Interchange Format version 3.5 (GLIF3.5), an ontology-driven clinical guideline representation language, was used for formulating the query tasks based on the GLIF3.5 flowchart in the Protégé environment. The flowchart-based data-querying model (FBDQM) query execution engine was developed and implemented for executing queries and presenting the results through a visual and graphical interface. To examine a broad variety of patient data, the clinical data generator was implemented to automatically generate the clinical data in the repository, and the generated data, thereby, were employed to evaluate the system. The accuracy and time performance of the system for three medical query tasks relevant to liver cancer were evaluated based on the clinical data generator in the experiments with varying numbers of patients. Results In this study, a prototype system was developed to test the feasibility of applying a methodology for building a query execution engine using FBDQMs by formulating query tasks using the existing GLIF. The FBDQM-based query execution engine was used to successfully retrieve the clinical data based on the query tasks formatted using the GLIF3.5 in the experiments with varying numbers of patients. The accuracy of the three queries (ie, “degree of liver damage,” “degree of liver damage when applying a mutually exclusive setting,” and “treatments for liver cancer”) was 100% for all four experiments (10 patients, 100 patients, 1000 patients, and 10,000 patients). Among the three measured query phases, (1) structured query language operations, (2) criteria verification, and (3) other, the first two had the longest execution time. Conclusions The ontology-driven FBDQM-based approach enriched the capabilities of the data-querying system. The adoption of the GLIF3.5 increased the potential for interoperability, shareability, and reusability of the query tasks. PMID:25600078
Music Identification System Using MPEG-7 Audio Signature Descriptors
You, Shingchern D.; Chen, Wei-Hwa; Chen, Woei-Kae
2013-01-01
This paper describes a multiresolution system based on MPEG-7 audio signature descriptors for music identification. Such an identification system may be used to detect illegally copied music circulated over the Internet. In the proposed system, low-resolution descriptors are used to search likely candidates, and then full-resolution descriptors are used to identify the unknown (query) audio. With this arrangement, the proposed system achieves both high speed and high accuracy. To deal with the problem that a piece of query audio may not be inside the system's database, we suggest two different methods to find the decision threshold. Simulation results show that the proposed method II can achieve an accuracy of 99.4% for query inputs both inside and outside the database. Overall, it is highly possible to use the proposed system for copyright control. PMID:23533359
FASH: A web application for nucleotides sequence search.
Veksler-Lublinksy, Isana; Barash, Danny; Avisar, Chai; Troim, Einav; Chew, Paul; Kedem, Klara
2008-05-27
: FASH (Fourier Alignment Sequence Heuristics) is a web application, based on the Fast Fourier Transform, for finding remote homologs within a long nucleic acid sequence. Given a query sequence and a long text-sequence (e.g, the human genome), FASH detects subsequences within the text that are remotely-similar to the query. FASH offers an alternative approach to Blast/Fasta for querying long RNA/DNA sequences. FASH differs from these other approaches in that it does not depend on the existence of contiguous seed-sequences in its initial detection phase. The FASH web server is user friendly and very easy to operate. FASH can be accessed athttps://fash.bgu.ac.il:8443/fash/default.jsp (secured website).
A natural language query system for Hubble Space Telescope proposal selection
NASA Technical Reports Server (NTRS)
Hornick, Thomas; Cohen, William; Miller, Glenn
1987-01-01
The proposal selection process for the Hubble Space Telescope is assisted by a robust and easy to use query program (TACOS). The system parses an English subset language sentence regardless of the order of the keyword phases, allowing the user a greater flexibility than a standard command query language. Capabilities for macro and procedure definition are also integrated. The system was designed for flexibility in both use and maintenance. In addition, TACOS can be applied to any knowledge domain that can be expressed in terms of a single reaction. The system was implemented mostly in Common LISP. The TACOS design is described in detail, with particular attention given to the implementation methods of sentence processing.
CUFID-query: accurate network querying through random walk based network flow estimation.
Jeong, Hyundoo; Qian, Xiaoning; Yoon, Byung-Jun
2017-12-28
Functional modules in biological networks consist of numerous biomolecules and their complicated interactions. Recent studies have shown that biomolecules in a functional module tend to have similar interaction patterns and that such modules are often conserved across biological networks of different species. As a result, such conserved functional modules can be identified through comparative analysis of biological networks. In this work, we propose a novel network querying algorithm based on the CUFID (Comparative network analysis Using the steady-state network Flow to IDentify orthologous proteins) framework combined with an efficient seed-and-extension approach. The proposed algorithm, CUFID-query, can accurately detect conserved functional modules as small subnetworks in the target network that are expected to perform similar functions to the given query functional module. The CUFID framework was recently developed for probabilistic pairwise global comparison of biological networks, and it has been applied to pairwise global network alignment, where the framework was shown to yield accurate network alignment results. In the proposed CUFID-query algorithm, we adopt the CUFID framework and extend it for local network alignment, specifically to solve network querying problems. First, in the seed selection phase, the proposed method utilizes the CUFID framework to compare the query and the target networks and to predict the probabilistic node-to-node correspondence between the networks. Next, the algorithm selects and greedily extends the seed in the target network by iteratively adding nodes that have frequent interactions with other nodes in the seed network, in a way that the conductance of the extended network is maximally reduced. Finally, CUFID-query removes irrelevant nodes from the querying results based on the personalized PageRank vector for the induced network that includes the fully extended network and its neighboring nodes. Through extensive performance evaluation based on biological networks with known functional modules, we show that CUFID-query outperforms the existing state-of-the-art algorithms in terms of prediction accuracy and biological significance of the predictions.
GIS-based accident location and analysis system (GIS-ALAS) : project report : phase I
DOT National Transportation Integrated Search
1998-04-06
This report summarizes progress made in Phase I of the geographic information system (GIS) based Accident Location and Analysis System (GIS-ALAS). The GIS-ALAS project builds on PC-ALAS, a locationally-referenced highway crash database query system d...
Video indexing based on image and sound
NASA Astrophysics Data System (ADS)
Faudemay, Pascal; Montacie, Claude; Caraty, Marie-Jose
1997-10-01
Video indexing is a major challenge for both scientific and economic reasons. Information extraction can sometimes be easier from sound channel than from image channel. We first present a multi-channel and multi-modal query interface, to query sound, image and script through 'pull' and 'push' queries. We then summarize the segmentation phase, which needs information from the image channel. Detection of critical segments is proposed. It should speed-up both automatic and manual indexing. We then present an overview of the information extraction phase. Information can be extracted from the sound channel, through speaker recognition, vocal dictation with unconstrained vocabularies, and script alignment with speech. We present experiment results for these various techniques. Speaker recognition methods were tested on the TIMIT and NTIMIT database. Vocal dictation as experimented on newspaper sentences spoken by several speakers. Script alignment was tested on part of a carton movie, 'Ivanhoe'. For good quality sound segments, error rates are low enough for use in indexing applications. Major issues are the processing of sound segments with noise or music, and performance improvement through the use of appropriate, low-cost architectures or networks of workstations.
Optimizability of OGC Standards Implementations - a Case Study
NASA Astrophysics Data System (ADS)
Misev, D.; Baumann, P.
2012-04-01
Why do we shop at Amazon? Because they have a unique offering that is nowhere else available? Certainly not. Rather, Amazon offers (i) simple, yet effective search; (ii) very simple payment; (iii) extremely rapid delivery. This is how scientific services will be distinguished in future: not for their data holding (there will be manifold choice), but for their service quality. We are facing the transition from data stewardship to service stewardship. One of the OGC standards which particularly enables flexible retrieval is the Web Coverage Processing Service (WCPS). It defines a high-level query language on large, multi-dimensional raster data, such as 1D timeseries, 2D EO imagery, 3D x/y/t image time series and x/y/z geophysical data, 4D x/y/z/t climate and ocean data. We have implemented WCPS based on an Array Database Management System, rasdaman, which is available in open source. In this demonstration, we study WCPS queries on 2D, 3D, and 4D data sets. Particular emphasis is placed on the computational load queries generate in such on-demand processing and filtering. We look at different techniques and their impact on performance, such as adaptive storage partitioning, query rewriting, and just-in-time compilation. Results show that there is significant potential for effective server-side optimization once a query language is sufficiently high-level and declarative.
CellLineNavigator: a workbench for cancer cell line analysis
Krupp, Markus; Itzel, Timo; Maass, Thorsten; Hildebrandt, Andreas; Galle, Peter R.; Teufel, Andreas
2013-01-01
The CellLineNavigator database, freely available at http://www.medicalgenomics.org/celllinenavigator, is a web-based workbench for large scale comparisons of a large collection of diverse cell lines. It aims to support experimental design in the fields of genomics, systems biology and translational biomedical research. Currently, this compendium holds genome wide expression profiles of 317 different cancer cell lines, categorized into 57 different pathological states and 28 individual tissues. To enlarge the scope of CellLineNavigator, the database was furthermore closely linked to commonly used bioinformatics databases and knowledge repositories. To ensure easy data access and search ability, a simple data and an intuitive querying interface were implemented. It allows the user to explore and filter gene expression, focusing on pathological or physiological conditions. For a more complex search, the advanced query interface may be used to query for (i) differentially expressed genes; (ii) pathological or physiological conditions; or (iii) gene names or functional attributes, such as Kyoto Encyclopaedia of Genes and Genomes pathway maps. These queries may also be combined. Finally, CellLineNavigator allows additional advanced analysis of differentially regulated genes by a direct link to the Database for Annotation, Visualization and Integrated Discovery (DAVID) Bioinformatics Resources. PMID:23118487
Improving integrative searching of systems chemical biology data using semantic annotation.
Chen, Bin; Ding, Ying; Wild, David J
2012-03-08
Systems chemical biology and chemogenomics are considered critical, integrative disciplines in modern biomedical research, but require data mining of large, integrated, heterogeneous datasets from chemistry and biology. We previously developed an RDF-based resource called Chem2Bio2RDF that enabled querying of such data using the SPARQL query language. Whilst this work has proved useful in its own right as one of the first major resources in these disciplines, its utility could be greatly improved by the application of an ontology for annotation of the nodes and edges in the RDF graph, enabling a much richer range of semantic queries to be issued. We developed a generalized chemogenomics and systems chemical biology OWL ontology called Chem2Bio2OWL that describes the semantics of chemical compounds, drugs, protein targets, pathways, genes, diseases and side-effects, and the relationships between them. The ontology also includes data provenance. We used it to annotate our Chem2Bio2RDF dataset, making it a rich semantic resource. Through a series of scientific case studies we demonstrate how this (i) simplifies the process of building SPARQL queries, (ii) enables useful new kinds of queries on the data and (iii) makes possible intelligent reasoning and semantic graph mining in chemogenomics and systems chemical biology. Chem2Bio2OWL is available at http://chem2bio2rdf.org/owl. The document is available at http://chem2bio2owl.wikispaces.com.
Secure quantum private information retrieval using phase-encoded queries
NASA Astrophysics Data System (ADS)
Olejnik, Lukasz
2011-08-01
We propose a quantum solution to the classical private information retrieval (PIR) problem, which allows one to query a database in a private manner. The protocol offers privacy thresholds and allows the user to obtain information from a database in a way that offers the potential adversary, in this model the database owner, no possibility of deterministically establishing the query contents. This protocol may also be viewed as a solution to the symmetrically private information retrieval problem in that it can offer database security (inability for a querying user to steal its contents). Compared to classical solutions, the protocol offers substantial improvement in terms of communication complexity. In comparison with the recent quantum private queries [Phys. Rev. Lett.PRLTAO0031-900710.1103/PhysRevLett.100.230502 100, 230502 (2008)] protocol, it is more efficient in terms of communication complexity and the number of rounds, while offering a clear privacy parameter. We discuss the security of the protocol and analyze its strengths and conclude that using this technique makes it challenging to obtain the unconditional (in the information-theoretic sense) privacy degree; nevertheless, in addition to being simple, the protocol still offers a privacy level. The oracle used in the protocol is inspired both by the classical computational PIR solutions as well as the Deutsch-Jozsa oracle.
Secure quantum private information retrieval using phase-encoded queries
DOE Office of Scientific and Technical Information (OSTI.GOV)
Olejnik, Lukasz
We propose a quantum solution to the classical private information retrieval (PIR) problem, which allows one to query a database in a private manner. The protocol offers privacy thresholds and allows the user to obtain information from a database in a way that offers the potential adversary, in this model the database owner, no possibility of deterministically establishing the query contents. This protocol may also be viewed as a solution to the symmetrically private information retrieval problem in that it can offer database security (inability for a querying user to steal its contents). Compared to classical solutions, the protocol offersmore » substantial improvement in terms of communication complexity. In comparison with the recent quantum private queries [Phys. Rev. Lett. 100, 230502 (2008)] protocol, it is more efficient in terms of communication complexity and the number of rounds, while offering a clear privacy parameter. We discuss the security of the protocol and analyze its strengths and conclude that using this technique makes it challenging to obtain the unconditional (in the information-theoretic sense) privacy degree; nevertheless, in addition to being simple, the protocol still offers a privacy level. The oracle used in the protocol is inspired both by the classical computational PIR solutions as well as the Deutsch-Jozsa oracle.« less
The national coal-resources data system of the U.S. geological survey
Carter, M.D.
1976-01-01
The National Coal Resources Data System (NCRDS) was designed by the U.S. Geological Survey (USGS) to meet the increasing demands for rapid retrieval of information on coal location, quantity, quality, and accessibility. An interactive conversational query system devised by the USGS retrieves information from the data bank through a standard computer terminal. The system is being developed in two phases. Phase I, which currently is available on a limited basis, contains published areal resource and chemical data. The primary objective of this phase is to retrieve, calculate, and tabulate coal-resource data by area on a local, regional, or national scale. Factors available for retrieval include: state, county, quadrangle, township, coal field, coal bed, formation, geologic age, source and reliability of data, and coal-bed rank, thickness, overburden, and tonnage, or any combinations of variables. In addition, the chemical data items include individual values for proximate and ultimate analyses, BTU value, and several other physical and chemical tests. Information will be validated and deleted or updated as needed. Phase II is being developed to store, retrieve, and manipulate basic point source coal data (e.g., field observations, drill-hole logs), including geodetic location; bed thickness; depth of burial; moisture; ash; sulfur; major-, minor-, and trace-element content; heat value; and characteristics of overburden, roof rocks, and floor rocks. The computer system may be used to generate interactively structure-contour or isoline maps of the physical and chemical characteristics of a coal bed or to calculate coal resources. ?? 1976.
Balaur, Irina; Saqi, Mansoor; Barat, Ana; Lysenko, Artem; Mazein, Alexander; Rawlings, Christopher J; Ruskin, Heather J; Auffray, Charles
2017-10-01
The development of colorectal cancer (CRC)-the third most common cancer type-has been associated with deregulations of cellular mechanisms stimulated by both genetic and epigenetic events. StatEpigen is a manually curated and annotated database, containing information on interdependencies between genetic and epigenetic signals, and specialized currently for CRC research. Although StatEpigen provides a well-developed graphical user interface for information retrieval, advanced queries involving associations between multiple concepts can benefit from more detailed graph representation of the integrated data. This can be achieved by using a graph database (NoSQL) approach. Data were extracted from StatEpigen and imported to our newly developed EpiGeNet, a graph database for storage and querying of conditional relationships between molecular (genetic and epigenetic) events observed at different stages of colorectal oncogenesis. We illustrate the enhanced capability of EpiGeNet for exploration of different queries related to colorectal tumor progression; specifically, we demonstrate the query process for (i) stage-specific molecular events, (ii) most frequently observed genetic and epigenetic interdependencies in colon adenoma, and (iii) paths connecting key genes reported in CRC and associated events. The EpiGeNet framework offers improved capability for management and visualization of data on molecular events specific to CRC initiation and progression.
EvoSNP-DB: A database of genetic diversity in East Asian populations.
Kim, Young Uk; Kim, Young Jin; Lee, Jong-Young; Park, Kiejung
2013-08-01
Genome-wide association studies (GWAS) have become popular as an approach for the identification of large numbers of phenotype-associated variants. However, differences in genetic architecture and environmental factors mean that the effect of variants can vary across populations. Understanding population genetic diversity is valuable for the investigation of possible population specific and independent effects of variants. EvoSNP-DB aims to provide information regarding genetic diversity among East Asian populations, including Chinese, Japanese, and Korean. Non-redundant SNPs (1.6 million) were genotyped in 54 Korean trios (162 samples) and were compared with 4 million SNPs from HapMap phase II populations. EvoSNP-DB provides two user interfaces for data query and visualization, and integrates scores of genetic diversity (Fst and VarLD) at the level of SNPs, genes, and chromosome regions. EvoSNP-DB is a web-based application that allows users to navigate and visualize measurements of population genetic differences in an interactive manner, and is available online at [http://biomi.cdc.go.kr/EvoSNP/].
PROTICdb: a web-based application to store, track, query, and compare plant proteome data.
Ferry-Dumazet, Hélène; Houel, Gwenn; Montalent, Pierre; Moreau, Luc; Langella, Olivier; Negroni, Luc; Vincent, Delphine; Lalanne, Céline; de Daruvar, Antoine; Plomion, Christophe; Zivy, Michel; Joets, Johann
2005-05-01
PROTICdb is a web-based application, mainly designed to store and analyze plant proteome data obtained by two-dimensional polyacrylamide gel electrophoresis (2-D PAGE) and mass spectrometry (MS). The purposes of PROTICdb are (i) to store, track, and query information related to proteomic experiments, i.e., from tissue sampling to protein identification and quantitative measurements, and (ii) to integrate information from the user's own expertise and other sources into a knowledge base, used to support data interpretation (e.g., for the determination of allelic variants or products of post-translational modifications). Data insertion into the relational database of PROTICdb is achieved either by uploading outputs of image analysis and MS identification software, or by filling web forms. 2-D PAGE annotated maps can be displayed, queried, and compared through a graphical interface. Links to external databases are also available. Quantitative data can be easily exported in a tabulated format for statistical analyses. PROTICdb is based on the Oracle or the PostgreSQL Database Management System and is freely available upon request at the following URL: http://moulon.inra.fr/ bioinfo/PROTICdb.
ERIC Educational Resources Information Center
Cornell Univ., Ithaca, NY. Dept. of Computer Science.
Part Two of the eighteenth report on Salton's Magical Automatic Retriever of Texts (SMART) project is composed of three papers: The first: "The Effect of Common Words and Synonyms on Retrieval Performance" by D. Bergmark discloses that removal of common words from the query and document vectors significantly increases precision and that…
SAFE: SPARQL Federation over RDF Data Cubes with Access Control.
Khan, Yasar; Saleem, Muhammad; Mehdi, Muntazir; Hogan, Aidan; Mehmood, Qaiser; Rebholz-Schuhmann, Dietrich; Sahay, Ratnesh
2017-02-01
Several query federation engines have been proposed for accessing public Linked Open Data sources. However, in many domains, resources are sensitive and access to these resources is tightly controlled by stakeholders; consequently, privacy is a major concern when federating queries over such datasets. In the Healthcare and Life Sciences (HCLS) domain real-world datasets contain sensitive statistical information: strict ownership is granted to individuals working in hospitals, research labs, clinical trial organisers, etc. Therefore, the legal and ethical concerns on (i) preserving the anonymity of patients (or clinical subjects); and (ii) respecting data ownership through access control; are key challenges faced by the data analytics community working within the HCLS domain. Likewise statistical data play a key role in the domain, where the RDF Data Cube Vocabulary has been proposed as a standard format to enable the exchange of such data. However, to the best of our knowledge, no existing approach has looked to optimise federated queries over such statistical data. We present SAFE: a query federation engine that enables policy-aware access to sensitive statistical datasets represented as RDF data cubes. SAFE is designed specifically to query statistical RDF data cubes in a distributed setting, where access control is coupled with source selection, user profiles and their access rights. SAFE proposes a join-aware source selection method that avoids wasteful requests to irrelevant and unauthorised data sources. In order to preserve anonymity and enforce stricter access control, SAFE's indexing system does not hold any data instances-it stores only predicates and endpoints. The resulting data summary has a significantly lower index generation time and size compared to existing engines, which allows for faster updates when sources change. We validate the performance of the system with experiments over real-world datasets provided by three clinical organisations as well as legacy linked datasets. We show that SAFE enables granular graph-level access control over distributed clinical RDF data cubes and efficiently reduces the source selection and overall query execution time when compared with general-purpose SPARQL query federation engines in the targeted setting.
HomPPI: a class of sequence homology based protein-protein interface prediction methods
2011-01-01
Background Although homology-based methods are among the most widely used methods for predicting the structure and function of proteins, the question as to whether interface sequence conservation can be effectively exploited in predicting protein-protein interfaces has been a subject of debate. Results We studied more than 300,000 pair-wise alignments of protein sequences from structurally characterized protein complexes, including both obligate and transient complexes. We identified sequence similarity criteria required for accurate homology-based inference of interface residues in a query protein sequence. Based on these analyses, we developed HomPPI, a class of sequence homology-based methods for predicting protein-protein interface residues. We present two variants of HomPPI: (i) NPS-HomPPI (Non partner-specific HomPPI), which can be used to predict interface residues of a query protein in the absence of knowledge of the interaction partner; and (ii) PS-HomPPI (Partner-specific HomPPI), which can be used to predict the interface residues of a query protein with a specific target protein. Our experiments on a benchmark dataset of obligate homodimeric complexes show that NPS-HomPPI can reliably predict protein-protein interface residues in a given protein, with an average correlation coefficient (CC) of 0.76, sensitivity of 0.83, and specificity of 0.78, when sequence homologs of the query protein can be reliably identified. NPS-HomPPI also reliably predicts the interface residues of intrinsically disordered proteins. Our experiments suggest that NPS-HomPPI is competitive with several state-of-the-art interface prediction servers including those that exploit the structure of the query proteins. The partner-specific classifier, PS-HomPPI can, on a large dataset of transient complexes, predict the interface residues of a query protein with a specific target, with a CC of 0.65, sensitivity of 0.69, and specificity of 0.70, when homologs of both the query and the target can be reliably identified. The HomPPI web server is available at http://homppi.cs.iastate.edu/. Conclusions Sequence homology-based methods offer a class of computationally efficient and reliable approaches for predicting the protein-protein interface residues that participate in either obligate or transient interactions. For query proteins involved in transient interactions, the reliability of interface residue prediction can be improved by exploiting knowledge of putative interaction partners. PMID:21682895
DOE Office of Scientific and Technical Information (OSTI.GOV)
Abe Lederman
This report contains the comprehensive summary of the work performed on the SBIR Phase II project (“Distributed Relevance Ranking in Heterogeneous Document Collections”) at Deep Web Technologies (http://www.deepwebtech.com). We have successfully completed all of the tasks defined in our SBIR Proposal work plan (See Table 1 - Phase II Tasks Status). The project was completed on schedule and we have successfully deployed an initial production release of the software architecture at DOE-OSTI for the Science.gov Alliance's search portal (http://www.science.gov). We have implemented a set of grid services that supports the extraction, filtering, aggregation, and presentation of search results from numerousmore » heterogeneous document collections. Illustration 3 depicts the services required to perform QuickRank™ filtering of content as defined in our architecture documentation. Functionality that has been implemented is indicated by the services highlighted in green. We have successfully tested our implementation in a multi-node grid deployment both within the Deep Web Technologies offices, and in a heterogeneous geographically distributed grid environment. We have performed a series of load tests in which we successfully simulated 100 concurrent users submitting search requests to the system. This testing was performed on deployments of one, two, and three node grids with services distributed in a number of different configurations. The preliminary results from these tests indicate that our architecture will scale well across multi-node grid deployments, but more work will be needed, beyond the scope of this project, to perform testing and experimentation to determine scalability and resiliency requirements. We are pleased to report that a production quality version (1.4) of the science.gov Alliance's search portal based on our grid architecture was released in June of 2006. This demonstration portal is currently available at http://science.gov/search30 . The portal allows the user to select from a number of collections grouped by category and enter a query expression (See Illustration 1 - Science.gov 3.0 Search Page). After the user clicks “search” a results page is displayed that provides a list of results from the selected collections ordered by relevance based on the query expression the user provided. Our grid based solution to deep web search and document ranking has already gained attention within DOE, other Government Agencies and a fortune 50 company. We are committed to the continued development of grid based solutions to large scale data access, filtering, and presentation problems within the domain of Information Retrieval and the more general categories of content management, data mining and data analysis.« less
G-Hash: Towards Fast Kernel-based Similarity Search in Large Graph Databases.
Wang, Xiaohong; Smalter, Aaron; Huan, Jun; Lushington, Gerald H
2009-01-01
Structured data including sets, sequences, trees and graphs, pose significant challenges to fundamental aspects of data management such as efficient storage, indexing, and similarity search. With the fast accumulation of graph databases, similarity search in graph databases has emerged as an important research topic. Graph similarity search has applications in a wide range of domains including cheminformatics, bioinformatics, sensor network management, social network management, and XML documents, among others.Most of the current graph indexing methods focus on subgraph query processing, i.e. determining the set of database graphs that contains the query graph and hence do not directly support similarity search. In data mining and machine learning, various graph kernel functions have been designed to capture the intrinsic similarity of graphs. Though successful in constructing accurate predictive and classification models for supervised learning, graph kernel functions have (i) high computational complexity and (ii) non-trivial difficulty to be indexed in a graph database.Our objective is to bridge graph kernel function and similarity search in graph databases by proposing (i) a novel kernel-based similarity measurement and (ii) an efficient indexing structure for graph data management. Our method of similarity measurement builds upon local features extracted from each node and their neighboring nodes in graphs. A hash table is utilized to support efficient storage and fast search of the extracted local features. Using the hash table, a graph kernel function is defined to capture the intrinsic similarity of graphs and for fast similarity query processing. We have implemented our method, which we have named G-hash, and have demonstrated its utility on large chemical graph databases. Our results show that the G-hash method achieves state-of-the-art performance for k-nearest neighbor (k-NN) classification. Most importantly, the new similarity measurement and the index structure is scalable to large database with smaller indexing size, faster indexing construction time, and faster query processing time as compared to state-of-the-art indexing methods such as C-tree, gIndex, and GraphGrep.
FPGA-based prototype storage system with phase change memory
NASA Astrophysics Data System (ADS)
Li, Gezi; Chen, Xiaogang; Chen, Bomy; Li, Shunfen; Zhou, Mi; Han, Wenbing; Song, Zhitang
2016-10-01
With the ever-increasing amount of data being stored via social media, mobile telephony base stations, and network devices etc. the database systems face severe bandwidth bottlenecks when moving vast amounts of data from storage to the processing nodes. At the same time, Storage Class Memory (SCM) technologies such as Phase Change Memory (PCM) with unique features like fast read access, high density, non-volatility, byte-addressability, positive response to increasing temperature, superior scalability, and zero standby leakage have changed the landscape of modern computing and storage systems. In such a scenario, we present a storage system called FLEET which can off-load partial or whole SQL queries to the storage engine from CPU. FLEET uses an FPGA rather than conventional CPUs to implement the off-load engine due to its highly parallel nature. We have implemented an initial prototype of FLEET with PCM-based storage. The results demonstrate that significant performance and CPU utilization gains can be achieved by pushing selected query processing components inside in PCM-based storage.
García-Remesal, M; Maojo, V; Billhardt, H; Crespo, J
2010-01-01
Bringing together structured and text-based sources is an exciting challenge for biomedical informaticians, since most relevant biomedical sources belong to one of these categories. In this paper we evaluate the feasibility of integrating relational and text-based biomedical sources using: i) an original logical schema acquisition method for textual databases developed by the authors, and ii) OntoFusion, a system originally designed by the authors for the integration of relational sources. We conducted an integration experiment involving a test set of seven differently structured sources covering the domain of genetic diseases. We used our logical schema acquisition method to generate schemas for all textual sources. The sources were integrated using the methods and tools provided by OntoFusion. The integration was validated using a test set of 500 queries. A panel of experts answered a questionnaire to evaluate i) the quality of the extracted schemas, ii) the query processing performance of the integrated set of sources, and iii) the relevance of the retrieved results. The results of the survey show that our method extracts coherent and representative logical schemas. Experts' feedback on the performance of the integrated system and the relevance of the retrieved results was also positive. Regarding the validation of the integration, the system successfully provided correct results for all queries in the test set. The results of the experiment suggest that text-based sources including a logical schema can be regarded as equivalent to structured databases. Using our method, previous research and existing tools designed for the integration of structured databases can be reused - possibly subject to minor modifications - to integrate differently structured sources.
An Improvement to a Multi-Client Searchable Encryption Scheme for Boolean Queries.
Jiang, Han; Li, Xue; Xu, Qiuliang
2016-12-01
The migration of e-health systems to the cloud computing brings huge benefits, as same as some security risks. Searchable Encryption(SE) is a cryptography encryption scheme that can protect the confidentiality of data and utilize the encrypted data at the same time. The SE scheme proposed by Cash et al. in Crypto2013 and its follow-up work in CCS2013 are most practical SE Scheme that support Boolean queries at present. In their scheme, the data user has to generate the search tokens by the counter number one by one and interact with server repeatedly, until he meets the correct one, or goes through plenty of tokens to illustrate that there is no search result. In this paper, we make an improvement to their scheme. We allow server to send back some information and help the user to generate exact search token in the search phase. In our scheme, there are only two round interaction between server and user, and the search token has [Formula: see text] elements, where n is the keywords number in query expression, and [Formula: see text] is the minimum documents number that contains one of keyword in query expression, and the computation cost of server is [Formula: see text] modular exponentiation operation.
NASA Technical Reports Server (NTRS)
Nguyen, Hung D.; Steele, Gynelle C.
2016-01-01
This report outlines the 2015 Small Business Innovation Research/Small Business Technology Transfer (SBIR/STTR) Phase I, Phase II, and Post-Phase II opportunity contract award results associated with NASA's Aeronautics Research Mission Directorate (ARMD), Human Exploration and Operations Mission Directorate (HEOMD), Science Mission Directorate (SMD), and Space Technology Mission Directorate (STMD) for NASA Glenn Research Center. The report also highlights the number of Phase I, Phase II, and Post-Phase II contracts awarded by mission directorate. The 2015 Phase I contract awards to companies in Ohio and their corresponding technologies are also discussed.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Frazier, Christopher Rawls; Durfee, Justin David; Bandlow, Alisa
The Contingency Contractor Optimization Tool – Prototype (CCOT-P) database is used to store input and output data for the linear program model described in [1]. The database allows queries to retrieve this data and updating and inserting new input data.
An infrastructure for ontology-based information systems in biomedicine: RICORDO case study.
Wimalaratne, Sarala M; Grenon, Pierre; Hoehndorf, Robert; Gkoutos, Georgios V; de Bono, Bernard
2012-02-01
The article presents an infrastructure for supporting the semantic interoperability of biomedical resources based on the management (storing and inference-based querying) of their ontology-based annotations. This infrastructure consists of: (i) a repository to store and query ontology-based annotations; (ii) a knowledge base server with an inference engine to support the storage of and reasoning over ontologies used in the annotation of resources; (iii) a set of applications and services allowing interaction with the integrated repository and knowledge base. The infrastructure is being prototyped and developed and evaluated by the RICORDO project in support of the knowledge management of biomedical resources, including physiology and pharmacology models and associated clinical data. The RICORDO toolkit and its source code are freely available from http://ricordo.eu/relevant-resources. sarala@ebi.ac.uk.
ERIC Educational Resources Information Center
Fugmann, Robert
1973-01-01
The personal literature search of a scientist is strongly governed by subjective influences. If he delegates his query, he must be aware of the problems emanating from this subjectivity. The information scientist, on the other hand, must consider the subjective character of the fundamental notions of information and order. (33 references)…
Chan, John K; Ueda, Stefanie M; Sugiyama, Valerie E; Stave, Christopher D; Shin, Jacob Y; Monk, Bradley J; Sikic, Branimir I; Osann, Kathryn; Kapp, Daniel S
2008-03-20
To identify the characteristics of phase II studies that predict for subsequent "positive" phase III trials (those that reached the proposed primary end points of study or those wherein the study drug was superior to the standard regimen investigating targeted agents in advanced tumors. We identified all phase III clinical trials of targeted therapies against advanced cancers published from 1985 to 2005. Characteristics of the preceding phase II studies were reviewed to identify predictive factors for success of the subsequent phase III trial. Data were analyzed using the chi(2) test and logistic regression models. Of 351 phase II studies, 167 (47.6%) subsequent phase III trials were positive and 184 (52.4%) negative. Phase II studies from multiple rather than single institutions were more likely to precede a successful trial (60.4% v 39.4%; P < .001). Positive phase II results were more likely to lead to a successful phase III trial (50.8% v 22.5%; P = .003). The percentage of successful trials from pharmaceutical companies was significantly higher compared with academic, cooperative groups, and research institutes (89.5% v 44.2%, 45.2%, and 46.3%, respectively; P = .002). On multivariate analysis, these factors and shorter time interval between publication of phase II results and III study publication were independent predictive factors for a positive phase III trial. In phase II studies of targeted agents, multiple- versus single-institution participation, positive phase II trial, pharmaceutical company-based trials, and shorter time period between publication of phase II to phase III trial were independent predictive factors of success in a phase III trial. Investigators should be cognizant of these factors in phase II studies before designing phase III trials.
A unified framework for managing provenance information in translational research
2011-01-01
Background A critical aspect of the NIH Translational Research roadmap, which seeks to accelerate the delivery of "bench-side" discoveries to patient's "bedside," is the management of the provenance metadata that keeps track of the origin and history of data resources as they traverse the path from the bench to the bedside and back. A comprehensive provenance framework is essential for researchers to verify the quality of data, reproduce scientific results published in peer-reviewed literature, validate scientific process, and associate trust value with data and results. Traditional approaches to provenance management have focused on only partial sections of the translational research life cycle and they do not incorporate "domain semantics", which is essential to support domain-specific querying and analysis by scientists. Results We identify a common set of challenges in managing provenance information across the pre-publication and post-publication phases of data in the translational research lifecycle. We define the semantic provenance framework (SPF), underpinned by the Provenir upper-level provenance ontology, to address these challenges in the four stages of provenance metadata: (a) Provenance collection - during data generation (b) Provenance representation - to support interoperability, reasoning, and incorporate domain semantics (c) Provenance storage and propagation - to allow efficient storage and seamless propagation of provenance as the data is transferred across applications (d) Provenance query - to support queries with increasing complexity over large data size and also support knowledge discovery applications We apply the SPF to two exemplar translational research projects, namely the Semantic Problem Solving Environment for Trypanosoma cruzi (T.cruzi SPSE) and the Biomedical Knowledge Repository (BKR) project, to demonstrate its effectiveness. Conclusions The SPF provides a unified framework to effectively manage provenance of translational research data during pre and post-publication phases. This framework is underpinned by an upper-level provenance ontology called Provenir that is extended to create domain-specific provenance ontologies to facilitate provenance interoperability, seamless propagation of provenance, automated querying, and analysis. PMID:22126369
Renin-Angiotensin System in Diabetes.
Rein, Johannes; Bader, Michael
2017-11-17
The renin-angiotensin system (RAS) has two different axes, the classical one with the effector peptide angiotensin II and the new one with the effector peptide angiotensin (1-7). Both peptides have been shown to be involved in the pathogenesis of diabetes mellitus and its consequences, nephropathy, retinopathy and cardiomyopathy in animal models and patients. In diabetes, angiotensin II acts mostly deleterious and angiotensin (1-7) protective. In this review we summarize the knowledge about the role of the different RAS axes in diabetes mellitus and the use of drugs interfering with the RAS in the therapy of the disease. Copyright© Bentham Science Publishers; For any queries, please email at epub@benthamscience.org.
NASA Technical Reports Server (NTRS)
1990-01-01
Synchrony, developed by St. Jude Medical's Cardiac Rhythm Management Division (formerly known as Pacesetter Systems, Inc.) is an advanced state-of-the-art implantable pacemaker that closely matches the natural rhythm of the heart. The companion element of the Synchrony Pacemaker System is the Programmer Analyzer APS-II which allows a doctor to reprogram and fine tune the pacemaker to each user's special requirements without surgery. The two-way communications capability that allows the physician to instruct and query the pacemaker is accomplished by bidirectional telemetry. APS-II features 28 pacing functions and thousands of programming combinations to accommodate diverse lifestyles. Microprocessor unit also records and stores pertinent patient data up to a year.
Seasonality in seeking mental health information on Google.
Ayers, John W; Althouse, Benjamin M; Allem, Jon-Patrick; Rosenquist, J Niels; Ford, Daniel E
2013-05-01
Population mental health surveillance is an important challenge limited by resource constraints, long time lags in data collection, and stigma. One promising approach to bridge similar gaps elsewhere has been the use of passively generated digital data. This article assesses the viability of aggregate Internet search queries for real-time monitoring of several mental health problems, specifically in regard to seasonal patterns of seeking out mental health information. All Google mental health queries were monitored in the U.S. and Australia from 2006 to 2010. Additionally, queries were subdivided among those including the terms ADHD (attention deficit-hyperactivity disorder); anxiety; bipolar; depression; anorexia or bulimia (eating disorders); OCD (obsessive-compulsive disorder); schizophrenia; and suicide. A wavelet phase analysis was used to isolate seasonal components in the trends, and based on this model, the mean search volume in winter was compared with that in summer, as performed in 2012. All mental health queries followed seasonal patterns with winter peaks and summer troughs amounting to a 14% (95% CI=11%, 16%) difference in volume for the U.S. and 11% (95% CI=7%, 15%) for Australia. These patterns also were evident for all specific subcategories of illness or problem. For instance, seasonal differences ranged from 7% (95% CI=5%, 10%) for anxiety (followed by OCD, bipolar, depression, suicide, ADHD, schizophrenia) to 37% (95% CI=31%, 44%) for eating disorder queries in the U.S. Several nonclinical motivators for query seasonality (such as media trends or academic interest) were explored and rejected. Information seeking on Google across all major mental illnesses and/or problems followed seasonal patterns similar to those found for seasonal affective disorder. These are the first data published on patterns of seasonality in information seeking encompassing all the major mental illnesses, notable also because they likely would have gone undetected using traditional surveillance. Copyright © 2013. Published by Elsevier Inc.
Query-based learning for aerospace applications.
Saad, E W; Choi, J J; Vian, J L; Wunsch, D C Ii
2003-01-01
Models of real-world applications often include a large number of parameters with a wide dynamic range, which contributes to the difficulties of neural network training. Creating the training data set for such applications becomes costly, if not impossible. In order to overcome the challenge, one can employ an active learning technique known as query-based learning (QBL) to add performance-critical data to the training set during the learning phase, thereby efficiently improving the overall learning/generalization. The performance-critical data can be obtained using an inverse mapping called network inversion (discrete network inversion and continuous network inversion) followed by oracle query. This paper investigates the use of both inversion techniques for QBL learning, and introduces an original heuristic to select the inversion target values for continuous network inversion method. Efficiency and generalization was further enhanced by employing node decoupled extended Kalman filter (NDEKF) training and a causality index (CI) as a means to reduce the input search dimensionality. The benefits of the overall QBL approach are experimentally demonstrated in two aerospace applications: a classification problem with large input space and a control distribution problem.
Ordered Backward XPath Axis Processing against XML Streams
NASA Astrophysics Data System (ADS)
Nizar M., Abdul; Kumar, P. Sreenivasa
Processing of backward XPath axes against XML streams is challenging for two reasons: (i) Data is not cached for future access. (ii) Query contains steps specifying navigation to the data that already passed by. While there are some attempts to process parent and ancestor axes, there are very few proposals to process ordered backward axes namely, preceding and preceding-sibling. For ordered backward axis processing, the algorithm, in addition to overcoming the limitations on data availability, has to take care of ordering constraints imposed by these axes. In this paper, we show how backward ordered axes can be effectively represented using forward constraints. We then discuss an algorithm for XML stream processing of XPath expressions containing ordered backward axes. The algorithm uses a layered cache structure to systematically accumulate query results. Our experiments show that the new algorithm gains remarkable speed up over the existing algorithm without compromising on bufferspace requirement.
Quality assurance for the query and distribution systems of the RCSB Protein Data Bank
Bluhm, Wolfgang F.; Beran, Bojan; Bi, Chunxiao; Dimitropoulos, Dimitris; Prlić, Andreas; Quinn, Gregory B.; Rose, Peter W.; Shah, Chaitali; Young, Jasmine; Yukich, Benjamin; Berman, Helen M.; Bourne, Philip E.
2011-01-01
The RCSB Protein Data Bank (RCSB PDB, www.pdb.org) is a key online resource for structural biology and related scientific disciplines. The website is used on average by 165 000 unique visitors per month, and more than 2000 other websites link to it. The amount and complexity of PDB data as well as the expectations on its usage are growing rapidly. Therefore, ensuring the reliability and robustness of the RCSB PDB query and distribution systems are crucially important and increasingly challenging. This article describes quality assurance for the RCSB PDB website at several distinct levels, including: (i) hardware redundancy and failover, (ii) testing protocols for weekly database updates, (iii) testing and release procedures for major software updates and (iv) miscellaneous monitoring and troubleshooting tools and practices. As such it provides suggestions for how other websites might be operated. Database URL: www.pdb.org PMID:21382834
NASA Technical Reports Server (NTRS)
Brown, David B.
1990-01-01
The results of research and development efforts are described for Task one, Phase two of a general project entitled The Development of a Program Analysis Environment for Ada. The scope of this task includes the design and development of a prototype system for testing Ada software modules at the unit level. The system is called Query Utility Environment for Software Testing of Ada (QUEST/Ada). The prototype for condition coverage provides a platform that implements expert system interaction with program testing. The expert system can modify data in the instrument source code in order to achieve coverage goals. Given this initial prototype, it is possible to evaluate the rule base in order to develop improved rules for test case generation. The goals of Phase two are the following: (1) to continue to develop and improve the current user interface to support the other goals of this research effort (i.e., those related to improved testing efficiency and increased code reliable); (2) to develop and empirically evaluate a succession of alternative rule bases for the test case generator such that the expert system achieves coverage in a more efficient manner; and (3) to extend the concepts of the current test environment to address the issues of Ada concurrency.
Visual perception-based criminal identification: a query-based approach
NASA Astrophysics Data System (ADS)
Singh, Avinash Kumar; Nandi, G. C.
2017-01-01
The visual perception of eyewitness plays a vital role in criminal identification scenario. It helps law enforcement authorities in searching particular criminal from their previous record. It has been reported that searching a criminal record manually requires too much time to get the accurate result. We have proposed a query-based approach which minimises the computational cost along with the reduction of search space. A symbolic database has been created to perform a stringent analysis on 150 public (Bollywood celebrities and Indian cricketers) and 90 local faces (our data-set). An expert knowledge has been captured to encapsulate every criminal's anatomical and facial attributes in the form of symbolic representation. A fast query-based searching strategy has been implemented using dynamic decision tree data structure which allows four levels of decomposition to fetch respective criminal records. Two types of case studies - viewed and forensic sketches have been considered to evaluate the strength of our proposed approach. We have derived 1200 views of the entire population by taking into consideration 80 participants as eyewitness. The system demonstrates an accuracy level of 98.6% for test case I and 97.8% for test case II. It has also been reported that experimental results reduce the search space up to 30 most relevant records.
The Protein Disease Database of human body fluids: II. Computer methods and data issues.
Lemkin, P F; Orr, G A; Goldstein, M P; Creed, G J; Myrick, J E; Merril, C R
1995-01-01
The Protein Disease Database (PDD) is a relational database of proteins and diseases. With this database it is possible to screen for quantitative protein abnormalities associated with disease states. These quantitative relationships use data drawn from the peer-reviewed biomedical literature. Assays may also include those observed in high-resolution electrophoretic gels that offer the potential to quantitate many proteins in a single test as well as data gathered by enzymatic or immunologic assays. We are using the Internet World Wide Web (WWW) and the Web browser paradigm as an access method for wide distribution and querying of the Protein Disease Database. The WWW hypertext transfer protocol and its Common Gateway Interface make it possible to build powerful graphical user interfaces that can support easy-to-use data retrieval using query specification forms or images. The details of these interactions are totally transparent to the users of these forms. Using a client-server SQL relational database, user query access, initial data entry and database maintenance are all performed over the Internet with a Web browser. We discuss the underlying design issues, mapping mechanisms and assumptions that we used in constructing the system, data entry, access to the database server, security, and synthesis of derived two-dimensional gel image maps and hypertext documents resulting from SQL database searches.
Holm, Liisa; Laakso, Laura M
2016-07-08
The Dali server (http://ekhidna2.biocenter.helsinki.fi/dali) is a network service for comparing protein structures in 3D. In favourable cases, comparing 3D structures may reveal biologically interesting similarities that are not detectable by comparing sequences. The Dali server has been running in various places for over 20 years and is used routinely by crystallographers on newly solved structures. The latest update of the server provides enhanced analytics for the study of sequence and structure conservation. The server performs three types of structure comparisons: (i) Protein Data Bank (PDB) search compares one query structure against those in the PDB and returns a list of similar structures; (ii) pairwise comparison compares one query structure against a list of structures specified by the user; and (iii) all against all structure comparison returns a structural similarity matrix, a dendrogram and a multidimensional scaling projection of a set of structures specified by the user. Structural superimpositions are visualized using the Java-free WebGL viewer PV. The structural alignment view is enhanced by sequence similarity searches against Uniprot. The combined structure-sequence alignment information is compressed to a stack of aligned sequence logos. In the stack, each structure is structurally aligned to the query protein and represented by a sequence logo. © The Author(s) 2016. Published by Oxford University Press on behalf of Nucleic Acids Research.
Using phase II data for the analysis of phase III studies: An application in rare diseases.
Wandel, Simon; Neuenschwander, Beat; Röver, Christian; Friede, Tim
2017-06-01
Clinical research and drug development in orphan diseases are challenging, since large-scale randomized studies are difficult to conduct. Formally synthesizing the evidence is therefore of great value, yet this is rarely done in the drug-approval process. Phase III designs that make better use of phase II data can facilitate drug development in orphan diseases. A Bayesian meta-analytic approach is used to inform the phase III study with phase II data. It is particularly attractive, since uncertainty of between-trial heterogeneity can be dealt with probabilistically, which is critical if the number of studies is small. Furthermore, it allows quantifying and discounting the phase II data through the predictive distribution relevant for phase III. A phase III design is proposed which uses the phase II data and considers approval based on a phase III interim analysis. The design is illustrated with a non-inferiority case study from a Food and Drug Administration approval in herpetic keratitis (an orphan disease). Design operating characteristics are compared to those of a traditional design, which ignores the phase II data. An analysis of the phase II data reveals good but insufficient evidence for non-inferiority, highlighting the need for a phase III study. For the phase III study supported by phase II data, the interim analysis is based on half of the patients. For this design, the meta-analytic interim results are conclusive and would justify approval. In contrast, based on the phase III data only, interim results are inconclusive and require further evidence. To accelerate drug development for orphan diseases, innovative study designs and appropriate methodology are needed. Taking advantage of randomized phase II data when analyzing phase III studies looks promising because the evidence from phase II supports informed decision-making. The implementation of the Bayesian design is straightforward with public software such as R.
IGF-1 as a Drug for Preterm Infants: A Step-Wise Clinical Development.
Hellstrom, Ann; Ley, David; Hallberg, Boubou; Lofqvist, Chatarina; Hansen-Pupp, Ingrid; Ramenghi, Luca A; Borg, Jan; Smith, Lois E H; Hard, Anna-Lena
2017-01-01
Insulin-like growth factor 1 (IGF-1) is a mitogenic hormone involved in many processes such as growth, metabolism, angiogenesis and differentiation. After very preterm birth, energy demands increase while maternal supplies of nutrients and other factors are lost and the infant may become dependent on parenteral nutrition for weeks. Low postnatal IGF-1 concentrations in preterm infants are associated with poor weight gain, retinopathy of prematurity (ROP) and other morbidities. We will describe the process by which we aim to develop supplementation with recombinant human (rh) IGF-1 and its binding protein rhIGFBP-3 as a possible therapy to promote growth and maturation and reduce morbidities in extremely preterm infants. In order to calculate a dose of IGF-1 tolerated by neonates, a pharmacokinetic study of transfusion with fresh frozen plasma was performed, which provided a relatively low dose of IGF-1, (on average 1.4 µg/kg), that increased serum IGF-1 to levels close to those observed in fetuses and preterm infants of similar GAs. Thereafter, a Phase I 3 hours IV infusion of rhIGF-1/rhIGFBP-3 was conducted in 5 infants, followed by a Phase II study with four sections (A-D). In the Phase II, sections A-D studies, time on infusion increased and younger gestational ages were included. IV infusion increased IGF-1 but with short half-life (0.5h) implying a need for continuous infusion. In order to obtain in utero levels of IGF-I, the dose was increased from 100 to 250 µg/kg/24 h and the infusion was prolonged from 3 weeks postnatal age until a postmenstrual age of 29 weeks and 6 days. The purpose has been to ensure high-quality research into the development of a new drug for preterm infants. We hope that our work will help to establish a new standard for the testing of medications for preterm infants. Copyright© Bentham Science Publishers; For any queries, please email at epub@benthamscience.org.
47 CFR 54.309 - Connect America Fund Phase II Public Interest Obligations.
Code of Federal Regulations, 2014 CFR
2014-10-01
... 47 Telecommunication 3 2014-10-01 2014-10-01 false Connect America Fund Phase II Public Interest Obligations. 54.309 Section 54.309 Telecommunication FEDERAL COMMUNICATIONS COMMISSION (CONTINUED) COMMON... Connect America Fund Phase II Public Interest Obligations. (a) A price cap carrier electing Phase II model...
NASA Technical Reports Server (NTRS)
Brown, David B.
1991-01-01
The results of research and development efforts of the first six months of Task 1, Phase 3 of the project are presented. The goals of Phase 3 are: (1) to further refine the rule base and complete the comparative rule base evaluation; (2) to implement and evaluate a concurrency testing prototype; (3) to convert the complete (unit-level and concurrency) testing prototype to a workstation environment; and (4) to provide a prototype development document to facilitate the transfer of research technology to a working environment. These goals were partially met and the results are summarized.
Isac Sc-Linac Phase-II Helium Refrigerator Commissioning and First Operational Experience at Triumf
NASA Astrophysics Data System (ADS)
Sekachev, I.; Kishi, D.; Laxdal, R. E.
2010-04-01
ISAC Phase-II is an upgrade of the radioactive isotope superconducting linear accelerator, SC-linac, at TRIUMF. The Phase-I section of the accelerator, medium-beta, is operational and is cooled with a 600 W helium refrigerator, commissioned in March 2005. An identical refrigerator is being used with the Phase-II segment of the accelerator; which is now under construction. The second refrigerator has been commissioned and tested with the Phase-I section of the linac and is used for Phase-II linac development, including new SC-cavity performance tests. The commissioning of the Phase-II refrigeration system and recent operational experience is presented.
Design of Phase II Non-inferiority Trials.
Jung, Sin-Ho
2017-09-01
With the development of inexpensive treatment regimens and less invasive surgical procedures, we are confronted with non-inferiority study objectives. A non-inferiority phase III trial requires a roughly four times larger sample size than that of a similar standard superiority trial. Because of the large required sample size, we often face feasibility issues to open a non-inferiority trial. Furthermore, due to lack of phase II non-inferiority trial design methods, we do not have an opportunity to investigate the efficacy of the experimental therapy through a phase II trial. As a result, we often fail to open a non-inferiority phase III trial and a large number of non-inferiority clinical questions still remain unanswered. In this paper, we want to develop some designs for non-inferiority randomized phase II trials with feasible sample sizes. At first, we review a design method for non-inferiority phase III trials. Subsequently, we propose three different designs for non-inferiority phase II trials that can be used under different settings. Each method is demonstrated with examples. Each of the proposed design methods is shown to require a reasonable sample size for non-inferiority phase II trials. The three different non-inferiority phase II trial designs are used under different settings, but require similar sample sizes that are typical for phase II trials.
Dual Nuclear/Fluorescence Imaging Potantial of Zinc(II) Phthalocyanine in MIA PaCa-2 Cell Line.
Lambrecht, Fatma Yurt; Ince, Mine; Er, Ozge; Ocakoglu, Kasim; Sarı, Fatma Aslıhan; Kayabasi, Cagla; Gunduz, Cumhur
2016-01-01
Pancreatic cancer is very common and difficult to diagnose in early stage. Imaging systems for diagnosing cancer have many disadvantages. However, combining different imaging modalities offers synergistic advantages. Optical imaging is the most multidirectional and widely used imaging modality in both clinical practice and research. In present study, Zinc(II) phthalocyanine [Zn(II)Pc] was synthesized, labeled with iodine- 131 and in vitro study was carried out. The intracellular uptake studies of radiolabeled Zn(II)Pc were performed in WI-38 [ATCC CCL-75™, tissue: human fibroblast lung] and MIA PaCa-2 [ATCC CRL-1420™, tissue: human epithelial pancreas carcinoma] cell lines. The intracellular uptake efficiency of radiolabeled Zn(II)Pc in MIA PaCa-2 cells was determined two times higher than WI-38 cells. Also, fluorescence imaging (FI) efficiency of synthesized Zn(II)Pc was investigated in MIA PaCa-2 cells and significant uptake was observed. Zn(II)Pc might be used as a new agent for dual fluorescence/nuclear imaging for pancreatic cancer. Copyright© Bentham Science Publishers; For any queries, please email at epub@benthamscience.org.
78 FR 76789 - Additional Connect America Fund Phase II Issues
Federal Register 2010, 2011, 2012, 2013, 2014
2013-12-19
... inspection and copying during normal business hours in the FCC Reference Information Center, Portals II, 445... Phase I to Phase II. 2. Timing of Phase II Support Disbursements. In the USF/ICC Transformation Order... language in paragraph 180 of the USF/ICC Transformation Order. We now seek to more fully develop the record...
48 CFR 1852.219-81 - Limitation on subcontracting-SBIR Phase II program.
Code of Federal Regulations, 2014 CFR
2014-10-01
... subcontracting-SBIR Phase II program. 1852.219-81 Section 1852.219-81 Federal Acquisition Regulations System... CLAUSES Texts of Provisions and Clauses 1852.219-81 Limitation on subcontracting—SBIR Phase II program. As prescribed in 1819.7302(b), insert the following clause: Limitation on Subcontracting—SBIR Phase II Program...
48 CFR 1852.219-81 - Limitation on subcontracting-SBIR Phase II program.
Code of Federal Regulations, 2013 CFR
2013-10-01
... subcontracting-SBIR Phase II program. 1852.219-81 Section 1852.219-81 Federal Acquisition Regulations System... CLAUSES Texts of Provisions and Clauses 1852.219-81 Limitation on subcontracting—SBIR Phase II program. As prescribed in 1819.7302(b), insert the following clause: Limitation on Subcontracting—SBIR Phase II Program...
48 CFR 1852.219-81 - Limitation on subcontracting-SBIR Phase II program.
Code of Federal Regulations, 2012 CFR
2012-10-01
... subcontracting-SBIR Phase II program. 1852.219-81 Section 1852.219-81 Federal Acquisition Regulations System... CLAUSES Texts of Provisions and Clauses 1852.219-81 Limitation on subcontracting—SBIR Phase II program. As prescribed in 1819.7302(b), insert the following clause: Limitation on Subcontracting—SBIR Phase II Program...
Halperin, Daniel M.; Lee, J. Jack; Dagohoy, Cecile Gonzales; Yao, James C.
2015-01-01
Purpose Despite a robust clinical trial enterprise and encouraging phase II results, the vast minority of oncologic drugs in development receive regulatory approval. In addition, clinicians occasionally make therapeutic decisions based on phase II data. Therefore, clinicians, investigators, and regulatory agencies require improved understanding of the implications of positive phase II studies. We hypothesized that prior probability of eventual drug approval was significantly different across GI cancers, with substantial ramifications for the predictive value of phase II studies. Methods We conducted a systematic search of phase II studies conducted between 1999 and 2004 and compared studies against US Food and Drug Administration and National Cancer Institute databases of approved indications for drugs tested in those studies. Results In all, 317 phase II trials were identified and followed for a median of 12.5 years. Following completion of phase III studies, eventual new drug application approval rates varied from 0% (zero of 45) in pancreatic adenocarcinoma to 34.8% (24 of 69) for colon adenocarcinoma. The proportion of drugs eventually approved was correlated with the disease under study (P < .001). The median type I error for all published trials was 0.05, and the median type II error was 0.1, with minimal variation. By using the observed median type I error for each disease, phase II studies have positive predictive values ranging from less than 1% to 90%, depending on primary site of the cancer. Conclusion Phase II trials in different GI malignancies have distinct prior probabilities of drug approval, yielding quantitatively and qualitatively different predictive values with similar statistical designs. Incorporation of prior probability into trial design may allow for more effective design and interpretation of phase II studies. PMID:26261263
Evaluation of Graph Pattern Matching Workloads in Graph Analysis Systems
DOE Office of Scientific and Technical Information (OSTI.GOV)
Hong, Seokyong; Lee, Sangkeun; Lim, Seung-Hwan
2016-01-01
Graph analysis has emerged as a powerful method for data scientists to represent, integrate, query, and explore heterogeneous data sources. As a result, graph data management and mining became a popular area of research, and led to the development of plethora of systems in recent years. Unfortunately, the number of emerging graph analysis systems and the wide range of applications, coupled with a lack of apples-to-apples comparisons, make it difficult to understand the trade-offs between different systems and the graph operations for which they are designed. A fair comparison of these systems is a challenging task for the following reasons:more » multiple data models, non-standardized serialization formats, various query interfaces to users, and diverse environments they operate in. To address these key challenges, in this paper we present a new benchmark suite by extending the Lehigh University Benchmark (LUBM) to cover the most common capabilities of various graph analysis systems. We provide the design process of the benchmark, which generalizes the workflow for data scientists to conduct the desired graph analysis on different graph analysis systems. Equipped with this extended benchmark suite, we present performance comparison for nine subgraph pattern retrieval operations over six graph analysis systems, namely NetworkX, Neo4j, Jena, Titan, GraphX, and uRiKA. Through the proposed benchmark suite, this study reveals both quantitative and qualitative findings in (1) implications in loading data into each system; (2) challenges in describing graph patterns for each query interface; and (3) different sensitivity of each system to query selectivity. We envision that this study will pave the road for: (i) data scientists to select the suitable graph analysis systems, and (ii) data management system designers to advance graph analysis systems.« less
Massett, Holly A; Mishkin, Grace; Rubinstein, Larry; Ivy, S Percy; Denicoff, Andrea; Godwin, Elizabeth; DiPiazza, Kate; Bolognese, Jennifer; Zwiebel, James A; Abrams, Jeffrey S
2016-11-15
Accruing patients in a timely manner represents a significant challenge to early phase cancer clinical trials. The NCI Cancer Therapy Evaluation Program analyzed 19 months of corrective action plans (CAP) received for slow-accruing phase I and II trials to identify slow accrual reasons, evaluate whether proposed corrective actions matched these reasons, and assess the CAP impact on trial accrual, duration, and likelihood of meeting primary scientific objectives. Of the 135 CAPs analyzed, 69 were for phase I trials and 66 for phase II trials. Primary reasons cited for slow accrual were safety/toxicity (phase I: 48%), design/protocol concerns (phase I: 42%, phase II: 33%), and eligibility criteria (phase I: 41%, phase II: 35%). The most commonly proposed corrective actions were adding institutions (phase I: 43%, phase II: 85%) and amending the trial to change eligibility or design (phase I: 55%, phase II: 44%). Only 40% of CAPs provided proposed corrective actions that matched the reasons given for slow accrual. Seventy percent of trials were closed to accrual at time of analysis (phase I = 48; phase II = 46). Of these, 67% of phase I and 70% of phase II trials met their primary objectives, but they were active three times longer than projected. Among closed trials, 24% had an accrual rate increase associated with a greater likelihood of meeting their primary scientific objectives. Ultimately, trials receiving CAPs saw improved accrual rates. Future trials may benefit from implementing CAPs early in trial life cycles, but it may be more beneficial to invest in earlier accrual planning. Clin Cancer Res; 22(22); 5408-16. ©2016 AACRSee related commentary by Mileham and Kim, p. 5397. ©2016 American Association for Cancer Research.
Element distinctness revisited
NASA Astrophysics Data System (ADS)
Portugal, Renato
2018-07-01
The element distinctness problem is the problem of determining whether the elements of a list are distinct, that is, if x=(x_1,\\ldots ,x_N) is a list with N elements, we ask whether the elements of x are distinct or not. The solution in a classical computer requires N queries because it uses sorting to check whether there are equal elements. In the quantum case, it is possible to solve the problem in O(N^{2/3}) queries. There is an extension which asks whether there are k colliding elements, known as element k-distinctness problem. This work obtains optimal values of two critical parameters of Ambainis' seminal quantum algorithm (SIAM J Comput 37(1):210-239, 2007). The first critical parameter is the number of repetitions of the algorithm's main block, which inverts the phase of the marked elements and calls a subroutine. The second parameter is the number of quantum walk steps interlaced by oracle queries. We show that, when the optimal values of the parameters are used, the algorithm's success probability is 1-O(N^{1/(k+1)}), quickly approaching 1. The specification of the exact running time and success probability is important in practical applications of this algorithm.
75 FR 13290 - Agency Information Collection Activities: Proposed Collection: Comment Request
Federal Register 2010, 2011, 2012, 2013, 2014
2010-03-19
... 60.11(a)(3) Adverse Action 480 2 960 45 720 25 18,000 60.11(c) Requests for Hearings by Entities 0 0 0 480 0 200 0 60.12(a)(1) & (2) Queries by Hospital 5,996 213 1,277,148 5 106,429 25 2,660,725 60.13(a)(1)(i) Disclosure to Hospitals \\2\\ 0 0 0 0 0 0 0 60.13(a)(1)(ii) Disclosure to Practitioners (Self...
MYCIN II: design and implementation of a therapy reference with complex content-based indexing.
Kim, D. K.; Fagan, L. M.; Jones, K. T.; Berrios, D. C.; Yu, V. L.
1998-01-01
We describe the construction of MYCIN II, a prototype system that provides for content-based markup and search of a forthcoming clinical therapeutics textbook, Antimicrobial Therapy and Vaccines. Existing commercial search technology for digital references utilizes generic tools such as textword-based searches with geographical or statistical refinements. We suggest that the drawbacks of such systems significantly restrict their use in everyday clinical practice. This is in spite of the fact that there is a great need for the information contained within these same references. The system we describe is intended to supplement keyword searching so that certain important questions can be asked easily and can be answered reliably (in terms of precision and recall). Our method attacks this problem in a restricted domain of knowledge-clinical infectious disease. For example, we would like to be able to answer the class of questions exemplified by the following query: "What antimicrobial agents can be used to treat endocarditis caused by Eikenella corrodens?" We have compiled and analyzed a list of such questions to develop a concept-based markup scheme. This scheme was then applied within an HTML markup to electronically "highlight" passages from three textbook chapters. We constructed a functioning web-based search interface. Our system also provides semi-automated querying of PubMed using our concept markup and the user's actions as a guide. PMID:9929205
MYCIN II: design and implementation of a therapy reference with complex content-based indexing.
Kim, D K; Fagan, L M; Jones, K T; Berrios, D C; Yu, V L
1998-01-01
We describe the construction of MYCIN II, a prototype system that provides for content-based markup and search of a forthcoming clinical therapeutics textbook, Antimicrobial Therapy and Vaccines. Existing commercial search technology for digital references utilizes generic tools such as textword-based searches with geographical or statistical refinements. We suggest that the drawbacks of such systems significantly restrict their use in everyday clinical practice. This is in spite of the fact that there is a great need for the information contained within these same references. The system we describe is intended to supplement keyword searching so that certain important questions can be asked easily and can be answered reliably (in terms of precision and recall). Our method attacks this problem in a restricted domain of knowledge-clinical infectious disease. For example, we would like to be able to answer the class of questions exemplified by the following query: "What antimicrobial agents can be used to treat endocarditis caused by Eikenella corrodens?" We have compiled and analyzed a list of such questions to develop a concept-based markup scheme. This scheme was then applied within an HTML markup to electronically "highlight" passages from three textbook chapters. We constructed a functioning web-based search interface. Our system also provides semi-automated querying of PubMed using our concept markup and the user's actions as a guide.
DOT National Transportation Integrated Search
2005-12-01
This volume provides an overview of the six studies that compose Phase II of the Enhanced Night Visibility project and the experimental plan for its third and final portion, Phase III. The Phase II studies evaluated up to 12 vision enhancement system...
Change of motion and localization of cholesterol molecule during L(alpha)-H(II) transition.
Hayakawa, E; Naganuma, M; Mukasa, K; Shimozawa, T; Araiso, T
1998-01-01
Formation of the inverted hexagonal (H(II)) phase from the lamellar (L(alpha)) phase of bovine brain-extracted phosphatidylcholine (BBPC) and phosphatidylethanolamine (BBPE) was investigated using 31P-NMR with or without cholesterol. When the ratio of BBPC to BBPE was 1:1, the H(II) formation was observed in the presence of 33 mol% cholesterol (i.e., BBPC:BBPE:cholesterol = 1:1:1) at 47 degrees C. The fraction of the H(II) phase in the BBPC/BBPE/cholesterol system could be controlled by the addition of dioleoylglycerol. The change of molecular motion of cholesterol affected by the H(II) formation was measured at various ratios of the L(alpha) to H(II) phase with the time-resolved fluorescence depolarization method, using dehydroergosterol as a fluorescent probe. It is observed that the motion of cholesterol became vigorous in the mixture state of the L(alpha) and the H(II) phases compared to that in the L(alpha) or the H(II) phase only. These facts show that cholesterol has the strong ability to induce the H(II) phase, probably by special molecular motion, which includes change of its location from the headgroup area to the acyl-chain area. PMID:9533700
Flight Test of the F/A-18 Active Aeroelastic Wing Airplane
NASA Technical Reports Server (NTRS)
Voracek, David
2007-01-01
A viewgraph presentation of flight tests performed on the F/A active aeroelastic wing airplane is shown. The topics include: 1) F/A-18 AAW Airplane; 2) F/A-18 AAW Control Surfaces; 3) Flight Test Background; 4) Roll Control Effectiveness Regions; 5) AAW Design Test Points; 6) AAW Phase I Test Maneuvers; 7) OBES Pitch Doublets; 8) OBES Roll Doublets; 9) AAW Aileron Flexibility; 10) Phase I - Lessons Learned; 11) Control Law Development and Verification & Validation Testing; 12) AAW Phase II RFCS Envelopes; 13) AAW 1-g Phase II Flight Test; 14) Region I - Subsonic 1-g Rolls; 15) Region I - Subsonic 1-g 360 Roll; 16) Region II - Supersonic 1-g Rolls; 17) Region II - Supersonic 1-g 360 Roll; 18) Region III - Subsonic 1-g Rolls; 19) Roll Axis HOS/LOS Comparison Region II - Supersonic (open-loop); 20) Roll Axis HOS/LOS Comparison Region II - Supersonic (closed-loop); 21) AAW Phase II Elevated-g Flight Test; 22) Region I - Subsonic 4-g RPO; and 23) Phase II - Lessons Learned
Any information, anywhere, anytime for the warfighter
NASA Astrophysics Data System (ADS)
Lazaroff, Mark B.; Sage, Philip A.
1997-06-01
The objective of the DARPA battlefield awareness data dissemination (BADD) program is to deliver battlefield awareness information to the warfighter -- anywhere, anytime. BADD is an advanced concept technology demonstration (ACTD) to support proof of concept technology demonstrations and experiments with a goal of introducing new technology to support the operational needs and acceptance of the warfighter. BADD's information management technology provides a 'smart' push of information to the users by providing information subscription services implemented via user- generated profiles. The system also provides services for warfighter pull or 'reach-back' of information via ad hoc query support. The high bandwidth delivery of informtion via the Global Broadcast System (GBS) satellites enables users to receive battlefield awareness information virtually anywhere. Very similar goals have been established for data warehousing technology -- that is, deliver the right information, to the right user, at the right time so that effective decisions can be made. In this paper, we examine the BADD Phase II architecture and underlying information management technoloyg in the context of data warehousing technology and a data warehouse reference architecture. In particular, we foucs on the BADD segment that PSR is building, the Interface to Information Sources (I2S).
Installation Restoration Program. Phase II--Confirmation/Quantification. Stage 1.
1985-03-01
four phases. Phase I, Initial Assessment/ Records Search, is designed to identify possible hazardous waste contami- nated sites and potential...7 71 -. - - IL’ -, 1% 33 AihlIII Is 33 n~iL t iiC UII! ii CL C LU 1-3, Phase II, Confirmation and Quantification, is designed to confirm the...additional monitoring data upon which design of mitigative actions are based. In Phase III, Technology Base Development, appropriate technology is selected and
An on-line image data base system: Managing image collections
Malchus B. Baker; Daniel P. Huebner; Peter F. Ffolliott
2000-01-01
Many researchers and land management personnel want photographic records of the phases of their studies or projects. Depending on the personnel and the type of project, a study can result in a few or hundreds of photographic images. A data base system allows users to query using various parameters, such as key words, dates, and project locations, and to view images...
1984-10-01
8 iii "i t-. Table of Contents (cont.) Section Title Page -APPENDIX A Acronyms, Definitions, Nomenclature and Units of Measure B Scope of Work, Task...Identification/Records Search Phase II - Problem Confirmation and Quantification Phase III - Technology Base Development Phase IV - Corrective Action Only...Problem Identification/Records Search Phase II - Problem Confirmation and Quantification Phase III - Technology Base Development Phase IV - Corrective
Calculation and application of activity discriminants in lead optimization.
Luo, Xincai; Krumrine, Jennifer R; Shenvi, Ashok B; Pierson, M Edward; Bernstein, Peter R
2010-11-01
We present a technique for computing activity discriminants of in vitro (pharmacological, DMPK, and safety) assays and the application to the prediction of in vitro activities of proposed synthetic targets during the lead optimization phase of drug discovery projects. This technique emulates how medicinal chemists perform SAR analysis and activity prediction. The activity discriminants that are functions of 6 commonly used medicinal chemistry descriptors can be interpreted easily by medicinal chemists. Further, visualization with Spotfire allows medicinal chemists to analyze how the query molecule is related to compounds tested previously, and to evaluate easily the relevance of the activity discriminants to the activities of the query molecule. Validation with all compounds synthesized and tested in AstraZeneca Wilmington since 2006 demonstrates that this approach is useful for prioritizing new synthetic targets for synthesis. Copyright © 2010 Elsevier Inc. All rights reserved.
Query Auto-Completion Based on Word2vec Semantic Similarity
NASA Astrophysics Data System (ADS)
Shao, Taihua; Chen, Honghui; Chen, Wanyu
2018-04-01
Query auto-completion (QAC) is the first step of information retrieval, which helps users formulate the entire query after inputting only a few prefixes. Regarding the models of QAC, the traditional method ignores the contribution from the semantic relevance between queries. However, similar queries always express extremely similar search intention. In this paper, we propose a hybrid model FS-QAC based on query semantic similarity as well as the query frequency. We choose word2vec method to measure the semantic similarity between intended queries and pre-submitted queries. By combining both features, our experiments show that FS-QAC model improves the performance when predicting the user’s query intention and helping formulate the right query. Our experimental results show that the optimal hybrid model contributes to a 7.54% improvement in terms of MRR against a state-of-the-art baseline using the public AOL query logs.
Unapparent Information Revelation: Text Mining for Counterterrorism
NASA Astrophysics Data System (ADS)
Srihari, Rohini K.
Unapparent information revelation (UIR) is a special case of text mining that focuses on detecting possible links between concepts across multiple text documents by generating an evidence trail explaining the connection. A traditional search involving, for example, two or more person names will attempt to find documents mentioning both these individuals. This research focuses on a different interpretation of such a query: what is the best evidence trail across documents that explains a connection between these individuals? For example, all may be good golfers. A generalization of this task involves query terms representing general concepts (e.g. indictment, foreign policy). Previous approaches to this problem have focused on graph mining involving hyperlinked documents, and link analysis exploiting named entities. A new robust framework is presented, based on (i) generating concept chain graphs, a hybrid content representation, (ii) performing graph matching to select candidate subgraphs, and (iii) subsequently using graphical models to validate hypotheses using ranked evidence trails. We adapt the DUC data set for cross-document summarization to evaluate evidence trails generated by this approach
EquiX-A Search and Query Language for XML.
ERIC Educational Resources Information Center
Cohen, Sara; Kanza, Yaron; Kogan, Yakov; Sagiv, Yehoshua; Nutt, Werner; Serebrenik, Alexander
2002-01-01
Describes EquiX, a search language for XML that combines querying with searching to query the data and the meta-data content of Web pages. Topics include search engines; a data model for XML documents; search query syntax; search query semantics; an algorithm for evaluating a query on a document; and indexing EquiX queries. (LRW)
Oral Sulforaphane increases Phase II antioxidant enzymes in the human upper airway
Riedl, Marc A.; Saxon, Andrew; Diaz-Sanchez, David
2009-01-01
Background Cellular oxidative stress is an important factor in asthma and is thought to be the principle mechanism by which oxidant pollutants such as ozone and particulates mediate their pro-inflammatory effects. Endogenous Phase II enzymes abrogate oxidative stress through the scavenging of reactive oxygen species and metabolism of reactive chemicals. Objective We conducted a placebo-controlled dose escalation trial to investigate the in vivo effects of sulforaphane, a naturally occurring potent inducer of Phase II enzymes, on the expression of glutathione-s-transferase M1 (GSTM1), glutathione-s-transferase P1 (GSTP1), NADPH quinone oxidoreductase (NQO1), and hemoxygenase-1 (HO-1) in the upper airway of human subjects. Methods Study subjects consumed oral sulforaphane doses contained in a standardized broccoli sprout homogenate (BSH). RNA expression for selected Phase II enzymes was measured in nasal lavage cells by RT-PCR before and after sulforaphane dosing. Results All subjects tolerated oral sulforaphane dosing without significant adverse events. Increased Phase II enzyme expression in nasal lavage cells occurred in a dose-dependent manner with maximal enzyme induction observed at the highest dose of 200 grams broccoli sprouts prepared as BSH. Significant increases were seen in all sentinel Phase II enzymes RNA expression compared to baseline. Phase II enzyme induction was not seen with ingestion of non-sulforaphane containing alfalfa sprouts. Conclusion Oral sulforaphane safely and effectively induces mucosal Phase II enzyme expression in the upper airway of human subjects. This study demonstrates the potential of antioxidant Phase II enzymes induction in the human airway as a strategy to reduce the inflammatory effects of oxidative stress. Clinical Implications This study demonstrates the potential of enhancement of Phase II enzyme expression as a novel therapeutic strategy for oxidant induced airway disease. Capsule Summary A placebo-controlled dose escalation trial demonstrated that naturally occurring sulforaphane from broccoli sprouts can induce a potent increase in antioxidant Phase II enzymes in airway cells. PMID:19028145
Centrifuge workers study. Phase II, completion report
DOE Office of Scientific and Technical Information (OSTI.GOV)
Wooten, H.D.
1994-09-01
Phase II of the Centrifuge Workers Study was a follow-up to the Phase I efforts. The Phase I results had indicated a higher risk than expected among centrifuge workers for developing bladder cancer when compared with the risk in the general population for developing this same type of cancer. However, no specific agent could be identified as the causative agent for these bladder cancers. As the Phase II Report states, Phase I had been limited to workers who had the greatest potential for exposure to substances used in the centrifuge process. Phase II was designed to expand the survey tomore » evaluate the health of all employees who had ever worked in Centrifuge Program Departments 1330-1339 but who had not been interviewed in Phase I. Employees in analytical laboratories and maintenance departments who provided support services for the Centrifuge Program were also included in Phase II. In December 1989, the Oak Ridge Associated Universities (ORAU), now known as Oak Ridge Institute for Science and Education (ORISE), was contracted to conduct a follow-up study (Phase II). Phase H of the Centrifuge Workers Study expanded the survey to include all former centrifuge workers who were not included in Phase I. ORISE was chosen because they had performed the Phase I tasks and summarized the corresponding survey data therefrom.« less
Spatial and symbolic queries for 3D image data
NASA Astrophysics Data System (ADS)
Benson, Daniel C.; Zick, Gregory L.
1992-04-01
We present a query system for an object-oriented biomedical imaging database containing 3-D anatomical structures and their corresponding 2-D images. The graphical interface facilitates the formation of spatial queries, nonspatial or symbolic queries, and combined spatial/symbolic queries. A query editor is used for the creation and manipulation of 3-D query objects as volumes, surfaces, lines, and points. Symbolic predicates are formulated through a combination of text fields and multiple choice selections. Query results, which may include images, image contents, composite objects, graphics, and alphanumeric data, are displayed in multiple views. Objects returned by the query may be selected directly within the views for further inspection or modification, or for use as query objects in subsequent queries. Our image database query system provides visual feedback and manipulation of spatial query objects, multiple views of volume data, and the ability to combine spatial and symbolic queries. The system allows for incremental enhancement of existing objects and the addition of new objects and spatial relationships. The query system is designed for databases containing symbolic and spatial data. This paper discuses its application to data acquired in biomedical 3- D image reconstruction, but it is applicable to other areas such as CAD/CAM, geographical information systems, and computer vision.
GenoQuery: a new querying module for functional annotation in a genomic warehouse
Lemoine, Frédéric; Labedan, Bernard; Froidevaux, Christine
2008-01-01
Motivation: We have to cope with both a deluge of new genome sequences and a huge amount of data produced by high-throughput approaches used to exploit these genomic features. Crossing and comparing such heterogeneous and disparate data will help improving functional annotation of genomes. This requires designing elaborate integration systems such as warehouses for storing and querying these data. Results: We have designed a relational genomic warehouse with an original multi-layer architecture made of a databases layer and an entities layer. We describe a new querying module, GenoQuery, which is based on this architecture. We use the entities layer to define mixed queries. These mixed queries allow searching for instances of biological entities and their properties in the different databases, without specifying in which database they should be found. Accordingly, we further introduce the central notion of alternative queries. Such queries have the same meaning as the original mixed queries, while exploiting complementarities yielded by the various integrated databases of the warehouse. We explain how GenoQuery computes all the alternative queries of a given mixed query. We illustrate how useful this querying module is by means of a thorough example. Availability: http://www.lri.fr/~lemoine/GenoQuery/ Contact: chris@lri.fr, lemoine@lri.fr PMID:18586731
SPARK: Adapting Keyword Query to Semantic Search
NASA Astrophysics Data System (ADS)
Zhou, Qi; Wang, Chong; Xiong, Miao; Wang, Haofen; Yu, Yong
Semantic search promises to provide more accurate result than present-day keyword search. However, progress with semantic search has been delayed due to the complexity of its query languages. In this paper, we explore a novel approach of adapting keywords to querying the semantic web: the approach automatically translates keyword queries into formal logic queries so that end users can use familiar keywords to perform semantic search. A prototype system named 'SPARK' has been implemented in light of this approach. Given a keyword query, SPARK outputs a ranked list of SPARQL queries as the translation result. The translation in SPARK consists of three major steps: term mapping, query graph construction and query ranking. Specifically, a probabilistic query ranking model is proposed to select the most likely SPARQL query. In the experiment, SPARK achieved an encouraging translation result.
1989-12-01
that can be easily understood. (9) Parallelism. Several system components may need to execute in parallel. For example, the processing of sensor data...knowledge base are not accessible for processing by the database. Also in the likely case that the expert system poses a series of related queries, the...hiharken nxpfilcs’Iog - Knowledge base for the automation of loCgistics rr-ovenet T’he Ii rectorY containing the strike aircraft replacement knowledge base
Griffon, N; Schuers, M; Dhombres, F; Merabti, T; Kerdelhué, G; Rollin, L; Darmoni, S J
2016-08-02
Despite international initiatives like Orphanet, it remains difficult to find up-to-date information about rare diseases. The aim of this study is to propose an exhaustive set of queries for PubMed based on terminological knowledge and to evaluate it versus the queries based on expertise provided by the most frequently used resource in Europe: Orphanet. Four rare disease terminologies (MeSH, OMIM, HPO and HRDO) were manually mapped to each other permitting the automatic creation of expended terminological queries for rare diseases. For 30 rare diseases, 30 citations retrieved by Orphanet expert query and/or query based on terminological knowledge were assessed for relevance by two independent reviewers unaware of the query's origin. An adjudication procedure was used to resolve any discrepancy. Precision, relative recall and F-measure were all computed. For each Orphanet rare disease (n = 8982), there was a corresponding terminological query, in contrast with only 2284 queries provided by Orphanet. Only 553 citations were evaluated due to queries with 0 or only a few hits. There were no significant differences between the Orpha query and terminological query in terms of precision, respectively 0.61 vs 0.52 (p = 0.13). Nevertheless, terminological queries retrieved more citations more often than Orpha queries (0.57 vs. 0.33; p = 0.01). Interestingly, Orpha queries seemed to retrieve older citations than terminological queries (p < 0.0001). The terminological queries proposed in this study are now currently available for all rare diseases. They may be a useful tool for both precision or recall oriented literature search.
SW#db: GPU-Accelerated Exact Sequence Similarity Database Search.
Korpar, Matija; Šošić, Martin; Blažeka, Dino; Šikić, Mile
2015-01-01
In recent years we have witnessed a growth in sequencing yield, the number of samples sequenced, and as a result-the growth of publicly maintained sequence databases. The increase of data present all around has put high requirements on protein similarity search algorithms with two ever-opposite goals: how to keep the running times acceptable while maintaining a high-enough level of sensitivity. The most time consuming step of similarity search are the local alignments between query and database sequences. This step is usually performed using exact local alignment algorithms such as Smith-Waterman. Due to its quadratic time complexity, alignments of a query to the whole database are usually too slow. Therefore, the majority of the protein similarity search methods prior to doing the exact local alignment apply heuristics to reduce the number of possible candidate sequences in the database. However, there is still a need for the alignment of a query sequence to a reduced database. In this paper we present the SW#db tool and a library for fast exact similarity search. Although its running times, as a standalone tool, are comparable to the running times of BLAST, it is primarily intended to be used for exact local alignment phase in which the database of sequences has already been reduced. It uses both GPU and CPU parallelization and was 4-5 times faster than SSEARCH, 6-25 times faster than CUDASW++ and more than 20 times faster than SSW at the time of writing, using multiple queries on Swiss-prot and Uniref90 databases.
An advanced web query interface for biological databases
Latendresse, Mario; Karp, Peter D.
2010-01-01
Although most web-based biological databases (DBs) offer some type of web-based form to allow users to author DB queries, these query forms are quite restricted in the complexity of DB queries that they can formulate. They can typically query only one DB, and can query only a single type of object at a time (e.g. genes) with no possible interaction between the objects—that is, in SQL parlance, no joins are allowed between DB objects. Writing precise queries against biological DBs is usually left to a programmer skillful enough in complex DB query languages like SQL. We present a web interface for building precise queries for biological DBs that can construct much more precise queries than most web-based query forms, yet that is user friendly enough to be used by biologists. It supports queries containing multiple conditions, and connecting multiple object types without using the join concept, which is unintuitive to biologists. This interactive web interface is called the Structured Advanced Query Page (SAQP). Users interactively build up a wide range of query constructs. Interactive documentation within the SAQP describes the schema of the queried DBs. The SAQP is based on BioVelo, a query language based on list comprehension. The SAQP is part of the Pathway Tools software and is available as part of several bioinformatics web sites powered by Pathway Tools, including the BioCyc.org site that contains more than 500 Pathway/Genome DBs. PMID:20624715
SPARQL Query Re-writing Using Partonomy Based Transformation Rules
NASA Astrophysics Data System (ADS)
Jain, Prateek; Yeh, Peter Z.; Verma, Kunal; Henson, Cory A.; Sheth, Amit P.
Often the information present in a spatial knowledge base is represented at a different level of granularity and abstraction than the query constraints. For querying ontology's containing spatial information, the precise relationships between spatial entities has to be specified in the basic graph pattern of SPARQL query which can result in long and complex queries. We present a novel approach to help users intuitively write SPARQL queries to query spatial data, rather than relying on knowledge of the ontology structure. Our framework re-writes queries, using transformation rules to exploit part-whole relations between geographical entities to address the mismatches between query constraints and knowledge base. Our experiments were performed on completely third party datasets and queries. Evaluations were performed on Geonames dataset using questions from National Geographic Bee serialized into SPARQL and British Administrative Geography Ontology using questions from a popular trivia website. These experiments demonstrate high precision in retrieval of results and ease in writing queries.
2006-06-01
SPARQL SPARQL Protocol and RDF Query Language SQL Structured Query Language SUMO Suggested Upper Merged Ontology SW... Query optimization algorithms are implemented in the Pellet reasoner in order to ensure querying a knowledge base is efficient . These algorithms...memory as a treelike structure in order for the data to be queried . XML Query (XQuery) is the standard language used when querying XML
Implementation of Quantum Private Queries Using Nuclear Magnetic Resonance
NASA Astrophysics Data System (ADS)
Wang, Chuan; Hao, Liang; Zhao, Lian-Jie
2011-08-01
We present a modified protocol for the realization of a quantum private query process on a classical database. Using one-qubit query and CNOT operation, the query process can be realized in a two-mode database. In the query process, the data privacy is preserved as the sender would not reveal any information about the database besides her query information, and the database provider cannot retain any information about the query. We implement the quantum private query protocol in a nuclear magnetic resonance system. The density matrix of the memory registers are constructed.
Monitoring Moving Queries inside a Safe Region
Al-Khalidi, Haidar; Taniar, David; Alamri, Sultan
2014-01-01
With mobile moving range queries, there is a need to recalculate the relevant surrounding objects of interest whenever the query moves. Therefore, monitoring the moving query is very costly. The safe region is one method that has been proposed to minimise the communication and computation cost of continuously monitoring a moving range query. Inside the safe region the set of objects of interest to the query do not change; thus there is no need to update the query while it is inside its safe region. However, when the query leaves its safe region the mobile device has to reevaluate the query, necessitating communication with the server. Knowing when and where the mobile device will leave a safe region is widely known as a difficult problem. To solve this problem, we propose a novel method to monitor the position of the query over time using a linear function based on the direction of the query obtained by periodic monitoring of its position. Periodic monitoring ensures that the query is aware of its location all the time. This method reduces the costs associated with communications in client-server architecture. Computational results show that our method is successful in handling moving query patterns. PMID:24696652
47 CFR 69.727 - Regulatory relief.
Code of Federal Regulations, 2010 CFR
2010-10-01
... customer. (b) Phase II relief. Upon satisfaction of the Phase II triggers specified in §§ 69.709(c) or 69... Pricing Flexibility § 69.727 Regulatory relief. (a) Phase I relief. Upon satisfaction of the Phase I... similarly situated customers; and (ii) The price cap LEC excludes all contract tariff offerings from price...
RDF-GL: A SPARQL-Based Graphical Query Language for RDF
NASA Astrophysics Data System (ADS)
Hogenboom, Frederik; Milea, Viorel; Frasincar, Flavius; Kaymak, Uzay
This chapter presents RDF-GL, a graphical query language (GQL) for RDF. The GQL is based on the textual query language SPARQL and mainly focuses on SPARQL SELECT queries. The advantage of a GQL over textual query languages is that complexity is hidden through the use of graphical symbols. RDF-GL is supported by a Java-based editor, SPARQLinG, which is presented as well. The editor does not only allow for RDF-GL query creation, but also converts RDF-GL queries to SPARQL queries and is able to subsequently execute these. Experiments show that using the GQL in combination with the editor makes RDF querying more accessible for end users.
47 CFR 90.769 - Construction and implementation of Phase II nationwide licenses.
Code of Federal Regulations, 2011 CFR
2011-10-01
... Use of Frequencies in the 220-222 MHz Band Policies Governing the Licensing and Use of Phase II Ea, Regional and Nationwide Systems § 90.769 Construction and implementation of Phase II nationwide licenses...
Cumulative query method for influenza surveillance using search engine data.
Seo, Dong-Woo; Jo, Min-Woo; Sohn, Chang Hwan; Shin, Soo-Yong; Lee, JaeHo; Yu, Maengsoo; Kim, Won Young; Lim, Kyoung Soo; Lee, Sang-Il
2014-12-16
Internet search queries have become an important data source in syndromic surveillance system. However, there is currently no syndromic surveillance system using Internet search query data in South Korea. The objective of this study was to examine correlations between our cumulative query method and national influenza surveillance data. Our study was based on the local search engine, Daum (approximately 25% market share), and influenza-like illness (ILI) data from the Korea Centers for Disease Control and Prevention. A quota sampling survey was conducted with 200 participants to obtain popular queries. We divided the study period into two sets: Set 1 (the 2009/10 epidemiological year for development set 1 and 2010/11 for validation set 1) and Set 2 (2010/11 for development Set 2 and 2011/12 for validation Set 2). Pearson's correlation coefficients were calculated between the Daum data and the ILI data for the development set. We selected the combined queries for which the correlation coefficients were .7 or higher and listed them in descending order. Then, we created a cumulative query method n representing the number of cumulative combined queries in descending order of the correlation coefficient. In validation set 1, 13 cumulative query methods were applied, and 8 had higher correlation coefficients (min=.916, max=.943) than that of the highest single combined query. Further, 11 of 13 cumulative query methods had an r value of ≥.7, but 4 of 13 combined queries had an r value of ≥.7. In validation set 2, 8 of 15 cumulative query methods showed higher correlation coefficients (min=.975, max=.987) than that of the highest single combined query. All 15 cumulative query methods had an r value of ≥.7, but 6 of 15 combined queries had an r value of ≥.7. Cumulative query method showed relatively higher correlation with national influenza surveillance data than combined queries in the development and validation set.
Barriers to participation in a phase II cardiac rehabilitation programme.
Mak, Y M W; Chan, W K; Yue, C S S
2005-12-01
To identify barriers to participation in a phase II cardiac rehabilitation programme and measures that may enhance participation. Prospective study. Regional hospital, Hong Kong. Cardiac patients recruited for a phase I cardiac rehabilitation programme from July 2002 to January 2003. Reasons for not participating in a phase II cardiac rehabilitation programme. Of the 193 patients recruited for a phase I cardiac rehabilitation programme, 152 (79%) patients, with a mean age of 70.3 years (standard deviation, 11.9 years), did not proceed to phase II programme. Eleven (7%) deaths occurred before commencement of phase II and 74 (49%) patients were considered physically unfit. Reasons for the latter included fractures, pain, or degenerative changes in the lower limbs (24%), and co-morbidities such as cerebrovascular accident (19%), chronic renal failure (11%), congestive heart failure (9%), and unstable angina (8%). Phase II rehabilitation was postponed until after completion of scheduled cardiac interventions in 13% of patients. Failure of physicians to arrange the pre-phase II exercise stress test as per protocol was reported in 7% of patients. Other reasons were reported: work or time conflicts (16%), non-compliance with cardiac treatment (5%), financial constraints (4%), self-exercise (3%), fear after exercise stress testing (3%), and patients returning to their original cardiologists for treatment (3%). A significant (79%) proportion of patients did not proceed to a phase II cardiac rehabilitation programme for a variety of reasons. These included physical unfitness, work or time conflicts, and need to attend scheduled cardiac interventions. Further studies are required to determine how to overcome obstacles to cardiac rehabilitation.
A Query Integrator and Manager for the Query Web
Brinkley, James F.; Detwiler, Landon T.
2012-01-01
We introduce two concepts: the Query Web as a layer of interconnected queries over the document web and the semantic web, and a Query Web Integrator and Manager (QI) that enables the Query Web to evolve. QI permits users to write, save and reuse queries over any web accessible source, including other queries saved in other installations of QI. The saved queries may be in any language (e.g. SPARQL, XQuery); the only condition for interconnection is that the queries return their results in some form of XML. This condition allows queries to chain off each other, and to be written in whatever language is appropriate for the task. We illustrate the potential use of QI for several biomedical use cases, including ontology view generation using a combination of graph-based and logical approaches, value set generation for clinical data management, image annotation using terminology obtained from an ontology web service, ontology-driven brain imaging data integration, small-scale clinical data integration, and wider-scale clinical data integration. Such use cases illustrate the current range of applications of QI and lead us to speculate about the potential evolution from smaller groups of interconnected queries into a larger query network that layers over the document and semantic web. The resulting Query Web could greatly aid researchers and others who now have to manually navigate through multiple information sources in order to answer specific questions. PMID:22531831
Conceptual search in electronic patient record.
Baud, R H; Lovis, C; Ruch, P; Rassinoux, A M
2001-01-01
Search by content in a large corpus of free texts in the medical domain is, today, only partially solved. The so-called GREP approach (Get Regular Expression and Print), based on highly efficient string matching techniques, is subject to inherent limitations, especially its inability to recognize domain specific knowledge. Such methods oblige the user to formulate his or her query in a logical Boolean style; if this constraint is not fulfilled, the results are poor. The authors present an enhancement to string matching search by the addition of a light conceptual model behind the word lexicon. The new system accepts any sentence as a query and radically improves the quality of results. Efficiency regarding execution time is obtained at the expense of implementing advanced indexing algorithms in a pre-processing phase. The method is described and commented and a brief account of the results illustrates this paper.
1991-03-01
1-2 1.4 CONCLUSIONS AND RECOMMENDATIONS ....................... 1-2 20. PHASE II MANAGEMENT PLAN...2-1 2.1 PROGRAM MANAGEMENT ................................... 2-1 2.2 IM IP TEAM...Barbier, reference Section 2.0 (Phase II Management Plan), is complete and this report provides the results of the Phase II study. 1.2 OBJECTIVES The
ERIC Educational Resources Information Center
Marzano Research Laboratory, 2010
2010-01-01
Phase II provides a more detailed examination of classroom variables important to achievement in Oklahoma schools. Where Phase I addressed all nine of the Oklahoma essential elements using survey data, Phase II focuses on what occurs in Oklahoma classrooms primarily using data from principal interviews, classroom observations (on-site), and video…
Using Generalized Annotated Programs to Solve Social Network Diffusion Optimization Problems
2013-01-01
as follows: —Let kall be the k value for the SNDOP-ALL query and for each SNDOP query i, let ki be the k for that query. For each query i, set ki... kall − 1. —Number each element of vi ∈ V such that gI(vi) and V C(vi) are true. For the ith SNDOP query, let vi be the corresponding element of V —Let...vertices of S. PROOF. We set up |V | SNDOP-queries as follows: —Let kall be the k value for the SNDOP-ALL query and and for each SNDOP-query i, let ki be
Alabama Coronary Artery Bypass Grafting Project
Holman, William L.; Sansom, Monique; Kiefe, Catarina I.; Peterson, Eric D.; Hubbard, Steve G.; Delong, James F.; Allman, Richard M.
2004-01-01
Objective/Background: This report describes the first round of results for Phase II of the Alabama CABG Project, a regional quality improvement initiative. Methods: Charts submitted by all hospitals in Alabama performing CABG (ICD-9 codes 36.10–36.20) were reviewed by a Clinical Data Abstraction Center (CDAC) (preintervention 1999–2000; postintervention 2000–2001). Variables that described quality in Phase I were abstracted for Phase II and data describing the new variables of β-blocker use and lipid management were collected. Data samples collected onsite by participating hospitals were used for rapid cycle improvement in Phase II. Results: CDAC data (n = 1927 cases in 1999; n = 2001 cases in 2000) showed that improvements from Phase I in aspirin prescription, internal mammary artery use, and duration of intubation persisted in Phase II. During Phase II, use of β-blockers before, during, or after CABG increased from 65% to 76% of patients (P < 0.05). Appropriate lipid management, an aggregate variable, occurred in 91% of patients before and 91% after the educational intervention. However, there were improvements in 3 of 5 subcategories for lipid management (documenting a lipid disorder [52%–57%], initiating drug therapy [45%–53%], and dietary counseling [74%–91%]; P < 0.05). Conclusions: In Phase II, this statewide process-oriented quality improvement program added two new measures of quality. Achievements of quality improvement from Phase I persisted in Phase II, and improvements were seen in the new variables of lipid management and perioperative use of β-blockers. PMID:14685107
Electric Utility Phase I Acid Rain Compliance Strategies for the Clean Air Act Amendments of 1990
1994-01-01
The Acid Rain Program is divided into two time periods; Phase I, from 1995 through 1999, and Phase II, starting in 2000. Phase I mostly affects power plants that are the largest sources of SO2 and NOx . Phase II affects virtually all electric power producers, including utilities and nonutilities. This report is a study of the effects of compliance with Phase I regulations on the costs and operations of electric utilities, but does not address any Phase II impacts.
DOT National Transportation Integrated Search
1966-12-01
This report describes a laboratory research program on the durability of lightweight concrete. Two phases of a three phase study are covered by this report, while the remaining phase is still under study. The two phases being reported are Phase II - ...
Ion Conduction Path and Low-Temperature Form:. Argyrodite-Type Superionic Conductors
NASA Astrophysics Data System (ADS)
Onoda, M.; Wada, H.; Sato, A.; Ishii, M.
2007-01-01
The structures of the orthorhombic room-temperature phase of Cu8GeS6 (phase II) and the monoclinic low-temperature phase of Ag7TaS6 (phase II) have been successfully refined based on X-ray diffraction data from 12-fold twinned (Cu8GeS6 II) and 24-fold twinned (Ag7TaS6 II) crystals. Respectively among 6 major and 6 minor twin domains of Cu8GeS6 II, or among 12 major and 12 minor twin domains of Ag7TaS6 II, the argyrodite-type frameworks, GeS6 or TaS6, can be superposed to each other in principle, and only Cu-Cu or Ag-Ag network directions differ. At higher temperature, the crystals were considered to be 2-fold twinned crystals of superionic-conductor phase I with a space group F 43m. On cooling, each domain transforms into 6 domains of orthorhombic Cu8GeS6 II or 12 domains of monoclinic Ag7TaS6 II. Superposed projections along 6 directions of the structure of Cu8GeS6 II and along 12 directions of the structure of Ag7TaS6 II seem to show approximate expressions for Cu-ion and Ag-ion conduction paths in superionic-conductor phases, Cu8GeS6 I and Ag7TaS6I.
Tan, H
1977-01-01
Estimates of general combining ability of parents for yield and girth obtained separately from seedlings and their corresponding clonal families in Phases II and IIIA of the RRIM breeding programme are compared. A highly significant positive correlation (r = 0.71***) is found between GCA estimates from seedling and clonal families for yield in Phase IIIA, but not in Phase II (r = -0.03(NS)) nor for girth (r= -0.27(NS)) in Phase IIIA. The correlations for Phase II yield and Phase IIIA girth, however, improve when the GCA estimates based on small sample size or reversed rankings are excluded.When the best selections (based on present clonal and seedling information) are compared, all five of the parents top-ranking for yield are common in Phase IIIA but only two parents are common for yield and girth in Phases II and IIIA respectively. However, only one parent for yield in Phase II and two parents for girth in Phase IIIA would, if selected on clonal performance, have been omitted from the top ranking selections made by previous workers using seedling information.These findings, therefore, justify the choice of parents based on GCA estimates for yield obtained from seedling performance. Similar justification cannot be offered for girth, for which analysis is confounded by uninterpretable site and seasonal effects.
Jadhav, Ashutosh; Andrews, Donna; Fiksdal, Alexander; Kumbamu, Ashok; McCormick, Jennifer B; Misitano, Andrew; Nelsen, Laurie; Ryu, Euijung; Sheth, Amit; Wu, Stephen
2014-01-01
Background The number of people using the Internet and mobile/smart devices for health information seeking is increasing rapidly. Although the user experience for online health information seeking varies with the device used, for example, smart devices (SDs) like smartphones/tablets versus personal computers (PCs) like desktops/laptops, very few studies have investigated how online health information seeking behavior (OHISB) may differ by device. Objective The objective of this study is to examine differences in OHISB between PCs and SDs through a comparative analysis of large-scale health search queries submitted through Web search engines from both types of devices. Methods Using the Web analytics tool, IBM NetInsight OnDemand, and based on the type of devices used (PCs or SDs), we obtained the most frequent health search queries between June 2011 and May 2013 that were submitted on Web search engines and directed users to the Mayo Clinic’s consumer health information website. We performed analyses on “Queries with considering repetition counts (QwR)” and “Queries without considering repetition counts (QwoR)”. The dataset contains (1) 2.74 million and 3.94 million QwoR, respectively for PCs and SDs, and (2) more than 100 million QwR for both PCs and SDs. We analyzed structural properties of the queries (length of the search queries, usage of query operators and special characters in health queries), types of search queries (keyword-based, wh-questions, yes/no questions), categorization of the queries based on health categories and information mentioned in the queries (gender, age-groups, temporal references), misspellings in the health queries, and the linguistic structure of the health queries. Results Query strings used for health information searching via PCs and SDs differ by almost 50%. The most searched health categories are “Symptoms” (1 in 3 search queries), “Causes”, and “Treatments & Drugs”. The distribution of search queries for different health categories differs with the device used for the search. Health queries tend to be longer and more specific than general search queries. Health queries from SDs are longer and have slightly fewer spelling mistakes than those from PCs. Users specify words related to women and children more often than that of men and any other age group. Most of the health queries are formulated using keywords; the second-most common are wh- and yes/no questions. Users ask more health questions using SDs than PCs. Almost all health queries have at least one noun and health queries from SDs are more descriptive than those from PCs. Conclusions This study is a large-scale comparative analysis of health search queries to understand the effects of device type (PCs vs SDs) used on OHISB. The study indicates that the device used for online health information search plays an important role in shaping how health information searches by consumers and patients are executed. PMID:25000537
Jadhav, Ashutosh; Andrews, Donna; Fiksdal, Alexander; Kumbamu, Ashok; McCormick, Jennifer B; Misitano, Andrew; Nelsen, Laurie; Ryu, Euijung; Sheth, Amit; Wu, Stephen; Pathak, Jyotishman
2014-07-04
The number of people using the Internet and mobile/smart devices for health information seeking is increasing rapidly. Although the user experience for online health information seeking varies with the device used, for example, smart devices (SDs) like smartphones/tablets versus personal computers (PCs) like desktops/laptops, very few studies have investigated how online health information seeking behavior (OHISB) may differ by device. The objective of this study is to examine differences in OHISB between PCs and SDs through a comparative analysis of large-scale health search queries submitted through Web search engines from both types of devices. Using the Web analytics tool, IBM NetInsight OnDemand, and based on the type of devices used (PCs or SDs), we obtained the most frequent health search queries between June 2011 and May 2013 that were submitted on Web search engines and directed users to the Mayo Clinic's consumer health information website. We performed analyses on "Queries with considering repetition counts (QwR)" and "Queries without considering repetition counts (QwoR)". The dataset contains (1) 2.74 million and 3.94 million QwoR, respectively for PCs and SDs, and (2) more than 100 million QwR for both PCs and SDs. We analyzed structural properties of the queries (length of the search queries, usage of query operators and special characters in health queries), types of search queries (keyword-based, wh-questions, yes/no questions), categorization of the queries based on health categories and information mentioned in the queries (gender, age-groups, temporal references), misspellings in the health queries, and the linguistic structure of the health queries. Query strings used for health information searching via PCs and SDs differ by almost 50%. The most searched health categories are "Symptoms" (1 in 3 search queries), "Causes", and "Treatments & Drugs". The distribution of search queries for different health categories differs with the device used for the search. Health queries tend to be longer and more specific than general search queries. Health queries from SDs are longer and have slightly fewer spelling mistakes than those from PCs. Users specify words related to women and children more often than that of men and any other age group. Most of the health queries are formulated using keywords; the second-most common are wh- and yes/no questions. Users ask more health questions using SDs than PCs. Almost all health queries have at least one noun and health queries from SDs are more descriptive than those from PCs. This study is a large-scale comparative analysis of health search queries to understand the effects of device type (PCs vs. SDs) used on OHISB. The study indicates that the device used for online health information search plays an important role in shaping how health information searches by consumers and patients are executed.
Benzocaine polymorphism: pressure-temperature phase diagram involving forms II and III.
Gana, Inès; Barrio, Maria; Do, Bernard; Tamarit, Josep-Lluís; Céolin, René; Rietveld, Ivo B
2013-11-18
Understanding the phase behavior of an active pharmaceutical ingredient in a drug formulation is required to avoid the occurrence of sudden phase changes resulting in decrease of bioavailability in a marketed product. Benzocaine is known to possess three crystalline polymorphs, but their stability hierarchy has so far not been determined. A topological method and direct calorimetric measurements under pressure have been used to construct the topological pressure-temperature diagram of the phase relationships between the solid phases II and III, the liquid, and the vapor phase. In the process, the transition temperature between solid phases III and II and its enthalpy change have been determined. Solid phase II, which has the highest melting point, is the more stable phase under ambient conditions in this phase diagram. Surprisingly, solid phase I has not been observed during the study, even though the scarce literature data on its thermal behavior appear to indicate that it might be the most stable one of the three solid phases. Copyright © 2013 Elsevier B.V. All rights reserved.
ASR-9 processor augmentation card (9-PAC) phase II scan-scan correlator algorithms
DOT National Transportation Integrated Search
2001-04-26
The report documents the scan-scan correlator (tracker) algorithm developed for Phase II of the ASR-9 Processor Augmentation Card (9-PAC) project. The improved correlation and tracking algorithms in 9-PAC Phase II decrease the incidence of false-alar...
SkyQuery - A Prototype Distributed Query and Cross-Matching Web Service for the Virtual Observatory
NASA Astrophysics Data System (ADS)
Thakar, A. R.; Budavari, T.; Malik, T.; Szalay, A. S.; Fekete, G.; Nieto-Santisteban, M.; Haridas, V.; Gray, J.
2002-12-01
We have developed a prototype distributed query and cross-matching service for the VO community, called SkyQuery, which is implemented with hierarchichal Web Services. SkyQuery enables astronomers to run combined queries on existing distributed heterogeneous astronomy archives. SkyQuery provides a simple, user-friendly interface to run distributed queries over the federation of registered astronomical archives in the VO. The SkyQuery client connects to the portal Web Service, which farms the query out to the individual archives, which are also Web Services called SkyNodes. The cross-matching algorithm is run recursively on each SkyNode. Each archive is a relational DBMS with a HTM index for fast spatial lookups. The results of the distributed query are returned as an XML DataSet that is automatically rendered by the client. SkyQuery also returns the image cutout corresponding to the query result. SkyQuery finds not only matches between the various catalogs, but also dropouts - objects that exist in some of the catalogs but not in others. This is often as important as finding matches. We demonstrate the utility of SkyQuery with a brown-dwarf search between SDSS and 2MASS, and a search for radio-quiet quasars in SDSS, 2MASS and FIRST. The importance of a service like SkyQuery for the worldwide astronomical community cannot be overstated: data on the same objects in various archives is mapped in different wavelength ranges and looks very different due to different errors, instrument sensitivities and other peculiarities of each archive. Our cross-matching algorithm preforms a fuzzy spatial join across multiple catalogs. This type of cross-matching is currently often done by eye, one object at a time. A static cross-identification table for a set of archives would become obsolete by the time it was built - the exponential growth of astronomical data means that a dynamic cross-identification mechanism like SkyQuery is the only viable option. SkyQuery was funded by a grant from the NASA AISR program.
Teng, Rui; Leibnitz, Kenji; Miura, Ryu
2013-01-01
An essential application of wireless sensor networks is to successfully respond to user queries. Query packet losses occur in the query dissemination due to wireless communication problems such as interference, multipath fading, packet collisions, etc. The losses of query messages at sensor nodes result in the failure of sensor nodes reporting the requested data. Hence, the reliable and successful dissemination of query messages to sensor nodes is a non-trivial problem. The target of this paper is to enable highly successful query delivery to sensor nodes by localized and energy-efficient discovery, and recovery of query losses. We adopt local and collective cooperation among sensor nodes to increase the success rate of distributed discoveries and recoveries. To enable the scalability in the operations of discoveries and recoveries, we employ a distributed name resolution mechanism at each sensor node to allow sensor nodes to self-detect the correlated queries and query losses, and then efficiently locally respond to the query losses. We prove that the collective discovery of query losses has a high impact on the success of query dissemination and reveal that scalability can be achieved by using the proposed approach. We further study the novel features of the cooperation and competition in the collective recovery at PHY and MAC layers, and show that the appropriate number of detectors can achieve optimal successful recovery rate. We evaluate the proposed approach with both mathematical analyses and computer simulations. The proposed approach enables a high rate of successful delivery of query messages and it results in short route lengths to recover from query losses. The proposed approach is scalable and operates in a fully distributed manner. PMID:23748172
ERIC Educational Resources Information Center
Illinois Univ., Urbana. Coordinated Science Lab.
In contrast to conventional information storage and retrieval systems in which a body of knowledge is thought of as an indexed codex of documents to which access is obtained by an appropriately indexed query, this interdisciplinary study aims at an understanding of what is "knowledge" as distinct from a "data file," how this knowledge is acquired,…
Horizon scanning for new genomic tests.
Gwinn, Marta; Grossniklaus, Daurice A; Yu, Wei; Melillo, Stephanie; Wulf, Anja; Flome, Jennifer; Dotson, W David; Khoury, Muin J
2011-02-01
The development of health-related genomic tests is decentralized and dynamic, involving government, academic, and commercial entities. Consequently, it is not easy to determine which tests are in development, currently available, or discontinued. We developed and assessed the usefulness of a systematic approach to identifying new genomic tests on the Internet. We devised targeted queries of Web pages, newspaper articles, and blogs (Google Alerts) to identify new genomic tests. We finalized search and review procedures during a pilot phase that ended in March 2010. Queries continue to run daily and are compiled weekly; selected data are indexed in an online database, the Genomic Applications in Practice and Prevention Finder. After the pilot phase, our scan detected approximately two to three new genomic tests per week. Nearly two thirds of all tests (122/188, 65%) were related to cancer; only 6% were related to hereditary disorders. Although 88 (47%) of the tests, including 2 marketed directly to consumers, were commercially available, only 12 (6%) claimed United States Food and Drug Administration licensure. Systematic surveillance of the Internet provides information about genomic tests that can be used in combination with other resources to evaluate genomic tests. The Genomic Applications in Practice and Prevention Finder makes this information accessible to a wide group of stakeholders.
Faster quantum searching with almost any diffusion operator
NASA Astrophysics Data System (ADS)
Tulsi, Avatar
2015-05-01
Grover's search algorithm drives a quantum system from an initial state |s > to a desired final state |t > by using selective phase inversions of these two states. Earlier, we studied a generalization of Grover's algorithm that relaxes the assumption of the efficient implementation of Is, the selective phase inversion of the initial state, also known as a diffusion operator. This assumption is known to become a serious handicap in cases of physical interest. Our general search algorithm works with almost any diffusion operator Ds with the only restriction of having |s > as one of its eigenstates. The price that we pay for using any operator is an increase in the number of oracle queries by a factor of O (B ) , where B is a characteristic of the eigenspectrum of Ds and can be large in some situations. Here we show that by using a quantum Fourier transform, we can regain the optimal query complexity of Grover's algorithm without losing the freedom of using any diffusion operator for quantum searching. However, the total number of operators required by the algorithm is still O (B ) times more than that of Grover's algorithm. So our algorithm offers an advantage only if the oracle operator is computationally more expensive than the diffusion operator, which is true in most search problems.
Ontological Approach to Military Knowledge Modeling and Management
2004-03-01
federated search mechanism has to reformulate user queries (expressed using the ontology) in the query languages of the different sources (e.g. SQL...ontologies as a common terminology – Unified query to perform federated search • Query processing – Ontology mapping to sources reformulate queries
NASA Astrophysics Data System (ADS)
Li, C.; Zhu, X.; Guo, W.; Liu, Y.; Huang, H.
2015-05-01
A method suitable for indoor complex semantic query considering the computation of indoor spatial relations is provided According to the characteristics of indoor space. This paper designs ontology model describing the space related information of humans, events and Indoor space objects (e.g. Storey and Room) as well as their relations to meet the indoor semantic query. The ontology concepts are used in IndoorSPARQL query language which extends SPARQL syntax for representing and querying indoor space. And four types specific primitives for indoor query, "Adjacent", "Opposite", "Vertical" and "Contain", are defined as query functions in IndoorSPARQL used to support quantitative spatial computations. Also a method is proposed to analysis the query language. Finally this paper adopts this method to realize indoor semantic query on the study area through constructing the ontology model for the study building. The experimental results show that the method proposed in this paper can effectively support complex indoor space semantic query.
VISAGE: Interactive Visual Graph Querying.
Pienta, Robert; Navathe, Shamkant; Tamersoy, Acar; Tong, Hanghang; Endert, Alex; Chau, Duen Horng
2016-06-01
Extracting useful patterns from large network datasets has become a fundamental challenge in many domains. We present VISAGE, an interactive visual graph querying approach that empowers users to construct expressive queries, without writing complex code (e.g., finding money laundering rings of bankers and business owners). Our contributions are as follows: (1) we introduce graph autocomplete , an interactive approach that guides users to construct and refine queries, preventing over-specification; (2) VISAGE guides the construction of graph queries using a data-driven approach, enabling users to specify queries with varying levels of specificity, from concrete and detailed (e.g., query by example), to abstract (e.g., with "wildcard" nodes of any types), to purely structural matching; (3) a twelve-participant, within-subject user study demonstrates VISAGE's ease of use and the ability to construct graph queries significantly faster than using a conventional query language; (4) VISAGE works on real graphs with over 468K edges, achieving sub-second response times for common queries.
VISAGE: Interactive Visual Graph Querying
Pienta, Robert; Navathe, Shamkant; Tamersoy, Acar; Tong, Hanghang; Endert, Alex; Chau, Duen Horng
2017-01-01
Extracting useful patterns from large network datasets has become a fundamental challenge in many domains. We present VISAGE, an interactive visual graph querying approach that empowers users to construct expressive queries, without writing complex code (e.g., finding money laundering rings of bankers and business owners). Our contributions are as follows: (1) we introduce graph autocomplete, an interactive approach that guides users to construct and refine queries, preventing over-specification; (2) VISAGE guides the construction of graph queries using a data-driven approach, enabling users to specify queries with varying levels of specificity, from concrete and detailed (e.g., query by example), to abstract (e.g., with “wildcard” nodes of any types), to purely structural matching; (3) a twelve-participant, within-subject user study demonstrates VISAGE’s ease of use and the ability to construct graph queries significantly faster than using a conventional query language; (4) VISAGE works on real graphs with over 468K edges, achieving sub-second response times for common queries. PMID:28553670
A Visual Interface for Querying Heterogeneous Phylogenetic Databases.
Jamil, Hasan M
2017-01-01
Despite the recent growth in the number of phylogenetic databases, access to these wealth of resources remain largely tool or form-based interface driven. It is our thesis that the flexibility afforded by declarative query languages may offer the opportunity to access these repositories in a better way, and to use such a language to pose truly powerful queries in unprecedented ways. In this paper, we propose a substantially enhanced closed visual query language, called PhyQL, that can be used to query phylogenetic databases represented in a canonical form. The canonical representation presented helps capture most phylogenetic tree formats in a convenient way, and is used as the storage model for our PhyloBase database for which PhyQL serves as the query language. We have implemented a visual interface for the end users to pose PhyQL queries using visual icons, and drag and drop operations defined over them. Once a query is posed, the interface translates the visual query into a Datalog query for execution over the canonical database. Responses are returned as hyperlinks to phylogenies that can be viewed in several formats using the tree viewers supported by PhyloBase. Results cached in PhyQL buffer allows secondary querying on the computed results making it a truly powerful querying architecture.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Tourassi, Georgia D.; Harrawood, Brian; Singh, Swatee
2007-08-15
We have previously presented a knowledge-based computer-assisted detection (KB-CADe) system for the detection of mammographic masses. The system is designed to compare a query mammographic region with mammographic templates of known ground truth. The templates are stored in an adaptive knowledge database. Image similarity is assessed with information theoretic measures (e.g., mutual information) derived directly from the image histograms. A previous study suggested that the diagnostic performance of the system steadily improves as the knowledge database is initially enriched with more templates. However, as the database increases in size, an exhaustive comparison of the query case with each stored templatemore » becomes computationally burdensome. Furthermore, blind storing of new templates may result in redundancies that do not necessarily improve diagnostic performance. To address these concerns we investigated an entropy-based indexing scheme for improving the speed of analysis and for satisfying database storage restrictions without compromising the overall diagnostic performance of our KB-CADe system. The indexing scheme was evaluated on two different datasets as (i) a search mechanism to sort through the knowledge database, and (ii) a selection mechanism to build a smaller, concise knowledge database that is easier to maintain but still effective. There were two important findings in the study. First, entropy-based indexing is an effective strategy to identify fast a subset of templates that are most relevant to a given query. Only this subset could be analyzed in more detail using mutual information for optimized decision making regarding the query. Second, a selective entropy-based deposit strategy may be preferable where only high entropy cases are maintained in the knowledge database. Overall, the proposed entropy-based indexing scheme was shown to reduce the computational cost of our KB-CADe system by 55% to 80% while maintaining the system's diagnostic performance.« less
NASA Technical Reports Server (NTRS)
Sandroni, P.; Novak, V.; Opfer-Gehrking, T. L.; Huck, C. A.; Low, P. A.
2000-01-01
The postural tachycardia syndrome (POTS) is characterized clinically by orthostatic lightheadedness and tachycardia. When these patients perform a Valsalva maneuver, there is an excessive blood pressure increment after cessation of the maneuver (phase IV) that is sometimes associated with headaches. It is not known whether excessive phase IV is due to excessive peripheral vascular tone (an alpha-adrenergic mechanism) or is a manifestation of increased beta-adrenergic tone (hyperadrenergic state). The authors undertook a pharmacologic study evaluating the effect of intravenous phentolamine (alpha-adrenergic antagonist) and propranolol (beta-adrenergic antagonist) on the different phases of the Valsalva maneuver in a group of patients with POTS and age-matched normal control subjects. Patients with POTS had mean phases, when compared with controls, that were characterized by more negative II_E (p = 0.07), smaller II_L (p = 0.04), and significantly larger phase IV (p = 0.001). The effect of phentolamine was qualitatively and quantitatively different in POTS when compared with controls. Ten mg phentolamine in controls resulted in a significant accentuation of phase II_E (p = 0.001), attenuation of phase II_L (p = 0.002), and increase of phase IV (57.6 vs 30.7 mm Hg; p = 0.025). These changes resembled those of patients with POTS at baseline. In patients with POTS, the phase II abnormalities, already present, were further accentuated (p <0.001), and phase IV became smaller (50.6 vs 73.8 mm Hg; p = 0.09). Propranolol had no significant effect on phases II_E and II_L, but significantly reduced phase IV in both controls (p <0.05) and in patients with POTS (p <0.001) and improved the headache symptoms, when present, during and after phase IV. The authors conclude that phase IV is mainly under beta-adrenergic regulation and that the exaggerated phase IV in POTS is a result of a hyperadrenergic state.
Which factors predict the time spent answering queries to a drug information centre?
Reppe, Linda A.; Spigset, Olav
2010-01-01
Objective To develop a model based upon factors able to predict the time spent answering drug-related queries to Norwegian drug information centres (DICs). Setting and method Drug-related queries received at 5 DICs in Norway from March to May 2007 were randomly assigned to 20 employees until each of them had answered a minimum of five queries. The employees reported the number of drugs involved, the type of literature search performed, and whether the queries were considered judgmental or not, using a specifically developed scoring system. Main outcome measures The scores of these three factors were added together to define a workload score for each query. Workload and its individual factors were subsequently related to the measured time spent answering the queries by simple or multiple linear regression analyses. Results Ninety-six query/answer pairs were analyzed. Workload significantly predicted the time spent answering the queries (adjusted R2 = 0.22, P < 0.001). Literature search was the individual factor best predicting the time spent answering the queries (adjusted R2 = 0.17, P < 0.001), and this variable also contributed the most in the multiple regression analyses. Conclusion The most important workload factor predicting the time spent handling the queries in this study was the type of literature search that had to be performed. The categorisation of queries as judgmental or not, also affected the time spent answering the queries. The number of drugs involved did not significantly influence the time spent answering drug information queries. PMID:20922480
Hadoop-GIS: A High Performance Spatial Data Warehousing System over MapReduce.
Aji, Ablimit; Wang, Fusheng; Vo, Hoang; Lee, Rubao; Liu, Qiaoling; Zhang, Xiaodong; Saltz, Joel
2013-08-01
Support of high performance queries on large volumes of spatial data becomes increasingly important in many application domains, including geospatial problems in numerous fields, location based services, and emerging scientific applications that are increasingly data- and compute-intensive. The emergence of massive scale spatial data is due to the proliferation of cost effective and ubiquitous positioning technologies, development of high resolution imaging technologies, and contribution from a large number of community users. There are two major challenges for managing and querying massive spatial data to support spatial queries: the explosion of spatial data, and the high computational complexity of spatial queries. In this paper, we present Hadoop-GIS - a scalable and high performance spatial data warehousing system for running large scale spatial queries on Hadoop. Hadoop-GIS supports multiple types of spatial queries on MapReduce through spatial partitioning, customizable spatial query engine RESQUE, implicit parallel spatial query execution on MapReduce, and effective methods for amending query results through handling boundary objects. Hadoop-GIS utilizes global partition indexing and customizable on demand local spatial indexing to achieve efficient query processing. Hadoop-GIS is integrated into Hive to support declarative spatial queries with an integrated architecture. Our experiments have demonstrated the high efficiency of Hadoop-GIS on query response and high scalability to run on commodity clusters. Our comparative experiments have showed that performance of Hadoop-GIS is on par with parallel SDBMS and outperforms SDBMS for compute-intensive queries. Hadoop-GIS is available as a set of library for processing spatial queries, and as an integrated software package in Hive.
Hadoop-GIS: A High Performance Spatial Data Warehousing System over MapReduce
Aji, Ablimit; Wang, Fusheng; Vo, Hoang; Lee, Rubao; Liu, Qiaoling; Zhang, Xiaodong; Saltz, Joel
2013-01-01
Support of high performance queries on large volumes of spatial data becomes increasingly important in many application domains, including geospatial problems in numerous fields, location based services, and emerging scientific applications that are increasingly data- and compute-intensive. The emergence of massive scale spatial data is due to the proliferation of cost effective and ubiquitous positioning technologies, development of high resolution imaging technologies, and contribution from a large number of community users. There are two major challenges for managing and querying massive spatial data to support spatial queries: the explosion of spatial data, and the high computational complexity of spatial queries. In this paper, we present Hadoop-GIS – a scalable and high performance spatial data warehousing system for running large scale spatial queries on Hadoop. Hadoop-GIS supports multiple types of spatial queries on MapReduce through spatial partitioning, customizable spatial query engine RESQUE, implicit parallel spatial query execution on MapReduce, and effective methods for amending query results through handling boundary objects. Hadoop-GIS utilizes global partition indexing and customizable on demand local spatial indexing to achieve efficient query processing. Hadoop-GIS is integrated into Hive to support declarative spatial queries with an integrated architecture. Our experiments have demonstrated the high efficiency of Hadoop-GIS on query response and high scalability to run on commodity clusters. Our comparative experiments have showed that performance of Hadoop-GIS is on par with parallel SDBMS and outperforms SDBMS for compute-intensive queries. Hadoop-GIS is available as a set of library for processing spatial queries, and as an integrated software package in Hive. PMID:24187650
Woo, Hyekyung; Cho, Youngtae; Shim, Eunyoung; Lee, Jong-Koo; Lee, Chang-Gun; Kim, Seong Hwan
2016-07-04
As suggested as early as in 2006, logs of queries submitted to search engines seeking information could be a source for detection of emerging influenza epidemics if changes in the volume of search queries are monitored (infodemiology). However, selecting queries that are most likely to be associated with influenza epidemics is a particular challenge when it comes to generating better predictions. In this study, we describe a methodological extension for detecting influenza outbreaks using search query data; we provide a new approach for query selection through the exploration of contextual information gleaned from social media data. Additionally, we evaluate whether it is possible to use these queries for monitoring and predicting influenza epidemics in South Korea. Our study was based on freely available weekly influenza incidence data and query data originating from the search engine on the Korean website Daum between April 3, 2011 and April 5, 2014. To select queries related to influenza epidemics, several approaches were applied: (1) exploring influenza-related words in social media data, (2) identifying the chief concerns related to influenza, and (3) using Web query recommendations. Optimal feature selection by least absolute shrinkage and selection operator (Lasso) and support vector machine for regression (SVR) were used to construct a model predicting influenza epidemics. In total, 146 queries related to influenza were generated through our initial query selection approach. A considerable proportion of optimal features for final models were derived from queries with reference to the social media data. The SVR model performed well: the prediction values were highly correlated with the recent observed influenza-like illness (r=.956; P<.001) and virological incidence rate (r=.963; P<.001). These results demonstrate the feasibility of using search queries to enhance influenza surveillance in South Korea. In addition, an approach for query selection using social media data seems ideal for supporting influenza surveillance based on search query data.
Woo, Hyekyung; Shim, Eunyoung; Lee, Jong-Koo; Lee, Chang-Gun; Kim, Seong Hwan
2016-01-01
Background As suggested as early as in 2006, logs of queries submitted to search engines seeking information could be a source for detection of emerging influenza epidemics if changes in the volume of search queries are monitored (infodemiology). However, selecting queries that are most likely to be associated with influenza epidemics is a particular challenge when it comes to generating better predictions. Objective In this study, we describe a methodological extension for detecting influenza outbreaks using search query data; we provide a new approach for query selection through the exploration of contextual information gleaned from social media data. Additionally, we evaluate whether it is possible to use these queries for monitoring and predicting influenza epidemics in South Korea. Methods Our study was based on freely available weekly influenza incidence data and query data originating from the search engine on the Korean website Daum between April 3, 2011 and April 5, 2014. To select queries related to influenza epidemics, several approaches were applied: (1) exploring influenza-related words in social media data, (2) identifying the chief concerns related to influenza, and (3) using Web query recommendations. Optimal feature selection by least absolute shrinkage and selection operator (Lasso) and support vector machine for regression (SVR) were used to construct a model predicting influenza epidemics. Results In total, 146 queries related to influenza were generated through our initial query selection approach. A considerable proportion of optimal features for final models were derived from queries with reference to the social media data. The SVR model performed well: the prediction values were highly correlated with the recent observed influenza-like illness (r=.956; P<.001) and virological incidence rate (r=.963; P<.001). Conclusions These results demonstrate the feasibility of using search queries to enhance influenza surveillance in South Korea. In addition, an approach for query selection using social media data seems ideal for supporting influenza surveillance based on search query data. PMID:27377323
Schuers, Matthieu; Joulakian, Mher; Kerdelhué, Gaetan; Segas, Léa; Grosjean, Julien; Darmoni, Stéfan J; Griffon, Nicolas
2017-07-03
MEDLINE is the most widely used medical bibliographic database in the world. Most of its citations are in English and this can be an obstacle for some researchers to access the information the database contains. We created a multilingual query builder to facilitate access to the PubMed subset using a language other than English. The aim of our study was to assess the impact of this multilingual query builder on the quality of PubMed queries for non-native English speaking physicians and medical researchers. A randomised controlled study was conducted among French speaking general practice residents. We designed a multi-lingual query builder to facilitate information retrieval, based on available MeSH translations and providing users with both an interface and a controlled vocabulary in their own language. Participating residents were randomly allocated either the French or the English version of the query builder. They were asked to translate 12 short medical questions into MeSH queries. The main outcome was the quality of the query. Two librarians blind to the arm independently evaluated each query, using a modified published classification that differentiated eight types of errors. Twenty residents used the French version of the query builder and 22 used the English version. 492 queries were analysed. There were significantly more perfect queries in the French group vs. the English group (respectively 37.9% vs. 17.9%; p < 0.01). It took significantly more time for the members of the English group than the members of the French group to build each query, respectively 194 sec vs. 128 sec; p < 0.01. This multi-lingual query builder is an effective tool to improve the quality of PubMed queries in particular for researchers whose first language is not English.
TNX GeoSiphon Cell (TGSC-1) Phase II Single Cell Deployment/Demonstration Final Report
DOE Office of Scientific and Technical Information (OSTI.GOV)
Phifer, M.A.
1999-04-15
This Phase II final report documents the Phase II testing conducted from June 18, 1998 through November 13, 1998, and it focuses on the application of the siphon technology as a sub-component of the overall GeoSiphon Cell technology. [Q-TPL-T-00004
40 CFR 72.44 - Phase II repowering extensions.
Code of Federal Regulations, 2012 CFR
2012-07-01
... (CONTINUED) PERMITS REGULATION Acid Rain Compliance Plan and Compliance Options § 72.44 Phase II repowering... the requirements of paragraph (a)(1)(i) of this section may include in the unit's Phase II Acid Rain... authority shall issue the Acid Rain portion of the operating permit including: (A) The approved repowering...
40 CFR 72.44 - Phase II repowering extensions.
Code of Federal Regulations, 2011 CFR
2011-07-01
... (CONTINUED) PERMITS REGULATION Acid Rain Compliance Plan and Compliance Options § 72.44 Phase II repowering... the requirements of paragraph (a)(1)(i) of this section may include in the unit's Phase II Acid Rain... authority shall issue the Acid Rain portion of the operating permit including: (A) The approved repowering...
40 CFR 72.44 - Phase II repowering extensions.
Code of Federal Regulations, 2013 CFR
2013-07-01
... (CONTINUED) PERMITS REGULATION Acid Rain Compliance Plan and Compliance Options § 72.44 Phase II repowering... the requirements of paragraph (a)(1)(i) of this section may include in the unit's Phase II Acid Rain... authority shall issue the Acid Rain portion of the operating permit including: (A) The approved repowering...
40 CFR 72.44 - Phase II repowering extensions.
Code of Federal Regulations, 2010 CFR
2010-07-01
... (CONTINUED) PERMITS REGULATION Acid Rain Compliance Plan and Compliance Options § 72.44 Phase II repowering... the requirements of paragraph (a)(1)(i) of this section may include in the unit's Phase II Acid Rain... authority shall issue the Acid Rain portion of the operating permit including: (A) The approved repowering...
40 CFR 72.44 - Phase II repowering extensions.
Code of Federal Regulations, 2014 CFR
2014-07-01
... (CONTINUED) PERMITS REGULATION Acid Rain Compliance Plan and Compliance Options § 72.44 Phase II repowering... the requirements of paragraph (a)(1)(i) of this section may include in the unit's Phase II Acid Rain... authority shall issue the Acid Rain portion of the operating permit including: (A) The approved repowering...
First results of GERDA Phase II and consistency with background models
NASA Astrophysics Data System (ADS)
Agostini, M.; Allardt, M.; Bakalyarov, A. M.; Balata, M.; Barabanov, I.; Baudis, L.; Bauer, C.; Bellotti, E.; Belogurov, S.; Belyaev, S. T.; Benato, G.; Bettini, A.; Bezrukov, L.; Bode1, T.; Borowicz, D.; Brudanin, V.; Brugnera, R.; Caldwell, A.; Cattadori, C.; Chernogorov, A.; D'Andrea, V.; Demidova, E. V.; Di Marco, N.; Domula, A.; Doroshkevich, E.; Egorov, V.; Falkenstein, R.; Frodyma, N.; Gangapshev, A.; Garfagnini, A.; Gooch, C.; Grabmayr, P.; Gurentsov, V.; Gusev, K.; Hakenmüller, J.; Hegai, A.; Heisel, M.; Hemmer, S.; Hofmann, W.; Hult, M.; Inzhechik, L. V.; Janicskó Csáthy, J.; Jochum, J.; Junker, M.; Kazalov, V.; Kihm, T.; Kirpichnikov, I. V.; Kirsch, A.; Kish, A.; Klimenko, A.; Kneißl, R.; Knöpfle, K. T.; Kochetov, O.; Kornoukhov, V. N.; Kuzminov, V. V.; Laubenstein, M.; Lazzaro, A.; Lebedev, V. I.; Lehnert, B.; Liao, H. Y.; Lindner, M.; Lippi, I.; Lubashevskiy, A.; Lubsandorzhiev, B.; Lutter, G.; Macolino, C.; Majorovits, B.; Maneschg, W.; Medinaceli, E.; Miloradovic, M.; Mingazheva, R.; Misiaszek, M.; Moseev, P.; Nemchenok, I.; Palioselitis, D.; Panas, K.; Pandola, L.; Pelczar, K.; Pullia, A.; Riboldi, S.; Rumyantseva, N.; Sada, C.; Salamida, F.; Salathe, M.; Schmitt, C.; Schneider, B.; Schönert, S.; Schreiner, J.; Schulz, O.; Schütz, A.-K.; Schwingenheuer, B.; Selivanenko, O.; Shevzik, E.; Shirchenko, M.; Simgen, H.; Smolnikov, A.; Stanco, L.; Vanhoefer, L.; Vasenko, A. A.; Veresnikova, A.; von Sturm, K.; Wagner, V.; Wegmann, A.; Wester, T.; Wiesinger, C.; Wojcik, M.; Yanovich, E.; Zhitnikov, I.; Zhukov, S. V.; Zinatulina, D.; Zuber, K.; Zuzel, G.
2017-01-01
The GERDA (GERmanium Detector Array) is an experiment for the search of neutrinoless double beta decay (0νββ) in 76Ge, located at Laboratori Nazionali del Gran Sasso of INFN (Italy). GERDA operates bare high purity germanium detectors submersed in liquid Argon (LAr). Phase II of data-taking started in Dec 2015 and is currently ongoing. In Phase II 35 kg of germanium detectors enriched in 76Ge including thirty newly produced Broad Energy Germanium (BEGe) detectors is operating to reach an exposure of 100 kg·yr within about 3 years data taking. The design goal of Phase II is to reduce the background by one order of magnitude to get the sensitivity for T1/20ν = O≤ft( {{{10}26}} \\right){{ yr}}. To achieve the necessary background reduction, the setup was complemented with LAr veto. Analysis of the background spectrum of Phase II demonstrates consistency with the background models. Furthermore 226Ra and 232Th contamination levels consistent with screening results. In the first Phase II data release we found no hint for a 0νββ decay signal and place a limit of this process T1/20ν > 5.3 \\cdot {1025} yr (90% C.L., sensitivity 4.0·1025 yr). First results of GERDA Phase II will be presented.
Dizdar, Omer; Bilgin, Emre; Akin, Serkan; Kilickap, Saadettin; Hayran, Mutlu
2017-01-01
Complementary and alternative medicine (CAM) products are increasingly used because they are perceived as natural, relatively low-cost and probably effective therapies for various diseases including cancer. We aimed to determine the quantity and major characteristics of recent herbal/alternative medicine trials registered in clinicaltrials. gov in patients with cancer. "Cancer AND (herbal OR complementary OR alternative)" key words were used to query clinicaltrials. gov (access date 17 April 2015). From the results, 163 trials which have been conducted in patients with the diagnosis of cancer were identified and included in this analysis. At the date of access, 72 trials were completed, 37 trials were still recruiting patients and 10 trials had been withdrawn. Most common cancer type was breast cancer. Eighty-eight percent of trials were interventional and 60% of trials were randomized. The rate of new trial submission were similar for 5-year periods after 2000. The majority of the trials were conducted in United States of America (55%) and People's Republic of China (11%). Nine and 4 of 37 recruiting trials were recorded as phase II and phase III, respectively. When browsing was restricted to "recruiting" and "interventional" studies, the ratio of herbal/complementary treatment trials to all chemotherapy trials was 1.8 %. CAM research in patients with cancer is currently limited, both in terms of quantity and quality. Until high quality scientific and clinical research establishes safety and efficacy of CAM practices, physicians should rigorously inform patients and the public on potential risks and caveats associated with CAM practices.
Mining Longitudinal Web Queries: Trends and Patterns.
ERIC Educational Resources Information Center
Wang, Peiling; Berry, Michael W.; Yang, Yiheng
2003-01-01
Analyzed user queries submitted to an academic Web site during a four-year period, using a relational database, to examine users' query behavior, to identify problems they encounter, and to develop techniques for optimizing query analysis and mining. Linguistic analyses focus on query structures, lexicon, and word associations using statistical…
Optimizing a Query by Transformation and Expansion.
Glocker, Katrin; Knurr, Alexander; Dieter, Julia; Dominick, Friederike; Forche, Melanie; Koch, Christian; Pascoe Pérez, Analie; Roth, Benjamin; Ückert, Frank
2017-01-01
In the biomedical sector not only the amount of information produced and uploaded into the web is enormous, but also the number of sources where these data can be found. Clinicians and researchers spend huge amounts of time on trying to access this information and to filter the most important answers to a given question. As the formulation of these queries is crucial, automated query expansion is an effective tool to optimize a query and receive the best possible results. In this paper we introduce the concept of a workflow for an optimization of queries in the medical and biological sector by using a series of tools for expansion and transformation of the query. After the definition of attributes by the user, the query string is compared to previous queries in order to add semantic co-occurring terms to the query. Additionally, the query is enlarged by an inclusion of synonyms. The translation into database specific ontologies ensures the optimal query formulation for the chosen database(s). As this process can be performed in various databases at once, the results are ranked and normalized in order to achieve a comparable list of answers for a question.
WATCHMAN: A Data Warehouse Intelligent Cache Manager
NASA Technical Reports Server (NTRS)
Scheuermann, Peter; Shim, Junho; Vingralek, Radek
1996-01-01
Data warehouses store large volumes of data which are used frequently by decision support applications. Such applications involve complex queries. Query performance in such an environment is critical because decision support applications often require interactive query response time. Because data warehouses are updated infrequently, it becomes possible to improve query performance by caching sets retrieved by queries in addition to query execution plans. In this paper we report on the design of an intelligent cache manager for sets retrieved by queries called WATCHMAN, which is particularly well suited for data warehousing environment. Our cache manager employs two novel, complementary algorithms for cache replacement and for cache admission. WATCHMAN aims at minimizing query response time and its cache replacement policy swaps out entire retrieved sets of queries instead of individual pages. The cache replacement and admission algorithms make use of a profit metric, which considers for each retrieved set its average rate of reference, its size, and execution cost of the associated query. We report on a performance evaluation based on the TPC-D and Set Query benchmarks. These experiments show that WATCHMAN achieves a substantial performance improvement in a decision support environment when compared to a traditional LRU replacement algorithm.
Suzuki, Kazuyuki; Endo, Ryujin; Takikawa, Yasuhiro; Moriyasu, Fuminori; Aoyagi, Yutaka; Moriwaki, Hisataka; Terai, Shuji; Sakaida, Isao; Sakai, Yoshiyuki; Nishiguchi, Shuhei; Ishikawa, Toru; Takagi, Hitoshi; Naganuma, Atsushi; Genda, Takuya; Ichida, Takafumi; Takaguchi, Koichi; Miyazawa, Katsuhiko; Okita, Kiwamu
2018-05-01
The efficacy and safety of rifaximin in the treatment of hepatic encephalopathy (HE) are widely known, but they have not been confirmed in Japanese patients with HE. Thus, two prospective, randomized studies (a phase II/III study and a phase III study) were carried out. Subjects with grade I or II HE and hyperammonemia were enrolled. The phase II/III study, which was a randomized, evaluator-blinded, active-comparator, parallel-group study, was undertaken at 37 institutions in Japan. Treatment periods were 14 days. Eligible patients were randomized to the rifaximin group (1200 mg/day) or the lactitol group (18-36 g/day). The phase III study was carried out in the same patients previously enrolled in the phase II/III study, and they were all treated with rifaximin (1200 mg/day) for 10 weeks. In the phase II/III study, 172 patients were enrolled. Blood ammonia (B-NH 3 ) concentration was significantly improved in the rifaximin group, but the difference between the two groups was not significant. The portal systemic encephalopathy index (PSE index), including HE grade, was significantly improved in both groups. In the phase III study, 87.3% of enrolled patients completed the treatment. The improved B-NH 3 concentration and PSE index were well maintained from the phase II/III study during the treatment period of the phase III study. Adverse drug reactions (ADRs) were seen in 13.4% of patients who received rifaximin, but there were no severe ADRs leading to death. The efficacy of rifaximin is sufficient and treatment is well tolerated in Japanese patients with HE and hyperammonemia. © 2017 The Japan Society of Hepatology.
PubDNA Finder: a web database linking full-text articles to sequences of nucleic acids.
García-Remesal, Miguel; Cuevas, Alejandro; Pérez-Rey, David; Martín, Luis; Anguita, Alberto; de la Iglesia, Diana; de la Calle, Guillermo; Crespo, José; Maojo, Víctor
2010-11-01
PubDNA Finder is an online repository that we have created to link PubMed Central manuscripts to the sequences of nucleic acids appearing in them. It extends the search capabilities provided by PubMed Central by enabling researchers to perform advanced searches involving sequences of nucleic acids. This includes, among other features (i) searching for papers mentioning one or more specific sequences of nucleic acids and (ii) retrieving the genetic sequences appearing in different articles. These additional query capabilities are provided by a searchable index that we created by using the full text of the 176 672 papers available at PubMed Central at the time of writing and the sequences of nucleic acids appearing in them. To automatically extract the genetic sequences occurring in each paper, we used an original method we have developed. The database is updated monthly by automatically connecting to the PubMed Central FTP site to retrieve and index new manuscripts. Users can query the database via the web interface provided. PubDNA Finder can be freely accessed at http://servet.dia.fi.upm.es:8080/pubdnafinder
DMINDA: an integrated web server for DNA motif identification and analyses
Ma, Qin; Zhang, Hanyuan; Mao, Xizeng; Zhou, Chuan; Liu, Bingqiang; Chen, Xin; Xu, Ying
2014-01-01
DMINDA (DNA motif identification and analyses) is an integrated web server for DNA motif identification and analyses, which is accessible at http://csbl.bmb.uga.edu/DMINDA/. This web site is freely available to all users and there is no login requirement. This server provides a suite of cis-regulatory motif analysis functions on DNA sequences, which are important to elucidation of the mechanisms of transcriptional regulation: (i) de novo motif finding for a given set of promoter sequences along with statistical scores for the predicted motifs derived based on information extracted from a control set, (ii) scanning motif instances of a query motif in provided genomic sequences, (iii) motif comparison and clustering of identified motifs, and (iv) co-occurrence analyses of query motifs in given promoter sequences. The server is powered by a backend computer cluster with over 150 computing nodes, and is particularly useful for motif prediction and analyses in prokaryotic genomes. We believe that DMINDA, as a new and comprehensive web server for cis-regulatory motif finding and analyses, will benefit the genomic research community in general and prokaryotic genome researchers in particular. PMID:24753419
2013-01-01
Background Accurate and complete identification of mobile elements is a challenging task in the current era of sequencing, given their large numbers and frequent truncations. Group II intron retroelements, which consist of a ribozyme and an intron-encoded protein (IEP), are usually identified in bacterial genomes through their IEP; however, the RNA component that defines the intron boundaries is often difficult to identify because of a lack of strong sequence conservation corresponding to the RNA structure. Compounding the problem of boundary definition is the fact that a majority of group II intron copies in bacteria are truncated. Results Here we present a pipeline of 11 programs that collect and analyze group II intron sequences from GenBank. The pipeline begins with a BLAST search of GenBank using a set of representative group II IEPs as queries. Subsequent steps download the corresponding genomic sequences and flanks, filter out non-group II introns, assign introns to phylogenetic subclasses, filter out incomplete and/or non-functional introns, and assign IEP sequences and RNA boundaries to the full-length introns. In the final step, the redundancy in the data set is reduced by grouping introns into sets of ≥95% identity, with one example sequence chosen to be the representative. Conclusions These programs should be useful for comprehensive identification of group II introns in sequence databases as data continue to rapidly accumulate. PMID:24359548
A web-based data visualization tool for the MIMIC-II database.
Lee, Joon; Ribey, Evan; Wallace, James R
2016-02-04
Although MIMIC-II, a public intensive care database, has been recognized as an invaluable resource for many medical researchers worldwide, becoming a proficient MIMIC-II researcher requires knowledge of SQL programming and an understanding of the MIMIC-II database schema. These are challenging requirements especially for health researchers and clinicians who may have limited computer proficiency. In order to overcome this challenge, our objective was to create an interactive, web-based MIMIC-II data visualization tool that first-time MIMIC-II users can easily use to explore the database. The tool offers two main features: Explore and Compare. The Explore feature enables the user to select a patient cohort within MIMIC-II and visualize the distributions of various administrative, demographic, and clinical variables within the selected cohort. The Compare feature enables the user to select two patient cohorts and visually compare them with respect to a variety of variables. The tool is also helpful to experienced MIMIC-II researchers who can use it to substantially accelerate the cumbersome and time-consuming steps of writing SQL queries and manually visualizing extracted data. Any interested researcher can use the MIMIC-II data visualization tool for free to quickly and conveniently conduct a preliminary investigation on MIMIC-II with a few mouse clicks. Researchers can also use the tool to learn the characteristics of the MIMIC-II patients. Since it is still impossible to conduct multivariable regression inside the tool, future work includes adding analytics capabilities. Also, the next version of the tool will aim to utilize MIMIC-III which contains more data.
Assisting Consumer Health Information Retrieval with Query Recommendations
Zeng, Qing T.; Crowell, Jonathan; Plovnick, Robert M.; Kim, Eunjung; Ngo, Long; Dibble, Emily
2006-01-01
Objective: Health information retrieval (HIR) on the Internet has become an important practice for millions of people, many of whom have problems forming effective queries. We have developed and evaluated a tool to assist people in health-related query formation. Design: We developed the Health Information Query Assistant (HIQuA) system. The system suggests alternative/additional query terms related to the user's initial query that can be used as building blocks to construct a better, more specific query. The recommended terms are selected according to their semantic distance from the original query, which is calculated on the basis of concept co-occurrences in medical literature and log data as well as semantic relations in medical vocabularies. Measurements: An evaluation of the HIQuA system was conducted and a total of 213 subjects participated in the study. The subjects were randomized into 2 groups. One group was given query recommendations and the other was not. Each subject performed HIR for both a predefined and a self-defined task. Results: The study showed that providing HIQuA recommendations resulted in statistically significantly higher rates of successful queries (odds ratio = 1.66, 95% confidence interval = 1.16–2.38), although no statistically significant impact on user satisfaction or the users' ability to accomplish the predefined retrieval task was found. Conclusion: Providing semantic-distance-based query recommendations can help consumers with query formation during HIR. PMID:16221944
Phenomenology of Polymorphism, III: p, TDiagram and Stability of Piracetam Polymorphs
NASA Astrophysics Data System (ADS)
Céolin, R.; Agafonov, V.; Louër, D.; Dzyabchenko, V. A.; Toscani, S.; Cense, J. M.
1996-02-01
The nootropic drug Piracetam is known to crystallize in three phases. In order to obtain their stability hierarchy from sublimation pressure inequalities, the drawing of a topologicalp,Tdiagram was attempted. For such a purpose and also for quality control, crystallographic and thermodynamic data were required. Powder X-ray diffractometry (XRD) and differential scanning calorimetry (DSC) were used. Molecular energy calculations were performed. Phase I melts at 426 K (ΔfusH(I) = +180 J·g-1). Phase II transforms into Phase I at 399 K (Δ(II→I)H= +24 J·g-1). Phase III transforms into phase I at 392 K (Δ(III→I)H= +28 J·g-1) or melts at 412 K (ΔfusH(III) = +210 J·g-1). Thep,Tdiagram shows that phase I is stable at higher temperature and phase II at lower temperature, like phase III, which is stable under high pressure. At room temperature, phase II is the more stable form, and phase I the less stable one. This agrees with the spontaneous I → II transformation observed at 298 K within a few hours, and with lattice energies, calculated previously. Molecular energy calculations and crystal structure comparison show how intermolecular hydrogen bonds and H-bonded dimers, in phases II and III, may stabilize conformations higher in energy than those of the isolated molecule and of phase I.
Solomon, Daniel H; Lu, Bing; Yu, Zhi; Corrigan, Cassandra; Harrold, Leslie R; Smolen, Josef S; Fraenkel, Liana; Katz, Jeffrey N; Losina, Elena
2018-01-05
We conducted a two-phase randomized controlled trial of a Learning Collaborative (LC) to facilitate implementation of treat to target (TTT) to manage rheumatoid arthritis (RA). We found substantial improvement in implementation of TTT in Phase I. Herein, we report on a second 9 months (Phase II) where we examined maintenance of response in Phase I and predictors of greater improvement in TTT adherence. We recruited 11 rheumatology sites and randomized them to either receive the LC during Phase I or to a wait-list control group that received the LC intervention during Phase II. The outcome was change in TTT implementation score (0 to 100, 100 is best) from pre- to post-intervention. TTT implementation score is defined as a percent of components documented in visit notes. Analyses examined: 1) the extent that the Phase I intervention teams sustained improvement in TTT; and, 2) predictors of TTT improvement. The analysis included 636 RA patients. At baseline, mean TTT implementation score was 11% in Phase I intervention sites and 13% in Phase II sites. After the intervention, TTT implementation score improved to 57% in the Phase I intervention sites and to 58% in the Phase II sites. Intervention sites from Phase I sustained the improvement during the Phase II (52%). Predictors of greater TTT improvement included only having rheumatologist providers at the site, academic affiliation of the site, fewer providers per site, and the rheumatologist provider being a trainee. Improvement in TTT remained relatively stable over a post-intervention period. This article is protected by copyright. All rights reserved. This article is protected by copyright. All rights reserved.
PAQ: Persistent Adaptive Query Middleware for Dynamic Environments
NASA Astrophysics Data System (ADS)
Rajamani, Vasanth; Julien, Christine; Payton, Jamie; Roman, Gruia-Catalin
Pervasive computing applications often entail continuous monitoring tasks, issuing persistent queries that return continuously updated views of the operational environment. We present PAQ, a middleware that supports applications' needs by approximating a persistent query as a sequence of one-time queries. PAQ introduces an integration strategy abstraction that allows composition of one-time query responses into streams representing sophisticated spatio-temporal phenomena of interest. A distinguishing feature of our middleware is the realization that the suitability of a persistent query's result is a function of the application's tolerance for accuracy weighed against the associated overhead costs. In PAQ, programmers can specify an inquiry strategy that dictates how information is gathered. Since network dynamics impact the suitability of a particular inquiry strategy, PAQ associates an introspection strategy with a persistent query, that evaluates the quality of the query's results. The result of introspection can trigger application-defined adaptation strategies that alter the nature of the query. PAQ's simple API makes developing adaptive querying systems easily realizable. We present the key abstractions, describe their implementations, and demonstrate the middleware's usefulness through application examples and evaluation.
NASA Astrophysics Data System (ADS)
Kuznetsov, Valentin; Riley, Daniel; Afaq, Anzar; Sekhri, Vijay; Guo, Yuyi; Lueking, Lee
2010-04-01
The CMS experiment has implemented a flexible and powerful system enabling users to find data within the CMS physics data catalog. The Dataset Bookkeeping Service (DBS) comprises a database and the services used to store and access metadata related to CMS physics data. To this, we have added a generalized query system in addition to the existing web and programmatic interfaces to the DBS. This query system is based on a query language that hides the complexity of the underlying database structure by discovering the join conditions between database tables. This provides a way of querying the system that is simple and straightforward for CMS data managers and physicists to use without requiring knowledge of the database tables or keys. The DBS Query Language uses the ANTLR tool to build the input query parser and tokenizer, followed by a query builder that uses a graph representation of the DBS schema to construct the SQL query sent to underlying database. We will describe the design of the query system, provide details of the language components and overview of how this component fits into the overall data discovery system architecture.
Spatial aggregation query in dynamic geosensor networks
NASA Astrophysics Data System (ADS)
Yi, Baolin; Feng, Dayang; Xiao, Shisong; Zhao, Erdun
2007-11-01
Wireless sensor networks have been widely used for civilian and military applications, such as environmental monitoring and vehicle tracking. In many of these applications, the researches mainly aim at building sensor network based systems to leverage the sensed data to applications. However, the existing works seldom exploited spatial aggregation query considering the dynamic characteristics of sensor networks. In this paper, we investigate how to process spatial aggregation query over dynamic geosensor networks where both the sink node and sensor nodes are mobile and propose several novel improvements on enabling techniques. The mobility of sensors makes the existing routing protocol based on information of fixed framework or the neighborhood infeasible. We present an improved location-based stateless implicit geographic forwarding (IGF) protocol for routing a query toward the area specified by query window, a diameter-based window aggregation query (DWAQ) algorithm for query propagation and data aggregation in the query window, finally considering the location changing of the sink node, we present two schemes to forward the result to the sink node. Simulation results show that the proposed algorithms can improve query latency and query accuracy.
Cellulose-lanthanum hydroxide nanocomposite as a selective marker for detection of toxic copper
2014-01-01
In this current report, a simple, reliable, and rapid method based on modifying the cellulose surface by doping it with different percentages of lanthanum hydroxide (i.e., 1% La(OH)3-cellulose (LC), 5% La(OH)3-cellulose (LC2), and 10% La(OH)3-cellulose (LC3)) was proposed as a selective marker for detection of copper (Cu(II)) in aqueous medium. Surface properties of the newly modified cellulose phases were confirmed by Fourier transform infrared spectroscopy, field emission scanning electron microscope, energy dispersive X-ray spectroscopy, X-ray diffraction, and X-ray photoelectron spectroscopic analysis. The effect of pH on the adsorption of modified cellulose phases for Cu(II) was evaluated, and LC3 was found to be the most selective for Cu(II) at pH 6.0. Other parameters, influencing the maximum uptake of Cu(II) on LC3, were also investigated for a deeper mechanistic understanding of the adsorption phenomena. Results showed that the adsorption capacity for Cu(II) was improved by 211% on the LC3 phase as compared to diethylaminoethyl cellulose phase after only 2 h contact time. Adsorption isotherm data established that the adsorption process nature was monolayer with a homogeneous adsorbent surface. Results displayed that the adsorption of Cu(II) onto the LC3 phase obeyed a pseudo-second-order kinetic model. Selectivity studies toward eight metal ions, i.e., Cd(II), Co(II), Cr(III), Cr(VI), Cu(II), Fe(III), Ni(II), and Zn(II), were further performed at the optimized pH value. Based on the selectivity study, it was found that Cu(II) is highly selective toward the LC3 phase. Moreover, the efficiency of the proposed method was supported by implementing it to real environmental water samples with adequate results. PMID:25258599
The report gives Phase II results of a combined experimental/theoretical study to define the mechanisms and kinetics of the formation of NOx and other combustion pollutants. Two experimental devices were used in Phase II. A special flat-flame burner with a controlled-temperature ...
Federal Register 2010, 2011, 2012, 2013, 2014
2011-01-20
... DEPARTMENT OF ENERGY Federal Energy Regulatory Commission [Docket No. ER11-2657-000] Milford Wind Corridor Phase II, LLC; Supplemental Notice That Initial Market-Based Rate Filing Includes Request for... proceeding Milford Wind Corridor Phase II, LLC's application for market-based rate authority, with an...
40 CFR 72.73 - State issuance of Phase II permits.
Code of Federal Regulations, 2010 CFR
2010-07-01
... (CONTINUED) PERMITS REGULATION Acid Rain Phase II Implementation § 72.73 State issuance of Phase II permits... permit program under part 70 of this chapter and that has a State Acid Rain program accepted by the Administrator under § 72.71 shall be responsible for administering and enforcing Acid Rain permits effective in...
40 CFR 72.73 - State issuance of Phase II permits.
Code of Federal Regulations, 2014 CFR
2014-07-01
... (CONTINUED) PERMITS REGULATION Acid Rain Phase II Implementation § 72.73 State issuance of Phase II permits... permit program under part 70 of this chapter and that has a State Acid Rain program accepted by the Administrator under § 72.71 shall be responsible for administering and enforcing Acid Rain permits effective in...
40 CFR 72.73 - State issuance of Phase II permits.
Code of Federal Regulations, 2012 CFR
2012-07-01
... (CONTINUED) PERMITS REGULATION Acid Rain Phase II Implementation § 72.73 State issuance of Phase II permits... permit program under part 70 of this chapter and that has a State Acid Rain program accepted by the Administrator under § 72.71 shall be responsible for administering and enforcing Acid Rain permits effective in...
40 CFR 72.73 - State issuance of Phase II permits.
Code of Federal Regulations, 2013 CFR
2013-07-01
... (CONTINUED) PERMITS REGULATION Acid Rain Phase II Implementation § 72.73 State issuance of Phase II permits... permit program under part 70 of this chapter and that has a State Acid Rain program accepted by the Administrator under § 72.71 shall be responsible for administering and enforcing Acid Rain permits effective in...
40 CFR 72.74 - Federal issuance of Phase II permits.
Code of Federal Regulations, 2010 CFR
2010-07-01
... PROGRAMS (CONTINUED) PERMITS REGULATION Acid Rain Phase II Implementation § 72.74 Federal issuance of Phase II permits. (a)(1) The Administrator will be responsible for administering and enforcing Acid Rain... and enforcing Acid Rain permits for such sources under § 72.73(a). (2) After and to the extent the...
40 CFR 72.74 - Federal issuance of Phase II permits.
Code of Federal Regulations, 2014 CFR
2014-07-01
... PROGRAMS (CONTINUED) PERMITS REGULATION Acid Rain Phase II Implementation § 72.74 Federal issuance of Phase II permits. (a)(1) The Administrator will be responsible for administering and enforcing Acid Rain... and enforcing Acid Rain permits for such sources under § 72.73(a). (2) After and to the extent the...
40 CFR 72.74 - Federal issuance of Phase II permits.
Code of Federal Regulations, 2012 CFR
2012-07-01
... PROGRAMS (CONTINUED) PERMITS REGULATION Acid Rain Phase II Implementation § 72.74 Federal issuance of Phase II permits. (a)(1) The Administrator will be responsible for administering and enforcing Acid Rain... and enforcing Acid Rain permits for such sources under § 72.73(a). (2) After and to the extent the...
40 CFR 72.74 - Federal issuance of Phase II permits.
Code of Federal Regulations, 2013 CFR
2013-07-01
... PROGRAMS (CONTINUED) PERMITS REGULATION Acid Rain Phase II Implementation § 72.74 Federal issuance of Phase II permits. (a)(1) The Administrator will be responsible for administering and enforcing Acid Rain... and enforcing Acid Rain permits for such sources under § 72.73(a). (2) After and to the extent the...
40 CFR 72.73 - State issuance of Phase II permits.
Code of Federal Regulations, 2011 CFR
2011-07-01
... (CONTINUED) PERMITS REGULATION Acid Rain Phase II Implementation § 72.73 State issuance of Phase II permits... permit program under part 70 of this chapter and that has a State Acid Rain program accepted by the Administrator under § 72.71 shall be responsible for administering and enforcing Acid Rain permits effective in...
40 CFR 72.74 - Federal issuance of Phase II permits.
Code of Federal Regulations, 2011 CFR
2011-07-01
... PROGRAMS (CONTINUED) PERMITS REGULATION Acid Rain Phase II Implementation § 72.74 Federal issuance of Phase II permits. (a)(1) The Administrator will be responsible for administering and enforcing Acid Rain... and enforcing Acid Rain permits for such sources under § 72.73(a). (2) After and to the extent the...
Doping-induced disappearance of ice II from water's phase diagram
NASA Astrophysics Data System (ADS)
Shephard, Jacob J.; Slater, Ben; Harvey, Peter; Hart, Martin; Bull, Craig L.; Bramwell, Steven T.; Salzmann, Christoph G.
2018-06-01
Water and the many phases of ice display a plethora of complex physical properties and phase relationships1-4 that are of paramount importance in a range of settings including processes in Earth's hydrosphere, the geology of icy moons, industry and even the evolution of life. Well-known examples include the unusual behaviour of supercooled water2, the emergent ferroelectric ordering in ice films4 and the fact that the `ordinary' ice Ih floats on water. We report the intriguing observation that ice II, one of the high-pressure phases of ice, disappears in a selective fashion from water's phase diagram following the addition of small amounts of ammonium fluoride. This finding exposes the strict topologically constrained nature of the ice II hydrogen-bond network, which is not found for the competing phases. In analogy to the behaviour of frustrated magnets5, the presence of the exceptional ice II is argued to have a wider impact on water's phase diagram, potentially explaining its general tendency to display anomalous behaviour. Furthermore, the impurity-induced disappearance of ice II raises the prospect that specific dopants may not only be able to suppress certain phases but also induce the formation of new phases of ice in future studies.
Hoogendam, Arjen; Stalenhoef, Anton FH; Robbé, Pieter F de Vries; Overbeke, A John PM
2008-01-01
Background The use of PubMed to answer daily medical care questions is limited because it is challenging to retrieve a small set of relevant articles and time is restricted. Knowing what aspects of queries are likely to retrieve relevant articles can increase the effectiveness of PubMed searches. The objectives of our study were to identify queries that are likely to retrieve relevant articles by relating PubMed search techniques and tools to the number of articles retrieved and the selection of articles for further reading. Methods This was a prospective observational study of queries regarding patient-related problems sent to PubMed by residents and internists in internal medicine working in an Academic Medical Centre. We analyzed queries, search results, query tools (Mesh, Limits, wildcards, operators), selection of abstract and full-text for further reading, using a portal that mimics PubMed. Results PubMed was used to solve 1121 patient-related problems, resulting in 3205 distinct queries. Abstracts were viewed in 999 (31%) of these queries, and in 126 (39%) of 321 queries using query tools. The average term count per query was 2.5. Abstracts were selected in more than 40% of queries using four or five terms, increasing to 63% if the use of four or five terms yielded 2–161 articles. Conclusion Queries sent to PubMed by physicians at our hospital during daily medical care contain fewer than three terms. Queries using four to five terms, retrieving less than 161 article titles, are most likely to result in abstract viewing. PubMed search tools are used infrequently by our population and are less effective than the use of four or five terms. Methods to facilitate the formulation of precise queries, using more relevant terms, should be the focus of education and research. PMID:18816391
Assessment of Operational Automated Guideway Systems - Airtrans (Phase II)
DOT National Transportation Integrated Search
1980-01-01
This study, Phase II, completes the assessment of AIRTRANS, the automated guideway system located at the Dallas-Fort Worth Airport. The Phase I assessment report: "Assessment of Operational Automated Guideway Systems--AIRTRANS (Phase I)" (PB-261 339)...
Courcier-Duplantier, Soizic; Bouhours, Philippe; Pinton, Philippe; Sibenaler, Claire; Lassale, Catherine
2004-01-01
In order to evaluate the attractiveness of France for conducting international clinical trials, a survey was performed among pharmaceutical companies that are based in France or that have affiliates in France. The survey concerned international phase II and III clinical studies carried out in 2002 and 2003. Ten pharmaceutical companies representing 36% of the French market completed the survey. 134 trials were analysed in total. France recruited 8.3% of the overall number of patients recruited, and 15.0% of those recruited within Europe. France was within the overall mean with regard to the percentage of active centres (78.5% versus 79.5%) and the percentage of patients evaluable according to protocol (86.8% versus 87.3%). In contrast, France ranked within the last third of analysed countries with respect to the speed of recruitment (1.5 versus 1.9 patients/centre/month), and the number of queries per observation (16.8 versus 10.9). The analysis of the qualitative indicators of performance showed that, although the perception of pharmaceutical companies towards the quality of French medicine and administrative authorities is positive, France notably needs to improve the productivity of its clinical research in order to enhance its attractiveness for the pharmaceutical sponsors of clinical trials.
McDonald, S A; Hutchinson, S J; Innes, H A; Allen, S; Bramley, P; Bhattacharyya, D; Carman, W; Dillon, J F; Fox, R; Fraser, A; Goldberg, D J; Kennedy, N; Mills, P R; Morris, J; Stanley, A J; Wilks, D; Hayes, P C
2014-05-01
Primary goals of the Hepatitis C Action Plan for Scotland Phase II (May 2008-March 2011) were to increase, among persons chronically infected with the hepatitis C (HCV) virus, attendance at specialist outpatient clinics and initiation on antiviral therapy. We evaluated progress towards these goals by comparing the odds, across time, of (a) first clinic attendance within 12 months of HCV diagnosis (n = 9747) and (b) initiation on antiviral treatment within 12 months of first attendance (n = 5736). Record linkage between the national HCV diagnosis (1996-2009) and HCV clinical (1996-2010) databases and logistic regression analyses were conducted for both outcomes. For outcome (a), 32% and 45% in the respective pre-Phase II (before 1 May 2008) and Phase II periods attended a specialist clinic within 12 months of diagnosis; the odds of attendance within 12 months increased over time (OR = 1.05 per year, 95% CI: 1.04-1.07), but was not significantly greater for persons diagnosed with HCV in the Phase II era, compared with the pre-Phase II era (OR = 1.1, 95% CI: 0.9-1.3), after adjustment for temporal trend. For outcome (b), 13% and 28% were initiated on treatment within 12 months of their first clinic attendance in the pre-Phase II and Phase II periods, respectively. Higher odds of treatment initiation were associated with first clinic attendance in the Phase II (OR = 1.9, 95% CI: 1.5-2.4), compared with the pre-Phase II era. Results were consistent with a positive impact of the Hepatitis C Action Plan on the treatment of chronically infected individuals, but further monitoring is required to confirm a sustained effect. © 2013 John Wiley & Sons Ltd.
Maximizing return on socioeconomic investment in phase II proof-of-concept trials.
Chen, Cong; Beckman, Robert A
2014-04-01
Phase II proof-of-concept (POC) trials play a key role in oncology drug development, determining which therapeutic hypotheses will undergo definitive phase III testing according to predefined Go-No Go (GNG) criteria. The number of possible POC hypotheses likely far exceeds available public or private resources. We propose a design strategy for maximizing return on socioeconomic investment in phase II trials that obtains the greatest knowledge with the minimum patient exposure. We compare efficiency using the benefit-cost ratio, defined to be the risk-adjusted number of truly active drugs correctly identified for phase III development divided by the risk-adjusted total sample size in phase II and III development, for different POC trial sizes, powering schemes, and associated GNG criteria. It is most cost-effective to conduct small POC trials and set the corresponding GNG bars high, so that more POC trials can be conducted under socioeconomic constraints. If δ is the minimum treatment effect size of clinical interest in phase II, the study design with the highest benefit-cost ratio has approximately 5% type I error rate and approximately 20% type II error rate (80% power) for detecting an effect size of approximately 1.5δ. A Go decision to phase III is made when the observed effect size is close to δ. With the phenomenal expansion of our knowledge in molecular biology leading to an unprecedented number of new oncology drug targets, conducting more small POC trials and setting high GNG bars maximize the return on socioeconomic investment in phase II POC trials. ©2014 AACR.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Abe, Katsunori; Kohyama, Akira; Tanaka, Satoru
This report describes an outline of the activities of the JUPITER-II collaboration (japan-USA program of Irradiation/Integration test for Fusion Research-II), Which has bee carried out through six years (2001-2006) under Phase 4 of the collabroation implemented by Amendment 4 of Annex 1 to the DOE (United States Department of Energy)-MEXT (Ministry of Education ,Culture,Sports,Science and Technology) Cooperation. This program followed the RTNS-II Program (Phase1:1982-4986), the FFTF/MOTA Program (Phase2:1987-1994) and the JUPITER Program (Phase 3: 1995-2000) [1].
Upgrade for Phase II of the Gerda experiment
NASA Astrophysics Data System (ADS)
Agostini, M.; Bakalyarov, A. M.; Balata, M.; Barabanov, I.; Baudis, L.; Bauer, C.; Bellotti, E.; Belogurov, S.; Belyaev, S. T.; Benato, G.; Bettini, A.; Bezrukov, L.; Bode, T.; Borowicz, D.; Brudanin, V.; Brugnera, R.; Caldwell, A.; Cattadori, C.; Chernogorov, A.; D'Andrea, V.; Demidova, E. V.; Di Marco, N.; Domula, A.; Doroshkevich, E.; Egorov, V.; Falkenstein, R.; Frodyma, N.; Gangapshev, A.; Garfagnini, A.; Grabmayr, P.; Gurentsov, V.; Gusev, K.; Hakenmüller, J.; Hegai, A.; Heisel, M.; Hemmer, S.; Hiller, R.; Hofmann, W.; Hult, M.; Inzhechik, L. V.; Ioannucci, L.; Janicskó Csáthy, J.; Jochum, J.; Junker, M.; Kazalov, V.; Kermaïdic, Y.; Kihm, T.; Kirpichnikov, I. V.; Kirsch, A.; Kish, A.; Klimenko, A.; Kneißl, R.; Knöpfle, K. T.; Kochetov, O.; Kornoukhov, V. N.; Kuzminov, V. V.; Laubenstein, M.; Lazzaro, A.; Lebedev, V. I.; Lehnert, B.; Lindner, M.; Lippi, I.; Lubashevskiy, A.; Lubsandorzhiev, B.; Lutter, G.; Macolino, C.; Majorovits, B.; Maneschg, W.; Medinaceli, E.; Miloradovic, M.; Mingazheva, R.; Misiaszek, M.; Moseev, P.; Nemchenok, I.; Nisi, S.; Panas, K.; Pandola, L.; Pelczar, K.; Pullia, A.; Ransom, C.; Riboldi, S.; Rumyantseva, N.; Sada, C.; Salamida, F.; Salathe, M.; Schmitt, C.; Schneider, B.; Schönert, S.; Schreiner, J.; Schütz, A.-K.; Schulz, O.; Schwingenheuer, B.; Selivanenko, O.; Shevchik, E.; Shirchenko, M.; Simgen, H.; Smolnikov, A.; Stanco, L.; Vanhoefer, L.; Vasenko, A. A.; Veresnikova, A.; von Sturm, K.; Wagner, V.; Wegmann, A.; Wester, T.; Wiesinger, C.; Wojcik, M.; Yanovich, E.; Zhitnikov, I.; Zhukov, S. V.; Zinatulina, D.; Zsigmond, A. J.; Zuber, K.; Zuzel, G.
2018-05-01
The Gerda collaboration is performing a sensitive search for neutrinoless double beta decay of ^{76}Ge at the INFN Laboratori Nazionali del Gran Sasso, Italy. The upgrade of the Gerda experiment from Phase I to Phase II has been concluded in December 2015. The first Phase II data release shows that the goal to suppress the background by one order of magnitude compared to Phase I has been achieved. Gerda is thus the first experiment that will remain "background-free" up to its design exposure (100 kg year). It will reach thereby a half-life sensitivity of more than 10^{26} year within 3 years of data collection. This paper describes in detail the modifications and improvements of the experimental setup for Phase II and discusses the performance of individual detector components.
40 CFR 80.45 - Complex emissions model.
Code of Federal Regulations, 2011 CFR
2011-07-01
...) VOCW% = Percentage change in winter VOC emissions from baseline levels (8) Phase II total VOC emissions... its domain Phase I = The years 1995-1999 Phase II = Year 2000 and beyond (b) Weightings and baselines... appropriate pollutant and Phase: Table 1—Normal and Higher Emitter Weightings for Exhaust Emissions Phase I...
Weinfurt, Kevin P.; Seils, Damon M.; Tzeng, Janice P.; Compton, Kate L.; Sulmasy, Daniel P.; Astrow, Alan B.; Solarino, Nicholas A.; Schulman, Kevin A.; Meropol, Neal J.
2009-01-01
Background Participants in early-phase clinical trials have reported high expectations of benefit from their participation. There is concern that participants misunderstand the trials to which they have consented. Such concern is based on assumptions about what patients mean when they respond to questions about likelihood of benefit. Methods Participants were 27 women and 18 men in early-phase oncology trials at 2 academic medical centers in the United States. To determine whether expectations of benefit differ depending on how patients are queried, we randomly assigned participants to 1 of 3 interviews corresponding to 3 questions about likelihood of benefit: frequency-type, belief-type, and vague. In semistructured interviews, we queried participants about how they understood and answered the question. Participants then answered and discussed one of the other questions. Results Expectations of benefit in response to the belief-type question were significantly greater than expectations in response to the frequency-type and vague questions (P = .02). The most common justifications involved positive attitude (n = 27 [60%]) and references to physical health (n = 23 [51%]). References to positive attitude were most common among participants with higher (> 70%) expectations (n = 11 [85%]) and least common among those with lower (< 50%) expectations (n = 3 [27%]). Conclusions The wording of questions about likelihood of benefit shapes the expectations that patients express. Also, patients who express high expectations may not do so to communicate understanding, but rather to register optimism. Ongoing research will clarify the meaning of high expectations and examine methods for assessing understanding in this context. PMID:18378940
LAILAPS-QSM: A RESTful API and JAVA library for semantic query suggestions.
Chen, Jinbo; Scholz, Uwe; Zhou, Ruonan; Lange, Matthias
2018-03-01
In order to access and filter content of life-science databases, full text search is a widely applied query interface. But its high flexibility and intuitiveness is paid for with potentially imprecise and incomplete query results. To reduce this drawback, query assistance systems suggest those combinations of keywords with the highest potential to match most of the relevant data records. Widespread approaches are syntactic query corrections that avoid misspelling and support expansion of words by suffixes and prefixes. Synonym expansion approaches apply thesauri, ontologies, and query logs. All need laborious curation and maintenance. Furthermore, access to query logs is in general restricted. Approaches that infer related queries by their query profile like research field, geographic location, co-authorship, affiliation etc. require user's registration and its public accessibility that contradict privacy concerns. To overcome these drawbacks, we implemented LAILAPS-QSM, a machine learning approach that reconstruct possible linguistic contexts of a given keyword query. The context is referred from the text records that are stored in the databases that are going to be queried or extracted for a general purpose query suggestion from PubMed abstracts and UniProt data. The supplied tool suite enables the pre-processing of these text records and the further computation of customized distributed word vectors. The latter are used to suggest alternative keyword queries. An evaluated of the query suggestion quality was done for plant science use cases. Locally present experts enable a cost-efficient quality assessment in the categories trait, biological entity, taxonomy, affiliation, and metabolic function which has been performed using ontology term similarities. LAILAPS-QSM mean information content similarity for 15 representative queries is 0.70, whereas 34% have a score above 0.80. In comparison, the information content similarity for human expert made query suggestions is 0.90. The software is either available as tool set to build and train dedicated query suggestion services or as already trained general purpose RESTful web service. The service uses open interfaces to be seamless embeddable into database frontends. The JAVA implementation uses highly optimized data structures and streamlined code to provide fast and scalable response for web service calls. The source code of LAILAPS-QSM is available under GNU General Public License version 2 in Bitbucket GIT repository: https://bitbucket.org/ipk_bit_team/bioescorte-suggestion.
Tao, Shiqiang; Cui, Licong; Wu, Xi; Zhang, Guo-Qiang
2017-01-01
To help researchers better access clinical data, we developed a prototype query engine called DataSphere for exploring large-scale integrated clinical data repositories. DataSphere expedites data importing using a NoSQL data management system and dynamically renders its user interface for concept-based querying tasks. DataSphere provides an interactive query-building interface together with query translation and optimization strategies, which enable users to build and execute queries effectively and efficiently. We successfully loaded a dataset of one million patients for University of Kentucky (UK) Healthcare into DataSphere with more than 300 million clinical data records. We evaluated DataSphere by comparing it with an instance of i2b2 deployed at UK Healthcare, demonstrating that DataSphere provides enhanced user experience for both query building and execution.
Tao, Shiqiang; Cui, Licong; Wu, Xi; Zhang, Guo-Qiang
2017-01-01
To help researchers better access clinical data, we developed a prototype query engine called DataSphere for exploring large-scale integrated clinical data repositories. DataSphere expedites data importing using a NoSQL data management system and dynamically renders its user interface for concept-based querying tasks. DataSphere provides an interactive query-building interface together with query translation and optimization strategies, which enable users to build and execute queries effectively and efficiently. We successfully loaded a dataset of one million patients for University of Kentucky (UK) Healthcare into DataSphere with more than 300 million clinical data records. We evaluated DataSphere by comparing it with an instance of i2b2 deployed at UK Healthcare, demonstrating that DataSphere provides enhanced user experience for both query building and execution. PMID:29854239
Improve Performance of Data Warehouse by Query Cache
NASA Astrophysics Data System (ADS)
Gour, Vishal; Sarangdevot, S. S.; Sharma, Anand; Choudhary, Vinod
2010-11-01
The primary goal of data warehouse is to free the information locked up in the operational database so that decision makers and business analyst can make queries, analysis and planning regardless of the data changes in operational database. As the number of queries is large, therefore, in certain cases there is reasonable probability that same query submitted by the one or multiple users at different times. Each time when query is executed, all the data of warehouse is analyzed to generate the result of that query. In this paper we will study how using query cache improves performance of Data Warehouse and try to find the common problems faced. These kinds of problems are faced by Data Warehouse administrators which are minimizes response time and improves the efficiency of query in data warehouse overall, particularly when data warehouse is updated at regular interval.
Safari, Leila; Patrick, Jon D
2018-06-01
This paper reports on a generic framework to provide clinicians with the ability to conduct complex analyses on elaborate research topics using cascaded queries to resolve internal time-event dependencies in the research questions, as an extension to the proposed Clinical Data Analytics Language (CliniDAL). A cascaded query model is proposed to resolve internal time-event dependencies in the queries which can have up to five levels of criteria starting with a query to define subjects to be admitted into a study, followed by a query to define the time span of the experiment. Three more cascaded queries can be required to define control groups, control variables and output variables which all together simulate a real scientific experiment. According to the complexity of the research questions, the cascaded query model has the flexibility of merging some lower level queries for simple research questions or adding a nested query to each level to compose more complex queries. Three different scenarios (one of them contains two studies) are described and used for evaluation of the proposed solution. CliniDAL's complex analyses solution enables answering complex queries with time-event dependencies at most in a few hours which manually would take many days. An evaluation of results of the research studies based on the comparison between CliniDAL and SQL solutions reveals high usability and efficiency of CliniDAL's solution. Copyright © 2018 Elsevier Inc. All rights reserved.
Chesapeake Bay Low Freshwater Inflow Study. Phase II. MAP FOLIO. Biota Assessment.
1982-05-01
conditions. These were: 1) Base Average -- average freshwater inflow conditions. by increased water consumption projected for the year 2020. 3) Base Drought...RESOLUTION TEST CHART NATIONAL BUREAU OF STANDARDS. 1963- A TAI m - ii J May 1982 Chesapeake Bay Low Freshwater Inflow Study Phase II Biota Assessment Map...A PERIOD ZOVERED change was found to CIESAPEAKE BAY LOW FRESHWATER INFLOW STUDY FINAL BIOTA ASSESSMENT PHASE II: FINAL REPORT MAP FOLIO s PERFORMING
Lee, Chih-Wei; Wang, Ji-Hung; Hsieh, Jen-Che; Hsieh, Tsung-Cheng; Huang, Chien-Hui
2013-01-01
[Purpose] To investigate the effects of cardiac exercise therapy (CET) on exercise capacity and coronary risk factors (CRFs) of patients with acute myocardial infarction (AMI). [Methods] Patients who participated in an 8-week supervised, hospital-based phase II and 6-month home-based phase III CET with monthly telephone and/or home visits were defined as the exercise group (EG) (n=20), while those who did not receive phase II or phase III CET were defined as the no-exercise group (NEG) (n=10). CRFs were evaluated pre- and post-phase II and eight months after discharge. One and two-way repeated measures ANOVA were used to perform intra- and inter-group comparisons. [Results] Thirty men with AMI aged 49.3 ± 8.3 years were studied. EG increased their exercise capacity (METs) (6.8 ± 1.6 vs.10.0 ± 1.9) after phase II CET and was able to maintain it at 8-month follow-up. Both groups had significantly fewer persons who kept on smoking compared to the first examination. High density lipoprotein cholesterol (HDL-C) increased from 38.1 ± 11.0 to 43.7 ± 8.7 mg/dl at follow-up in EG while no significant difference was noted in NEG. [Conclusion] After phase III CET subjects had maintained the therapeutic effects of smoking cessation, and increasing exercise capacity obtained in phase II CET. HDL-C in EG continued to improve during phase III CET. PMID:24396201
Performance Prediction of a MongoDB-Based Traceability System in Smart Factory Supply Chains
Kang, Yong-Shin; Park, Il-Ha; Youm, Sekyoung
2016-01-01
In the future, with the advent of the smart factory era, manufacturing and logistics processes will become more complex, and the complexity and criticality of traceability will further increase. This research aims at developing a performance assessment method to verify scalability when implementing traceability systems based on key technologies for smart factories, such as Internet of Things (IoT) and BigData. To this end, based on existing research, we analyzed traceability requirements and an event schema for storing traceability data in MongoDB, a document-based Not Only SQL (NoSQL) database. Next, we analyzed the algorithm of the most representative traceability query and defined a query-level performance model, which is composed of response times for the components of the traceability query algorithm. Next, this performance model was solidified as a linear regression model because the response times increase linearly by a benchmark test. Finally, for a case analysis, we applied the performance model to a virtual automobile parts logistics. As a result of the case study, we verified the scalability of a MongoDB-based traceability system and predicted the point when data node servers should be expanded in this case. The traceability system performance assessment method proposed in this research can be used as a decision-making tool for hardware capacity planning during the initial stage of construction of traceability systems and during their operational phase. PMID:27983654
Performance Prediction of a MongoDB-Based Traceability System in Smart Factory Supply Chains.
Kang, Yong-Shin; Park, Il-Ha; Youm, Sekyoung
2016-12-14
In the future, with the advent of the smart factory era, manufacturing and logistics processes will become more complex, and the complexity and criticality of traceability will further increase. This research aims at developing a performance assessment method to verify scalability when implementing traceability systems based on key technologies for smart factories, such as Internet of Things (IoT) and BigData. To this end, based on existing research, we analyzed traceability requirements and an event schema for storing traceability data in MongoDB, a document-based Not Only SQL (NoSQL) database. Next, we analyzed the algorithm of the most representative traceability query and defined a query-level performance model, which is composed of response times for the components of the traceability query algorithm. Next, this performance model was solidified as a linear regression model because the response times increase linearly by a benchmark test. Finally, for a case analysis, we applied the performance model to a virtual automobile parts logistics. As a result of the case study, we verified the scalability of a MongoDB-based traceability system and predicted the point when data node servers should be expanded in this case. The traceability system performance assessment method proposed in this research can be used as a decision-making tool for hardware capacity planning during the initial stage of construction of traceability systems and during their operational phase.
47 CFR 54.310 - Connect America Fund for Price Cap Territories-Phase II
Code of Federal Regulations, 2014 CFR
2014-10-01
... 47 Telecommunication 3 2014-10-01 2014-10-01 false Connect America Fund for Price Cap Territories... Connect America Fund for Price Cap Territories—Phase II (a) Geographic areas eligible for support. Connect America Phase II support may be made available for census blocks or other areas identified as eligible by...
Evaluation of Sub Query Performance in SQL Server
NASA Astrophysics Data System (ADS)
Oktavia, Tanty; Sujarwo, Surya
2014-03-01
The paper explores several sub query methods used in a query and their impact on the query performance. The study uses experimental approach to evaluate the performance of each sub query methods combined with indexing strategy. The sub query methods consist of in, exists, relational operator and relational operator combined with top operator. The experimental shows that using relational operator combined with indexing strategy in sub query has greater performance compared with using same method without indexing strategy and also other methods. In summary, for application that emphasized on the performance of retrieving data from database, it better to use relational operator combined with indexing strategy. This study is done on Microsoft SQL Server 2012.
Phase I/II adaptive design for drug combination oncology trials
Wages, Nolan A.; Conaway, Mark R.
2014-01-01
Existing statistical methodology on dose finding for combination chemotherapies has focused on toxicity considerations alone in finding a maximum tolerated dose combination to recommend for further testing of efficacy in a phase II setting. Recently, there has been increasing interest in integrating phase I and phase II trials in order to facilitate drug development. In this article, we propose a new adaptive phase I/II method for dual-agent combinations that takes into account both toxicity and efficacy after each cohort inclusion. The primary objective, both within and at the conclusion of the trial, becomes finding a single dose combination with an acceptable level of toxicity that maximizes efficacious response. We assume that there exist monotone dose–toxicity and dose–efficacy relationships among doses of one agent when the dose of other agent is fixed. We perform extensive simulation studies that demonstrate the operating characteristics of our proposed approach, and we compare simulated results to existing methodology in phase I/II design for combinations of agents. PMID:24470329
NASA Astrophysics Data System (ADS)
Yurtseven, H.; Kavruk, D.
In this study, we calculate the Raman frequencies as a function of temperature for the fixed pressures of 706, 1080 and 6355 bars using the volume data for phase II of ammonium iodide. The Raman frequencies calculated here are for the translational optic ν5 TOM (125 cm-1) lattice mode that is located at the zone boundary (M point) of the Brillouin zone of phase II for NH4I. For this calculation the volume data obtained at zero pressure, is used through the mode Grüneisen parameter for the disordered phase II (β phase) which has the CsCl structure of NH4I. Our predicted frequencies of the ν5 TOM (125 cm-1) mode can be compared when the Raman data for this lattice mode is available at various temperatures for fixed pressures of 706, 1080 and 6355 bars in the disordered phase II of ammonium iodide.
Secure Skyline Queries on Cloud Platform.
Liu, Jinfei; Yang, Juncheng; Xiong, Li; Pei, Jian
2017-04-01
Outsourcing data and computation to cloud server provides a cost-effective way to support large scale data storage and query processing. However, due to security and privacy concerns, sensitive data (e.g., medical records) need to be protected from the cloud server and other unauthorized users. One approach is to outsource encrypted data to the cloud server and have the cloud server perform query processing on the encrypted data only. It remains a challenging task to support various queries over encrypted data in a secure and efficient way such that the cloud server does not gain any knowledge about the data, query, and query result. In this paper, we study the problem of secure skyline queries over encrypted data. The skyline query is particularly important for multi-criteria decision making but also presents significant challenges due to its complex computations. We propose a fully secure skyline query protocol on data encrypted using semantically-secure encryption. As a key subroutine, we present a new secure dominance protocol, which can be also used as a building block for other queries. Finally, we provide both serial and parallelized implementations and empirically study the protocols in terms of efficiency and scalability under different parameter settings, verifying the feasibility of our proposed solutions.
Towards Hybrid Online On-Demand Querying of Realtime Data with Stateful Complex Event Processing
DOE Office of Scientific and Technical Information (OSTI.GOV)
Zhou, Qunzhi; Simmhan, Yogesh; Prasanna, Viktor K.
Emerging Big Data applications in areas like e-commerce and energy industry require both online and on-demand queries to be performed over vast and fast data arriving as streams. These present novel challenges to Big Data management systems. Complex Event Processing (CEP) is recognized as a high performance online query scheme which in particular deals with the velocity aspect of the 3-V’s of Big Data. However, traditional CEP systems do not consider data variety and lack the capability to embed ad hoc queries over the volume of data streams. In this paper, we propose H2O, a stateful complex event processing framework,more » to support hybrid online and on-demand queries over realtime data. We propose a semantically enriched event and query model to address data variety. A formal query algebra is developed to precisely capture the stateful and containment semantics of online and on-demand queries. We describe techniques to achieve the interactive query processing over realtime data featured by efficient online querying, dynamic stream data persistence and on-demand access. The system architecture is presented and the current implementation status reported.« less
Demonstration of Hadoop-GIS: A Spatial Data Warehousing System Over MapReduce.
Aji, Ablimit; Sun, Xiling; Vo, Hoang; Liu, Qioaling; Lee, Rubao; Zhang, Xiaodong; Saltz, Joel; Wang, Fusheng
2013-11-01
The proliferation of GPS-enabled devices, and the rapid improvement of scientific instruments have resulted in massive amounts of spatial data in the last decade. Support of high performance spatial queries on large volumes data has become increasingly important in numerous fields, which requires a scalable and efficient spatial data warehousing solution as existing approaches exhibit scalability limitations and efficiency bottlenecks for large scale spatial applications. In this demonstration, we present Hadoop-GIS - a scalable and high performance spatial query system over MapReduce. Hadoop-GIS provides an efficient spatial query engine to process spatial queries, data and space based partitioning, and query pipelines that parallelize queries implicitly on MapReduce. Hadoop-GIS also provides an expressive, SQL-like spatial query language for workload specification. We will demonstrate how spatial queries are expressed in spatially extended SQL queries, and submitted through a command line/web interface for execution. Parallel to our system demonstration, we explain the system architecture and details on how queries are translated to MapReduce operators, optimized, and executed on Hadoop. In addition, we will showcase how the system can be used to support two representative real world use cases: large scale pathology analytical imaging, and geo-spatial data warehousing.
Query Health: standards-based, cross-platform population health surveillance
Klann, Jeffrey G; Buck, Michael D; Brown, Jeffrey; Hadley, Marc; Elmore, Richard; Weber, Griffin M; Murphy, Shawn N
2014-01-01
Objective Understanding population-level health trends is essential to effectively monitor and improve public health. The Office of the National Coordinator for Health Information Technology (ONC) Query Health initiative is a collaboration to develop a national architecture for distributed, population-level health queries across diverse clinical systems with disparate data models. Here we review Query Health activities, including a standards-based methodology, an open-source reference implementation, and three pilot projects. Materials and methods Query Health defined a standards-based approach for distributed population health queries, using an ontology based on the Quality Data Model and Consolidated Clinical Document Architecture, Health Quality Measures Format (HQMF) as the query language, the Query Envelope as the secure transport layer, and the Quality Reporting Document Architecture as the result language. Results We implemented this approach using Informatics for Integrating Biology and the Bedside (i2b2) and hQuery for data analytics and PopMedNet for access control, secure query distribution, and response. We deployed the reference implementation at three pilot sites: two public health departments (New York City and Massachusetts) and one pilot designed to support Food and Drug Administration post-market safety surveillance activities. The pilots were successful, although improved cross-platform data normalization is needed. Discussions This initiative resulted in a standards-based methodology for population health queries, a reference implementation, and revision of the HQMF standard. It also informed future directions regarding interoperability and data access for ONC's Data Access Framework initiative. Conclusions Query Health was a test of the learning health system that supplied a functional methodology and reference implementation for distributed population health queries that has been validated at three sites. PMID:24699371
Query Health: standards-based, cross-platform population health surveillance.
Klann, Jeffrey G; Buck, Michael D; Brown, Jeffrey; Hadley, Marc; Elmore, Richard; Weber, Griffin M; Murphy, Shawn N
2014-01-01
Understanding population-level health trends is essential to effectively monitor and improve public health. The Office of the National Coordinator for Health Information Technology (ONC) Query Health initiative is a collaboration to develop a national architecture for distributed, population-level health queries across diverse clinical systems with disparate data models. Here we review Query Health activities, including a standards-based methodology, an open-source reference implementation, and three pilot projects. Query Health defined a standards-based approach for distributed population health queries, using an ontology based on the Quality Data Model and Consolidated Clinical Document Architecture, Health Quality Measures Format (HQMF) as the query language, the Query Envelope as the secure transport layer, and the Quality Reporting Document Architecture as the result language. We implemented this approach using Informatics for Integrating Biology and the Bedside (i2b2) and hQuery for data analytics and PopMedNet for access control, secure query distribution, and response. We deployed the reference implementation at three pilot sites: two public health departments (New York City and Massachusetts) and one pilot designed to support Food and Drug Administration post-market safety surveillance activities. The pilots were successful, although improved cross-platform data normalization is needed. This initiative resulted in a standards-based methodology for population health queries, a reference implementation, and revision of the HQMF standard. It also informed future directions regarding interoperability and data access for ONC's Data Access Framework initiative. Query Health was a test of the learning health system that supplied a functional methodology and reference implementation for distributed population health queries that has been validated at three sites. Published by the BMJ Publishing Group Limited. For permission to use (where not already granted under a licence) please go to http://group.bmj.com/group/rights-licensing/permissions.
Using search engine query data to track pharmaceutical utilization: a study of statins.
Schuster, Nathaniel M; Rogers, Mary A M; McMahon, Laurence F
2010-08-01
To examine temporal and geographic associations between Google queries for health information and healthcare utilization benchmarks. Retrospective longitudinal study. Using Google Trends and Google Insights for Search data, the search terms Lipitor (atorvastatin calcium; Pfizer, Ann Arbor, MI) and simvastatin were evaluated for change over time and for association with Lipitor revenues. The relationship between query data and community-based resource use per Medicare beneficiary was assessed for 35 US metropolitan areas. Google queries for Lipitor significantly decreased from January 2004 through June 2009 and queries for simvastatin significantly increased (P <.001 for both), particularly after Lipitor came off patent (P <.001 for change in slope). The mean number of Google queries for Lipitor correlated (r = 0.98) with the percentage change in Lipitor global revenues from 2004 to 2008 (P <.001). Query preference for Lipitor over simvastatin was positively associated (r = 0.40) with a community's use of Medicare services. For every 1% increase in utilization of Medicare services in a community, there was a 0.2-unit increase in the ratio of Lipitor queries to simvastatin queries in that community (P = .02). Specific search engine queries for medical information correlate with pharmaceutical revenue and with overall healthcare utilization in a community. This suggests that search query data can track community-wide characteristics in healthcare utilization and have the potential for informing payers and policy makers regarding trends in utilization.
The National Geographic Names Data Base: Phase II instructions
Orth, Donald J.; Payne, Roger L.
1987-01-01
not recorded on topographic maps be added. The systematic collection of names from other sources, including maps, charts, and texts, is termed Phase II. In addition, specific types of features not compiled during Phase I are encoded and added to the data base. Other names of importance to researchers and users, such as historical and variant names, are also included. The rules and procedures for Phase II research, compilation, and encoding are contained in this publication.
CSRQ: Communication-Efficient Secure Range Queries in Two-Tiered Sensor Networks
Dai, Hua; Ye, Qingqun; Yang, Geng; Xu, Jia; He, Ruiliang
2016-01-01
In recent years, we have seen many applications of secure query in two-tiered wireless sensor networks. Storage nodes are responsible for storing data from nearby sensor nodes and answering queries from Sink. It is critical to protect data security from a compromised storage node. In this paper, the Communication-efficient Secure Range Query (CSRQ)—a privacy and integrity preserving range query protocol—is proposed to prevent attackers from gaining information of both data collected by sensor nodes and queries issued by Sink. To preserve privacy and integrity, in addition to employing the encoding mechanisms, a novel data structure called encrypted constraint chain is proposed, which embeds the information of integrity verification. Sink can use this encrypted constraint chain to verify the query result. The performance evaluation shows that CSRQ has lower communication cost than the current range query protocols. PMID:26907293
SPARQLGraph: a web-based platform for graphically querying biological Semantic Web databases.
Schweiger, Dominik; Trajanoski, Zlatko; Pabinger, Stephan
2014-08-15
Semantic Web has established itself as a framework for using and sharing data across applications and database boundaries. Here, we present a web-based platform for querying biological Semantic Web databases in a graphical way. SPARQLGraph offers an intuitive drag & drop query builder, which converts the visual graph into a query and executes it on a public endpoint. The tool integrates several publicly available Semantic Web databases, including the databases of the just recently released EBI RDF platform. Furthermore, it provides several predefined template queries for answering biological questions. Users can easily create and save new query graphs, which can also be shared with other researchers. This new graphical way of creating queries for biological Semantic Web databases considerably facilitates usability as it removes the requirement of knowing specific query languages and database structures. The system is freely available at http://sparqlgraph.i-med.ac.at.
Mixed response and time-to-event endpoints for multistage single-arm phase II design.
Lai, Xin; Zee, Benny Chung-Ying
2015-06-04
The objective of phase II cancer clinical trials is to determine if a treatment has sufficient activity to warrant further study. The efficiency of a conventional phase II trial design has been the object of considerable debate, particularly when the study regimen is characteristically cytostatic. At the time of development of a phase II cancer trial, we accumulated clinical experience regarding the time to progression (TTP) for similar classes of drugs and for standard therapy. By considering the time to event (TTE) in addition to the tumor response endpoint, a mixed-endpoint phase II design may increase the efficiency and ability of selecting promising cytotoxic and cytostatic agents for further development. We proposed a single-arm phase II trial design by extending the Zee multinomial method to fully use mixed endpoints with tumor response and the TTE. In this design, the dependence between the probability of response and the TTE outcome is modeled through a Gaussian copula. Given the type I and type II errors and the hypothesis as defined by the response rate (RR) and median TTE, such as median TTP, the decision rules for a two-stage phase II trial design can be generated. We demonstrated through simulation that the proposed design has a smaller expected sample size and higher early stopping probability under the null hypothesis than designs based on a single-response endpoint or a single TTE endpoint. The proposed design is more efficient for screening new cytotoxic or cytostatic agents and less likely to miss an effective agent than the alternative single-arm design.
NASA Astrophysics Data System (ADS)
Antani, Sameer K.; Natarajan, Mukil; Long, Jonathan L.; Long, L. Rodney; Thoma, George R.
2005-04-01
The article describes the status of our ongoing R&D at the U.S. National Library of Medicine (NLM) towards the development of an advanced multimedia database biomedical information system that supports content-based image retrieval (CBIR). NLM maintains a collection of 17,000 digitized spinal X-rays along with text survey data from the Second National Health and Nutritional Examination Survey (NHANES II). These data serve as a rich data source for epidemiologists and researchers of osteoarthritis and musculoskeletal diseases. It is currently possible to access these through text keyword queries using our Web-based Medical Information Retrieval System (WebMIRS). CBIR methods developed specifically for biomedical images could offer direct visual searching of these images by means of example image or user sketch. We are building a system which supports hybrid queries that have text and image-content components. R&D goals include developing algorithms for robust image segmentation for localizing and identifying relevant anatomy, labeling the segmented anatomy based on its pathology, developing suitable indexing and similarity matching methods for images and image features, and associating the survey text information for query and retrieval along with the image data. Some highlights of the system developed in MATLAB and Java are: use of a networked or local centralized database for text and image data; flexibility to incorporate new research work; provides a means to control access to system components under development; and use of XML for structured reporting. The article details the design, features, and algorithms in this third revision of this prototype system, CBIR3.
Improving accuracy for identifying related PubMed queries by an integrated approach.
Lu, Zhiyong; Wilbur, W John
2009-10-01
PubMed is the most widely used tool for searching biomedical literature online. As with many other online search tools, a user often types a series of multiple related queries before retrieving satisfactory results to fulfill a single information need. Meanwhile, it is also a common phenomenon to see a user type queries on unrelated topics in a single session. In order to study PubMed users' search strategies, it is necessary to be able to automatically separate unrelated queries and group together related queries. Here, we report a novel approach combining both lexical and contextual analyses for segmenting PubMed query sessions and identifying related queries and compare its performance with the previous approach based solely on concept mapping. We experimented with our integrated approach on sample data consisting of 1539 pairs of consecutive user queries in 351 user sessions. The prediction results of 1396 pairs agreed with the gold-standard annotations, achieving an overall accuracy of 90.7%. This demonstrates that our approach is significantly better than the previously published method. By applying this approach to a one day query log of PubMed, we found that a significant proportion of information needs involved more than one PubMed query, and that most of the consecutive queries for the same information need are lexically related. Finally, the proposed PubMed distance is shown to be an accurate and meaningful measure for determining the contextual similarity between biological terms. The integrated approach can play a critical role in handling real-world PubMed query log data as is demonstrated in our experiments.
Improving accuracy for identifying related PubMed queries by an integrated approach
Lu, Zhiyong; Wilbur, W. John
2009-01-01
PubMed is the most widely used tool for searching biomedical literature online. As with many other online search tools, a user often types a series of multiple related queries before retrieving satisfactory results to fulfill a single information need. Meanwhile, it is also a common phenomenon to see a user type queries on unrelated topics in a single session. In order to study PubMed users’ search strategies, it is necessary to be able to automatically separate unrelated queries and group together related queries. Here, we report a novel approach combining both lexical and contextual analyses for segmenting PubMed query sessions and identifying related queries and compare its performance with the previous approach based solely on concept mapping. We experimented with our integrated approach on sample data consisting of 1,539 pairs of consecutive user queries in 351 user sessions. The prediction results of 1,396 pairs agreed with the gold-standard annotations, achieving an overall accuracy of 90.7%. This demonstrates that our approach is significantly better than the previously published method. By applying this approach to a one day query log of PubMed, we found that a significant proportion of information needs involved more than one PubMed query, and that most of the consecutive queries for the same information need are lexically related. Finally, the proposed PubMed distance is shown to be an accurate and meaningful measure for determining the contextual similarity between biological terms. The integrated approach can play a critical role in handling real-world PubMed query log data as is demonstrated in our experiments. PMID:19162232
Dattilo, David J; Drooger, Scott A
2004-02-01
The purpose of this study was to compare the subjective findings of the Epworth Sleepiness Scale (ESS) to the objective findings of the overnight sleep study (OSS) in 57 patients who underwent phase I and phase II surgery for the correction of obstructive sleep apnea (OSA). Forty-two patients in phase I category (hyoid suspension, palatal surgery, and/or genioglossus advancement) and 15 patients in phase II category (maxillomandibular advancement) were examined. All patients had an OSS and completion of an ESS preoperatively and at a minimum of 8 weeks postoperatively. The results of each test were evaluated to examine any relationship between the improvements of the findings of the OSS to the changes in the ESS. Using accepted criteria, phase I surgery produced an 80% success rate and phase II surgery produced a greater than 95% success rate in both the respiratory disturbance index and the ESS. 1) Both phase I and phase II procedures are effective in treating OSA. 2) Phase II appears to be more effective in treating OSA using both objective and subjective evaluations. 3) Improvement in ESS scores and excessive daytime sleepiness seems to parallel the improvement in OSS scores in patients undergoing surgical correction of OSA.
Multi-Bit Quantum Private Query
NASA Astrophysics Data System (ADS)
Shi, Wei-Xu; Liu, Xing-Tong; Wang, Jian; Tang, Chao-Jing
2015-09-01
Most of the existing Quantum Private Queries (QPQ) protocols provide only single-bit queries service, thus have to be repeated several times when more bits are retrieved. Wei et al.'s scheme for block queries requires a high-dimension quantum key distribution system to sustain, which is still restricted in the laboratory. Here, based on Markus Jakobi et al.'s single-bit QPQ protocol, we propose a multi-bit quantum private query protocol, in which the user can get access to several bits within one single query. We also extend the proposed protocol to block queries, using a binary matrix to guard database security. Analysis in this paper shows that our protocol has better communication complexity, implementability and can achieve a considerable level of security.
Estimating Missing Features to Improve Multimedia Information Retrieval
DOE Office of Scientific and Technical Information (OSTI.GOV)
Bagherjeiran, A; Love, N S; Kamath, C
Retrieval in a multimedia database usually involves combining information from different modalities of data, such as text and images. However, all modalities of the data may not be available to form the query. The retrieval results from such a partial query are often less than satisfactory. In this paper, we present an approach to complete a partial query by estimating the missing features in the query. Our experiments with a database of images and their associated captions show that, with an initial text-only query, our completion method has similar performance to a full query with both image and text features.more » In addition, when we use relevance feedback, our approach outperforms the results obtained using a full query.« less
NASA Astrophysics Data System (ADS)
Liao, S.; Chen, L.; Li, J.; Xiong, W.; Wu, Q.
2015-07-01
Existing spatiotemporal database supports spatiotemporal aggregation query over massive moving objects datasets. Due to the large amounts of data and single-thread processing method, the query speed cannot meet the application requirements. On the other hand, the query efficiency is more sensitive to spatial variation then temporal variation. In this paper, we proposed a spatiotemporal aggregation query method using multi-thread parallel technique based on regional divison and implemented it on the server. Concretely, we divided the spatiotemporal domain into several spatiotemporal cubes, computed spatiotemporal aggregation on all cubes using the technique of multi-thread parallel processing, and then integrated the query results. By testing and analyzing on the real datasets, this method has improved the query speed significantly.
ERIC Educational Resources Information Center
Walker, Bonnie L.
This report describes Phase II of a project which developed a system for delivering fire safety training to board and care providers who serve adults with developmental disabilities. Phase II focused on developing and pilot testing a "train the trainers" workshop for instructors and field testing the provider's workshop. Evaluation of…
Single-arm phase II trial design under parametric cure models.
Wu, Jianrong
2015-01-01
The current practice of designing single-arm phase II survival trials is limited under the exponential model. Trial design under the exponential model may not be appropriate when a portion of patients are cured. There is no literature available for designing single-arm phase II trials under the parametric cure model. In this paper, a test statistic is proposed, and a sample size formula is derived for designing single-arm phase II trials under a class of parametric cure models. Extensive simulations showed that the proposed test and sample size formula perform very well under different scenarios. Copyright © 2015 John Wiley & Sons, Ltd.
Cull, Brooke J; Dzewaltowski, David A; Guagliano, Justin M; Rosenkranz, Sara K; Knutson, Cassandra K; Rosenkranz, Richard R
2018-01-01
To evaluate the effectiveness of in-person versus online Girl Scout leader wellness training for implementation of wellness-promoting practices during troop meetings (phase I) and to assess training adoption and current practices across the council (phase II). Pragmatic superiority trial (phase 1) followed by serial cross-sectional study (phase II). Girl Scout troop meetings in Northeast Kansas. Eighteen troop leaders from 3 counties (phase 1); 113 troop leaders from 7 counties (phase II). Phase I: Troop leaders attended 2 wellness training sessions (first in groups, second individually), wherein leaders set wellness-promoting practice implementation goals, self-monitored progress, and received guidance and resources for implementation. Leaders received the intervention in person or online. Phase I: At baseline and postintervention, leaders completed a wellness-promoting practice implementation questionnaire assessing practices during troop meetings (max score = 11). Phase II: Leaders completed a survey about typical troop practices and interest in further training. Phase I: Generalized linear mixed modeling. Phase I: In-person training increased wellness-promoting practice implementation more than online training (in person = 2.1 ± 1.8; online = 0.2 ± 1.2; P = .022). Phase II: Fifty-six percent of leaders adopted the training. For 8 of 11 wellness categories, greater than 50% of leaders employed wellness-promoting practices. In-person training was superior to online training for improvements in wellness-promoting practices. Wellness training was adopted by the majority of leaders across the council.
NASA Astrophysics Data System (ADS)
Pi, E. I.; Siegel, E.
2010-03-01
Siegel[AMS Natl.Mtg.(2002)-Abs.973-60-124] digits logarithmic- law inversion to ONLY BEQS BEC:Quanta/Bosons=#: EMP-like SEVERE VULNERABILITY of ONLY #-networks(VS.ANALOG INvulnerability) via Barabasi NP(VS.dynamics[Not.AMS(5/2009)] critique);(so called)``quantum-computing''(QC) = simple-arithmetic (sansdivision);algorithmiccomplexities:INtractibility/UNdecidabi lity/INefficiency/NONcomputability/HARDNESS(so MIScalled) ``noise''-induced-phase-transition(NIT)ACCELERATION:Cook-Levin theorem Reducibility = RG fixed-points; #-Randomness DEFINITION via WHAT? Query(VS. Goldreich[Not.AMS(2002)] How? mea culpa)= ONLY MBCS hot-plasma v #-clumping NON-random BEC; Modular-Arithmetic Congruences = Signal x Noise PRODUCTS = clock-model; NON-Shor[Physica A,341,586(04)]BEC logarithmic-law inversion factorization: Watkins #-theory U statistical- physics); P=/=NP C-S TRIVIAL Proof: Euclid!!! [(So Miscalled) computational-complexity J-O obviation(3 millennia AGO geometry: NO:CC,``CS'';``Feet of Clay!!!'']; Query WHAT?:Definition: (so MIScalled)``complexity''=UTTER-SIMPLICITY!! v COMPLICATEDNESS MEASURE(S).
A Framework for WWW Query Processing
NASA Technical Reports Server (NTRS)
Wu, Binghui Helen; Wharton, Stephen (Technical Monitor)
2000-01-01
Query processing is the most common operation in a DBMS. Sophisticated query processing has been mainly targeted at a single enterprise environment providing centralized control over data and metadata. Submitting queries by anonymous users on the web is different in such a way that load balancing or DBMS' accessing control becomes the key issue. This paper provides a solution by introducing a framework for WWW query processing. The success of this framework lies in the utilization of query optimization techniques and the ontological approach. This methodology has proved to be cost effective at the NASA Goddard Space Flight Center Distributed Active Archive Center (GDAAC).
QBIC project: querying images by content, using color, texture, and shape
NASA Astrophysics Data System (ADS)
Niblack, Carlton W.; Barber, Ron; Equitz, Will; Flickner, Myron D.; Glasman, Eduardo H.; Petkovic, Dragutin; Yanker, Peter; Faloutsos, Christos; Taubin, Gabriel
1993-04-01
In the query by image content (QBIC) project we are studying methods to query large on-line image databases using the images' content as the basis of the queries. Examples of the content we use include color, texture, and shape of image objects and regions. Potential applications include medical (`Give me other images that contain a tumor with a texture like this one'), photo-journalism (`Give me images that have blue at the top and red at the bottom'), and many others in art, fashion, cataloging, retailing, and industry. Key issues include derivation and computation of attributes of images and objects that provide useful query functionality, retrieval methods based on similarity as opposed to exact match, query by image example or user drawn image, the user interfaces, query refinement and navigation, high dimensional database indexing, and automatic and semi-automatic database population. We currently have a prototype system written in X/Motif and C running on an RS/6000 that allows a variety of queries, and a test database of over 1000 images and 1000 objects populated from commercially available photo clip art images. In this paper we present the main algorithms for color texture, shape and sketch query that we use, show example query results, and discuss future directions.
Pentoney, Christopher; Harwell, Jeff; Leroy, Gondy
2014-01-01
Searching for medical information online is a common activity. While it has been shown that forming good queries is difficult, Google's query suggestion tool, a type of query expansion, aims to facilitate query formation. However, it is unknown how this expansion, which is based on what others searched for, affects the information gathering of the online community. To measure the impact of social-based query expansion, this study compared it with content-based expansion, i.e., what is really in the text. We used 138,906 medical queries from the AOL User Session Collection and expanded them using Google's Autocomplete method (social-based) and the content of the Google Web Corpus (content-based). We evaluated the specificity and ambiguity of the expansion terms for trigram queries. We also looked at the impact on the actual results using domain diversity and expansion edit distance. Results showed that the social-based method provided more precise expansion terms as well as terms that were less ambiguous. Expanded queries do not differ significantly in diversity when expanded using the social-based method (6.72 different domains returned in the first ten results, on average) vs. content-based method (6.73 different domains, on average).
a Novel Approach of Indexing and Retrieving Spatial Polygons for Efficient Spatial Region Queries
NASA Astrophysics Data System (ADS)
Zhao, J. H.; Wang, X. Z.; Wang, F. Y.; Shen, Z. H.; Zhou, Y. C.; Wang, Y. L.
2017-10-01
Spatial region queries are more and more widely used in web-based applications. Mechanisms to provide efficient query processing over geospatial data are essential. However, due to the massive geospatial data volume, heavy geometric computation, and high access concurrency, it is difficult to get response in real time. Spatial indexes are usually used in this situation. In this paper, based on k-d tree, we introduce a distributed KD-Tree (DKD-Tree) suitbable for polygon data, and a two-step query algorithm. The spatial index construction is recursive and iterative, and the query is an in memory process. Both the index and query methods can be processed in parallel, and are implemented based on HDFS, Spark and Redis. Experiments on a large volume of Remote Sensing images metadata have been carried out, and the advantages of our method are investigated by comparing with spatial region queries executed on PostgreSQL and PostGIS. Results show that our approach not only greatly improves the efficiency of spatial region query, but also has good scalability, Moreover, the two-step spatial range query algorithm can also save cluster resources to support a large number of concurrent queries. Therefore, this method is very useful when building large geographic information systems.
Secure Skyline Queries on Cloud Platform
Liu, Jinfei; Yang, Juncheng; Xiong, Li; Pei, Jian
2017-01-01
Outsourcing data and computation to cloud server provides a cost-effective way to support large scale data storage and query processing. However, due to security and privacy concerns, sensitive data (e.g., medical records) need to be protected from the cloud server and other unauthorized users. One approach is to outsource encrypted data to the cloud server and have the cloud server perform query processing on the encrypted data only. It remains a challenging task to support various queries over encrypted data in a secure and efficient way such that the cloud server does not gain any knowledge about the data, query, and query result. In this paper, we study the problem of secure skyline queries over encrypted data. The skyline query is particularly important for multi-criteria decision making but also presents significant challenges due to its complex computations. We propose a fully secure skyline query protocol on data encrypted using semantically-secure encryption. As a key subroutine, we present a new secure dominance protocol, which can be also used as a building block for other queries. Finally, we provide both serial and parallelized implementations and empirically study the protocols in terms of efficiency and scalability under different parameter settings, verifying the feasibility of our proposed solutions. PMID:28883710
NASA Astrophysics Data System (ADS)
Indrayana, I. N. E.; P, N. M. Wirasyanti D.; Sudiartha, I. KG
2018-01-01
Mobile application allow many users to access data from the application without being limited to space, space and time. Over time the data population of this application will increase. Data access time will cause problems if the data record has reached tens of thousands to millions of records.The objective of this research is to maintain the performance of data execution for large data records. One effort to maintain data access time performance is to apply query optimization method. The optimization used in this research is query heuristic optimization method. The built application is a mobile-based financial application using MySQL database with stored procedure therein. This application is used by more than one business entity in one database, thus enabling rapid data growth. In this stored procedure there is an optimized query using heuristic method. Query optimization is performed on a “Select” query that involves more than one table with multiple clausa. Evaluation is done by calculating the average access time using optimized and unoptimized queries. Access time calculation is also performed on the increase of population data in the database. The evaluation results shown the time of data execution with query heuristic optimization relatively faster than data execution time without using query optimization.
Demonstration of Hadoop-GIS: A Spatial Data Warehousing System Over MapReduce
Aji, Ablimit; Sun, Xiling; Vo, Hoang; Liu, Qioaling; Lee, Rubao; Zhang, Xiaodong; Saltz, Joel; Wang, Fusheng
2016-01-01
The proliferation of GPS-enabled devices, and the rapid improvement of scientific instruments have resulted in massive amounts of spatial data in the last decade. Support of high performance spatial queries on large volumes data has become increasingly important in numerous fields, which requires a scalable and efficient spatial data warehousing solution as existing approaches exhibit scalability limitations and efficiency bottlenecks for large scale spatial applications. In this demonstration, we present Hadoop-GIS – a scalable and high performance spatial query system over MapReduce. Hadoop-GIS provides an efficient spatial query engine to process spatial queries, data and space based partitioning, and query pipelines that parallelize queries implicitly on MapReduce. Hadoop-GIS also provides an expressive, SQL-like spatial query language for workload specification. We will demonstrate how spatial queries are expressed in spatially extended SQL queries, and submitted through a command line/web interface for execution. Parallel to our system demonstration, we explain the system architecture and details on how queries are translated to MapReduce operators, optimized, and executed on Hadoop. In addition, we will showcase how the system can be used to support two representative real world use cases: large scale pathology analytical imaging, and geo-spatial data warehousing. PMID:27617325
DOE Office of Scientific and Technical Information (OSTI.GOV)
Starcher, Autumn N.; Elzinga, Evert J.; Sparks, Donald L.
Previous research demonstrated the formation of single divalent metal (Co, Ni, and ZnAl) and mixed divalent metal (NiZnAl) layered double hydroxide (LDH) phases from reactions of the divalent metal with Al-bearing substrates and soils in both laboratory experiments and in the natural environment. Recently Fe(II)-Al-LDH phases have been found in laboratory batch reaction studies, and although they have yet to be found in the natural environment. Potential locations of Fe(II)-Al-LDH phases in nature include areas with suboxic and anoxic conditions. Because these areas can be environments of significant contaminant accumulation, it is important to understand the possible interactions and impactsmore » of contaminant elements on LDH phase formation. One such contaminant, Zn, can also form as an LDH and has been found to form as a mixed divalent layered hydroxide phase. To understand how Zn impacts the formation of Fe(II)-Al-LDH phase formation and kinetics, 3 mM or 0.8 mM Fe(II) and 0.8 mM Zn were batch reacted with either 10 g/L pyrophyllite or 7.5 g/L γ-Al2O3 for up to three months under anoxic conditions. Aqueous samples were analyzed by inductively coupled plasma optical emission spectrometry (ICP-OES) and solid samples were analyzed with X-ray absorption spectroscopy (XAS). Shell-by-shell fits of Fe(II) and co-sorption samples with pyrophyllite show the formation of a mixed divalent metal (Fe(II)-Zn-Al) layered hydroxide phase, while Fe(II) and Zn co-sorption samples with γ-Al2O3 produce Fe(II)-Al-LDH phases and Zn in inner-sphere complexation with the γ-Al2O3. This study demonstrates the formation of a mixed divalent metal layered hydroxide and further iterates the importance of sorbent reactivity on LDH phase formation.« less
Secure Encapsulation and Publication of Biological Services in the Cloud Computing Environment
Zhang, Weizhe; Wang, Xuehui; Lu, Bo; Kim, Tai-hoon
2013-01-01
Secure encapsulation and publication for bioinformatics software products based on web service are presented, and the basic function of biological information is realized in the cloud computing environment. In the encapsulation phase, the workflow and function of bioinformatics software are conducted, the encapsulation interfaces are designed, and the runtime interaction between users and computers is simulated. In the publication phase, the execution and management mechanisms and principles of the GRAM components are analyzed. The functions such as remote user job submission and job status query are implemented by using the GRAM components. The services of bioinformatics software are published to remote users. Finally the basic prototype system of the biological cloud is achieved. PMID:24078906
Secure encapsulation and publication of biological services in the cloud computing environment.
Zhang, Weizhe; Wang, Xuehui; Lu, Bo; Kim, Tai-hoon
2013-01-01
Secure encapsulation and publication for bioinformatics software products based on web service are presented, and the basic function of biological information is realized in the cloud computing environment. In the encapsulation phase, the workflow and function of bioinformatics software are conducted, the encapsulation interfaces are designed, and the runtime interaction between users and computers is simulated. In the publication phase, the execution and management mechanisms and principles of the GRAM components are analyzed. The functions such as remote user job submission and job status query are implemented by using the GRAM components. The services of bioinformatics software are published to remote users. Finally the basic prototype system of the biological cloud is achieved.
Swoboda, Sandra M; Earsing, Karen; Strauss, Kevin; Lane, Stephen; Lipsett, Pamela A
2004-02-01
To determine whether electronic monitoring of hand hygiene and voice prompts can improve hand hygiene and decrease nosocomial infection rates in a surgical intermediate care unit. Three-phase quasi-experimental design. Phase I was electronic monitoring and direct observation; phase II was electronic monitoring and computerized voice prompts for failure to perform hand hygiene on room exit; and phase III was electronic monitoring only. Nine-room, 14-bed intermediate care unit in a university, tertiary-care institution. All patient rooms, utility room, and staff lavatory were monitored electronically. All healthcare personnel including physicians, nurses, nursing support personnel, ancillary staff, all visitors and family members, and any other personnel interacting with patients on the intermediate care unit. All patients with an intermediate care unit length of stay >48 hrs were followed for nosocomial infection. Electronic monitoring during all phases, computerized voice prompts during phase II only. We evaluated a total of 283,488 electronically monitored entries into a patient room with 251,526 exits for 420 days (10,080 hrs and 3,549 patient days). Compared with phase I, hand hygiene compliance in patient rooms improved 37% during phase II (odds ratio, 1.38; 95% confidence interval, 1.04-1.83) and 41% in phase III (odds ratio, 1.41; 95% confidence interval, 1.07-1.84). When adjusting for patient admissions during each phase, point estimates of nosocomial infections decreased by 22% during phase II and 48% during phase III; when adjusting for patient days, the number of infections decreased by 10% during phase II and 40% during phase III. Although the overall rate of nosocomial infections significantly decreased when combining phases II and III, the association between nosocomial infection and individual phase was not significant. Electronic monitoring provided effective ongoing feedback about hand hygiene compliance. During both the voice prompt phase and post-intervention phase, hand hygiene compliance and nosocomial infection rates improved suggesting that ongoing monitoring and feedback had both a short-term and, perhaps, a longer-term effect.
A high performance, ad-hoc, fuzzy query processing system for relational databases
NASA Technical Reports Server (NTRS)
Mansfield, William H., Jr.; Fleischman, Robert M.
1992-01-01
Database queries involving imprecise or fuzzy predicates are currently an evolving area of academic and industrial research. Such queries place severe stress on the indexing and I/O subsystems of conventional database environments since they involve the search of large numbers of records. The Datacycle architecture and research prototype is a database environment that uses filtering technology to perform an efficient, exhaustive search of an entire database. It has recently been modified to include fuzzy predicates in its query processing. The approach obviates the need for complex index structures, provides unlimited query throughput, permits the use of ad-hoc fuzzy membership functions, and provides a deterministic response time largely independent of query complexity and load. This paper describes the Datacycle prototype implementation of fuzzy queries and some recent performance results.
Systems and methods for an extensible business application framework
NASA Technical Reports Server (NTRS)
Bell, David G. (Inventor); Crawford, Michael (Inventor)
2012-01-01
Method and systems for editing data from a query result include requesting a query result using a unique collection identifier for a collection of individual files and a unique identifier for a configuration file that specifies a data structure for the query result. A query result is generated that contains a plurality of fields as specified by the configuration file, by combining each of the individual files associated with a unique identifier for a collection of individual files. The query result data is displayed with a plurality of labels as specified in the configuration file. Edits can be performed by querying a collection of individual files using the configuration file, editing a portion of the query result, and transmitting only the edited information for storage back into a data repository.
CMIF ECLS system test findings
NASA Technical Reports Server (NTRS)
Schunk, Richard G.; Carrasquillo, Robyn L.; Ogle, Kathyrn Y.; Wieland, Paul O.; Bagdigian, Robert M.
1989-01-01
During 1987 three Space Station integrated Environmental Control and Life Support System (ECLSS) tests were conducted at the Marshall Space Flight Center (MSFC) Core Module Integration Facility (CMIF) as part of the MSFC ECLSS Phase II test program. The three tests ranged in duration from 50 to 150 hours and were conducted inside of the CMIF module simulator. The Phase II partial integrated system test configuration consisted of four regenerative air revitalization subsystems and one regenerative water reclamation subsystem. This paper contains a discussion of results and lessons learned from the Phase II test program. The design of the Phase II test configuration and improvements made throughout the program are detailed. Future plans for the MSFC CMIF test program are provided, including an overview of planned improvements for the Phase III program.
Analysis of Information Needs of Users of MEDLINEplus, 2002 – 2003
Scott-Wright, Alicia; Crowell, Jon; Zeng, Qing; Bates, David W.; Greenes, Robert
2006-01-01
We analyzed query logs from use of MEDLINEplus to answer the questions: Are consumers’ health information needs stable over time? and To what extent do users’ queries change over time? To determine log stability, we assessed an Overlap Rate (OR) defined as the number of unique queries common to two adjacent months divided by the total number of unique queries in those months. All exactly matching queries were considered as one unique query. We measured ORs for the top 10 and 100 unique queries of a month and compared these to ORs for the following month. Over ten months, users submitted 12,234,737 queries; only 2,179,571 (17.8%) were unique and these had a mean word count of 2.73 (S.D., 0.24); 121 of 137 (88.3%) unique queries each comprised of exactly matching search term(s) used at least 5000 times were of only one word. We could predict with 95% confidence that the monthly OR for the top 100 unique queries would lie between 67% – 87% when compared with the top 100 from the previous month. The mean month-to-month OR for top 10 queries was 62% (S.D., 20%) indicating significant variability; the lowest OR of 33% between the top 10 in Mar. compared to Apr. was likely due to “new” interest in information about SARS pneumonia in Apr. 2003. Consumers’ health information needs are relatively stable and the 100 most common unique queries are about 77% the same from month to month. Website sponsors should provide a broad range of information about a relatively stable number of topics. Analyses of log similarity may identify media-induced, cyclical, or seasonal changes in areas of consumer interest. PMID:17238431
Big Data and Dysmenorrhea: What Questions Do Women and Men Ask About Menstrual Pain?
Chen, Chen X; Groves, Doyle; Miller, Wendy R; Carpenter, Janet S
2018-04-30
Menstrual pain is highly prevalent among women of reproductive age. As the general public increasingly obtains health information online, Big Data from online platforms provide novel sources to understand the public's perspectives and information needs about menstrual pain. The study's purpose was to describe salient queries about dysmenorrhea using Big Data from a question and answer platform. We performed text-mining of 1.9 billion queries from ChaCha, a United States-based question and answer platform. Dysmenorrhea-related queries were identified by using keyword searching. Each relevant query was split into token words (i.e., meaningful words or phrases) and stop words (i.e., not meaningful functional words). Word Adjacency Graph (WAG) modeling was used to detect clusters of queries and visualize the range of dysmenorrhea-related topics. We constructed two WAG models respectively from queries by women of reproductive age and bymen. Salient themes were identified through inspecting clusters of WAG models. We identified two subsets of queries: Subset 1 contained 507,327 queries from women aged 13-50 years. Subset 2 contained 113,888 queries from men aged 13 or above. WAG modeling revealed topic clusters for each subset. Between female and male subsets, topic clusters overlapped on dysmenorrhea symptoms and management. Among female queries, there were distinctive topics on approaching menstrual pain at school and menstrual pain-related conditions; while among male queries, there was a distinctive cluster of queries on menstrual pain from male's perspectives. Big Data mining of the ChaCha ® question and answer service revealed a series of information needs among women and men on menstrual pain. Findings may be useful in structuring the content and informing the delivery platform for educational interventions.
Investing in Our Nation's Youth. National Youth Anti-Drug Media Campaign: Phase II (Final Report).
ERIC Educational Resources Information Center
Office of National Drug Control Policy, Washington, DC.
This publication presents the findings from an evaluation of Phase II of the National Youth Anti-Drug Media Campaign. The number one goal of the campaign was to educate youth to reject illegal drugs. This report evaluates Phase II and focuses on the effect of paid television advertising on awareness of anti-drug messages among youth, teens, and…
Labeled carbon dioxide (C18O2): an indicator gas for phase II in expirograms.
Schulz, Holger; Schulz, Anne; Eder, Gunter; Heyder, Joachim
2004-11-01
Carbon dioxide labeled with 18O (C18O2) was used as a tracer gas for single-breath measurements in six anesthetized, mechanically ventilated beagle dogs. C18O2 is taken up quasi-instantaneously in the gas-exchanging region of the lungs but much less so in the conducting airways. Its use allows a clear separation of phase II in an expirogram even from diseased individuals and excludes the influence of alveolar concentration differences. Phase II of a C18O2 expirogram mathematically corresponds to the cumulative distribution of bronchial pathways to be traversed completely in the course of exhalation. The derivative of this cumulative distribution with respect to respired volume was submitted to a power moment analysis to characterize volumetric mean (position), standard deviation (broadness), and skewness (asymmetry) of phase II. Position is an estimate of dead space volume, whereas broadness and skewness are measures of the range and asymmetry of functional airway pathway lengths. The effects of changing ventilatory patterns and of changes in airway size (via carbachol-induced bronchoconstriction) were studied. Increasing inspiratory or expiratory flow rates or tidal volume had only minor influence on position and shape of phase II. With the introduction of a postinspiratory breath hold, phase II was continually shifted toward the airway opening (maximum 45% at 16 s) and became steeper by up to 16%, whereas skewness showed a biphasic response with a moderate decrease at short breath holding and a significant increase at longer breath holds. Stepwise bronchoconstriction decreased position up to 45 +/- 2% and broadness of phase II up to 43 +/- 4%, whereas skewness was increased up to twofold at high-carbachol concentrations. Under all circumstances, position of phase II by power moment analysis and dead space volume by the Fowler technique agreed closely in our healthy dogs. Overall, power moment analysis provides a more comprehensive view on phase II of single-breath expirograms than conventional dead space volume determinations and may be useful for respiratory physiology studies as well as for the study of diseased lungs.
Generation of phase II in vitro metabolites using homogenized horse liver.
Wong, Jenny K Y; Chan, George H M; Leung, David K K; Tang, Francis P W; Wan, Terence S M
2016-02-01
The successful use of homogenized horse liver for the generation of phase I in vitro metabolites has been previously reported by the authors' laboratory. Prior to the use of homogenized liver, the authors' laboratory had been using mainly horse liver microsomes for carrying out equine in vitro metabolism studies. Homogenized horse liver has shown significant advantages over liver microsomes for in vitro metabolism studies as the procedures are much quicker and have higher capability for generating more in vitro metabolites. In this study, the use of homogenized liver has been extended to the generation of phase II in vitro metabolites (glucuronide and/or sulfate conjugates) using 17β-estradiol, morphine, and boldenone undecylenate as model substrates. It was observed that phase II metabolites could also be generated even without the addition of cofactors. To the authors' knowledge, this is the first report of the successful use of homogenized horse liver for the generation of phase II metabolites. It also demonstrates the ease with which both phase I and phase II metabolites can now be generated in vitro simply by using homogenized liver without the need for ultracentrifuges or tedious preparation steps. Copyright © 2015 John Wiley & Sons, Ltd.
Multiple Query Evaluation Based on an Enhanced Genetic Algorithm.
ERIC Educational Resources Information Center
Tamine, Lynda; Chrisment, Claude; Boughanem, Mohand
2003-01-01
Explains the use of genetic algorithms to combine results from multiple query evaluations to improve relevance in information retrieval. Discusses niching techniques, relevance feedback techniques, and evolution heuristics, and compares retrieval results obtained by both genetic multiple query evaluation and classical single query evaluation…
Relational Algebra and SQL: Better Together
ERIC Educational Resources Information Center
McMaster, Kirby; Sambasivam, Samuel; Hadfield, Steven; Wolthuis, Stuart
2013-01-01
In this paper, we describe how database instructors can teach Relational Algebra and Structured Query Language together through programming. Students write query programs consisting of sequences of Relational Algebra operations vs. Structured Query Language SELECT statements. The query programs can then be run interactively, allowing students to…
A Firefly Algorithm-based Approach for Pseudo-Relevance Feedback: Application to Medical Database.
Khennak, Ilyes; Drias, Habiba
2016-11-01
The difficulty of disambiguating the sense of the incomplete and imprecise keywords that are extensively used in the search queries has caused the failure of search systems to retrieve the desired information. One of the most powerful and promising method to overcome this shortcoming and improve the performance of search engines is Query Expansion, whereby the user's original query is augmented by new keywords that best characterize the user's information needs and produce more useful query. In this paper, a new Firefly Algorithm-based approach is proposed to enhance the retrieval effectiveness of query expansion while maintaining low computational complexity. In contrast to the existing literature, the proposed approach uses a Firefly Algorithm to find the best expanded query among a set of expanded query candidates. Moreover, this new approach allows the determination of the length of the expanded query empirically. Experimental results on MEDLINE, the on-line medical information database, show that our proposed approach is more effective and efficient compared to the state-of-the-art.
RiPPAS: A Ring-Based Privacy-Preserving Aggregation Scheme in Wireless Sensor Networks
Zhang, Kejia; Han, Qilong; Cai, Zhipeng; Yin, Guisheng
2017-01-01
Recently, data privacy in wireless sensor networks (WSNs) has been paid increased attention. The characteristics of WSNs determine that users’ queries are mainly aggregation queries. In this paper, the problem of processing aggregation queries in WSNs with data privacy preservation is investigated. A Ring-based Privacy-Preserving Aggregation Scheme (RiPPAS) is proposed. RiPPAS adopts ring structure to perform aggregation. It uses pseudonym mechanism for anonymous communication and uses homomorphic encryption technique to add noise to the data easily to be disclosed. RiPPAS can handle both sum() queries and min()/max() queries, while the existing privacy-preserving aggregation methods can only deal with sum() queries. For processing sum() queries, compared with the existing methods, RiPPAS has advantages in the aspects of privacy preservation and communication efficiency, which can be proved by theoretical analysis and simulation results. For processing min()/max() queries, RiPPAS provides effective privacy preservation and has low communication overhead. PMID:28178197
The development, deployment, and impact of the virtual observatory, Part II
NASA Astrophysics Data System (ADS)
Hanisch, R. J.
2015-06-01
This is the second special issue of Astronomy and Computing devoted to the Virtual Observatory, and we again see a combination of papers covering various aspects of the VO, from infrastructure to applications to programmatics. The critical role of data models is described by Louys, and the method by which applications communicate amongst each other through the Simple Applications Messaging Protocol (SAMP) is described by Taylor et al. Demleitner et al. explain the client interfaces to the VO registry, that is, how applications developers can query the registry for information about VO-compliant data collections and services.1
Phenotypic variation of Pseudomonas brassicacearum as a plant root-colonization strategy.
Achouak, Wafa; Conrod, Sandrine; Cohen, Valérie; Heulin, Thierry
2004-08-01
Pseudomonas brassicacearum was isolated as a major root-colonizing population from Arabidopsis thaliana. The strain NFM421 of P. brassicacearum undergoes phenotypic variation during A. thaliana and Brassica napus root colonization in vitro as well as in soil, resulting in different colony appearance on agar surfaces. Bacteria forming translucent colonies (phase II cells) essentially were localized at the surface of young roots and root tips, whereas wild-type cells (phase I cells) were localized at the basal part of roots. The ability of phase II cells to spread and colonize new sites on root surface correlates with over-production of flagellin as evidenced by sodium dodecyl sulfate-polyacrylamide gel electrophoresis analysis of surface proteins and microsequencing. Moreover, phase II cells showed a higher ability to swim and to swarm on semisolid agar medium. Phase I and phase II cells of P. brassicacearum NFM421 were tagged genetically with green fluorescent protein and red fluorescent protein. Confocal scanning laser microscopy was used to localize phase II cells on secondary roots and root tips of A. thaliana, whereas phase I cells essentially were localized at the basal part of roots. These experiments were conducted in vitro and in soil. Phenotypic variation on plant roots is likely to be a colonization strategy that may explain the high colonization power of P. brassicacearum.
Probability of success for phase III after exploratory biomarker analysis in phase II.
Götte, Heiko; Kirchner, Marietta; Sailer, Martin Oliver
2017-05-01
The probability of success or average power describes the potential of a future trial by weighting the power with a probability distribution of the treatment effect. The treatment effect estimate from a previous trial can be used to define such a distribution. During the development of targeted therapies, it is common practice to look for predictive biomarkers. The consequence is that the trial population for phase III is often selected on the basis of the most extreme result from phase II biomarker subgroup analyses. In such a case, there is a tendency to overestimate the treatment effect. We investigate whether the overestimation of the treatment effect estimate from phase II is transformed into a positive bias for the probability of success for phase III. We simulate a phase II/III development program for targeted therapies. This simulation allows to investigate selection probabilities and allows to compare the estimated with the true probability of success. We consider the estimated probability of success with and without subgroup selection. Depending on the true treatment effects, there is a negative bias without selection because of the weighting by the phase II distribution. In comparison, selection increases the estimated probability of success. Thus, selection does not lead to a bias in probability of success if underestimation due to the phase II distribution and overestimation due to selection cancel each other out. We recommend to perform similar simulations in practice to get the necessary information about the risk and chances associated with such subgroup selection designs. Copyright © 2017 John Wiley & Sons, Ltd.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Fujihisa, Hiroshi; Yamawaki, Hiroshi; Sakashita, Mami
2004-10-01
The structure of high pressure phases, selenium-II{sup '} (Se-II{sup '}) and sulfur-II (S-II), for {alpha}-Se{sub 8} (monoclinic Se-I) and {alpha}-S{sub 8} (orthorhombic S-I) was studied by powder x-ray diffraction experiments. Se-II{sup '} and S-II were found to be isostructural and to belong to the tetragonal space group I4{sub 1}/acd, which is made up of 16 atoms in the unit cell. The structure consisted of unique spiral chains with both 4{sub 1} and 4{sub 3} screws. The results confirmed that the structure sequence of the pressure-induced phase transitions for the group VIb elements depended on the initial molecular form. The chemicalmore » bonds of the phases are also discussed from the interatomic distances that were obtained.« less
Dynamic Querying of Mass-Storage RDF Data with Rule-Based Entailment Regimes
NASA Astrophysics Data System (ADS)
Ianni, Giovambattista; Krennwallner, Thomas; Martello, Alessandra; Polleres, Axel
RDF Schema (RDFS) as a lightweight ontology language is gaining popularity and, consequently, tools for scalable RDFS inference and querying are needed. SPARQL has become recently a W3C standard for querying RDF data, but it mostly provides means for querying simple RDF graphs only, whereas querying with respect to RDFS or other entailment regimes is left outside the current specification. In this paper, we show that SPARQL faces certain unwanted ramifications when querying ontologies in conjunction with RDF datasets that comprise multiple named graphs, and we provide an extension for SPARQL that remedies these effects. Moreover, since RDFS inference has a close relationship with logic rules, we generalize our approach to select a custom ruleset for specifying inferences to be taken into account in a SPARQL query. We show that our extensions are technically feasible by providing benchmark results for RDFS querying in our prototype system GiaBATA, which uses Datalog coupled with a persistent Relational Database as a back-end for implementing SPARQL with dynamic rule-based inference. By employing different optimization techniques like magic set rewriting our system remains competitive with state-of-the-art RDFS querying systems.
Mining the SDSS SkyServer SQL queries log
NASA Astrophysics Data System (ADS)
Hirota, Vitor M.; Santos, Rafael; Raddick, Jordan; Thakar, Ani
2016-05-01
SkyServer, the Internet portal for the Sloan Digital Sky Survey (SDSS) astronomic catalog, provides a set of tools that allows data access for astronomers and scientific education. One of SkyServer data access interfaces allows users to enter ad-hoc SQL statements to query the catalog. SkyServer also presents some template queries that can be used as basis for more complex queries. This interface has logged over 330 million queries submitted since 2001. It is expected that analysis of this data can be used to investigate usage patterns, identify potential new classes of queries, find similar queries, etc. and to shed some light on how users interact with the Sloan Digital Sky Survey data and how scientists have adopted the new paradigm of e-Science, which could in turn lead to enhancements on the user interfaces and experience in general. In this paper we review some approaches to SQL query mining, apply the traditional techniques used in the literature and present lessons learned, namely, that the general text mining approach for feature extraction and clustering does not seem to be adequate for this type of data, and, most importantly, we find that this type of analysis can result in very different queries being clustered together.
Applying Query Structuring in Cross-language Retrieval.
ERIC Educational Resources Information Center
Pirkola, Ari; Puolamaki, Deniz; Jarvelin, Kalervo
2003-01-01
Explores ways to apply query structuring in cross-language information retrieval. Tested were: English queries translated into Finnish using an electronic dictionary, and run in a Finnish newspaper databases; effects of compound-based structuring using a proximity operator for translation equivalents of query language compound components; and a…
Querying and Ranking XML Documents.
ERIC Educational Resources Information Center
Schlieder, Torsten; Meuss, Holger
2002-01-01
Discussion of XML, information retrieval, precision, and recall focuses on a retrieval technique that adopts the similarity measure of the vector space model, incorporates the document structure, and supports structured queries. Topics include a query model based on tree matching; structured queries and term-based ranking; and term frequency and…
Advanced Query Formulation in Deductive Databases.
ERIC Educational Resources Information Center
Niemi, Timo; Jarvelin, Kalervo
1992-01-01
Discusses deductive databases and database management systems (DBMS) and introduces a framework for advanced query formulation for end users. Recursive processing is described, a sample extensional database is presented, query types are explained, and criteria for advanced query formulation from the end user's viewpoint are examined. (31…
A Semantic Graph Query Language
DOE Office of Scientific and Technical Information (OSTI.GOV)
Kaplan, I L
2006-10-16
Semantic graphs can be used to organize large amounts of information from a number of sources into one unified structure. A semantic query language provides a foundation for extracting information from the semantic graph. The graph query language described here provides a simple, powerful method for querying semantic graphs.
Harris, Daniel R.; Henderson, Darren W.; Kavuluru, Ramakanth; Stromberg, Arnold J.; Johnson, Todd R.
2015-01-01
We present a custom, Boolean query generator utilizing common-table expressions (CTEs) that is capable of scaling with big datasets. The generator maps user-defined Boolean queries, such as those interactively created in clinical-research and general-purpose healthcare tools, into SQL. We demonstrate the effectiveness of this generator by integrating our work into the Informatics for Integrating Biology and the Bedside (i2b2) query tool and show that it is capable of scaling. Our custom generator replaces and outperforms the default query generator found within the Clinical Research Chart (CRC) cell of i2b2. In our experiments, sixteen different types of i2b2 queries were identified by varying four constraints: date, frequency, exclusion criteria, and whether selected concepts occurred in the same encounter. We generated non-trivial, random Boolean queries based on these 16 types; the corresponding SQL queries produced by both generators were compared by execution times. The CTE-based solution significantly outperformed the default query generator and provided a much more consistent response time across all query types (M=2.03, SD=6.64 vs. M=75.82, SD=238.88 seconds). Without costly hardware upgrades, we provide a scalable solution based on CTEs with very promising empirical results centered on performance gains. The evaluation methodology used for this provides a means of profiling clinical data warehouse performance. PMID:25192572
The Origins of [C ii] Emission in Local Star-forming Galaxies
DOE Office of Scientific and Technical Information (OSTI.GOV)
Croxall, K. V.; Smith, J. D.; Pellegrini, E.
The [C ii] 158 μ m fine-structure line is the brightest emission line observed in local star-forming galaxies. As a major coolant of the gas-phase interstellar medium, [C ii] balances the heating, including that due to far-ultraviolet photons, which heat the gas via the photoelectric effect. However, the origin of [C ii] emission remains unclear because C{sup +} can be found in multiple phases of the interstellar medium. Here we measure the fractions of [C ii] emission originating in the ionized and neutral gas phases of a sample of nearby galaxies. We use the [N ii] 205 μ m fine-structuremore » line to trace the ionized medium, thereby eliminating the strong density dependence that exists in the ratio of [C ii]/[N ii] 122 μ m. Using the FIR [C ii] and [N ii] emission detected by the KINGFISH (Key Insights on Nearby Galaxies: a Far- Infrared Survey with Herschel ) and Beyond the Peak Herschel programs, we show that 60%–80% of [C ii] emission originates from neutral gas. We find that the fraction of [C ii] originating in the neutral medium has a weak dependence on dust temperature and the surface density of star formation, and has a stronger dependence on the gas-phase metallicity. In metal-rich environments, the relatively cooler ionized gas makes substantially larger contributions to total [C ii] emission than at low abundance, contrary to prior expectations. Approximate calibrations of this metallicity trend are provided.« less
Query Language for Location-Based Services: A Model Checking Approach
NASA Astrophysics Data System (ADS)
Hoareau, Christian; Satoh, Ichiro
We present a model checking approach to the rationale, implementation, and applications of a query language for location-based services. Such query mechanisms are necessary so that users, objects, and/or services can effectively benefit from the location-awareness of their surrounding environment. The underlying data model is founded on a symbolic model of space organized in a tree structure. Once extended to a semantic model for modal logic, we regard location query processing as a model checking problem, and thus define location queries as hybrid logicbased formulas. Our approach is unique to existing research because it explores the connection between location models and query processing in ubiquitous computing systems, relies on a sound theoretical basis, and provides modal logic-based query mechanisms for expressive searches over a decentralized data structure. A prototype implementation is also presented and will be discussed.
Towards Building a High Performance Spatial Query System for Large Scale Medical Imaging Data.
Aji, Ablimit; Wang, Fusheng; Saltz, Joel H
2012-11-06
Support of high performance queries on large volumes of scientific spatial data is becoming increasingly important in many applications. This growth is driven by not only geospatial problems in numerous fields, but also emerging scientific applications that are increasingly data- and compute-intensive. For example, digital pathology imaging has become an emerging field during the past decade, where examination of high resolution images of human tissue specimens enables more effective diagnosis, prediction and treatment of diseases. Systematic analysis of large-scale pathology images generates tremendous amounts of spatially derived quantifications of micro-anatomic objects, such as nuclei, blood vessels, and tissue regions. Analytical pathology imaging provides high potential to support image based computer aided diagnosis. One major requirement for this is effective querying of such enormous amount of data with fast response, which is faced with two major challenges: the "big data" challenge and the high computation complexity. In this paper, we present our work towards building a high performance spatial query system for querying massive spatial data on MapReduce. Our framework takes an on demand index building approach for processing spatial queries and a partition-merge approach for building parallel spatial query pipelines, which fits nicely with the computing model of MapReduce. We demonstrate our framework on supporting multi-way spatial joins for algorithm evaluation and nearest neighbor queries for microanatomic objects. To reduce query response time, we propose cost based query optimization to mitigate the effect of data skew. Our experiments show that the framework can efficiently support complex analytical spatial queries on MapReduce.
Towards Building a High Performance Spatial Query System for Large Scale Medical Imaging Data
Aji, Ablimit; Wang, Fusheng; Saltz, Joel H.
2013-01-01
Support of high performance queries on large volumes of scientific spatial data is becoming increasingly important in many applications. This growth is driven by not only geospatial problems in numerous fields, but also emerging scientific applications that are increasingly data- and compute-intensive. For example, digital pathology imaging has become an emerging field during the past decade, where examination of high resolution images of human tissue specimens enables more effective diagnosis, prediction and treatment of diseases. Systematic analysis of large-scale pathology images generates tremendous amounts of spatially derived quantifications of micro-anatomic objects, such as nuclei, blood vessels, and tissue regions. Analytical pathology imaging provides high potential to support image based computer aided diagnosis. One major requirement for this is effective querying of such enormous amount of data with fast response, which is faced with two major challenges: the “big data” challenge and the high computation complexity. In this paper, we present our work towards building a high performance spatial query system for querying massive spatial data on MapReduce. Our framework takes an on demand index building approach for processing spatial queries and a partition-merge approach for building parallel spatial query pipelines, which fits nicely with the computing model of MapReduce. We demonstrate our framework on supporting multi-way spatial joins for algorithm evaluation and nearest neighbor queries for microanatomic objects. To reduce query response time, we propose cost based query optimization to mitigate the effect of data skew. Our experiments show that the framework can efficiently support complex analytical spatial queries on MapReduce. PMID:24501719
ERIC Educational Resources Information Center
Calfee, Robert; Calfee, Kathryn Hoover
The Beginning Teacher Evaluation Study (BTES), Phase II, was a research project on effective teaching behavior--what teachers do that significantly affects what and how pupils learn. The purposes of Phase II were to (1) develop an assessment system for measuring teacher and pupil behaviors and other factors which could influence each of them and…
Query Expansion and Query Translation as Logical Inference.
ERIC Educational Resources Information Center
Nie, Jian-Yun
2003-01-01
Examines query expansion during query translation in cross language information retrieval and develops a general framework for inferential information retrieval in two particular contexts: using fuzzy logic and probability theory. Obtains evaluation formulas that are shown to strongly correspond to those used in other information retrieval models.…
End-User Use of Data Base Query Language: Pros and Cons.
ERIC Educational Resources Information Center
Nicholes, Walter
1988-01-01
Man-machine interface, the concept of a computer "query," a review of database technology, and a description of the use of query languages at Brigham Young University are discussed. The pros and cons of end-user use of database query languages are explored. (Author/MLW)
Information Retrieval Using UMLS-based Structured Queries
Fagan, Lawrence M.; Berrios, Daniel C.; Chan, Albert; Cucina, Russell; Datta, Anupam; Shah, Maulik; Surendran, Sujith
2001-01-01
During the last three years, we have developed and described components of ELBook, a semantically based information-retrieval system [1-4]. Using these components, domain experts can specify a query model, indexers can use the query model to index documents, and end-users can search these documents for instances of indexed queries.
A Relational Algebra Query Language for Programming Relational Databases
ERIC Educational Resources Information Center
McMaster, Kirby; Sambasivam, Samuel; Anderson, Nicole
2011-01-01
In this paper, we describe a Relational Algebra Query Language (RAQL) and Relational Algebra Query (RAQ) software product we have developed that allows database instructors to teach relational algebra through programming. Instead of defining query operations using mathematical notation (the approach commonly taken in database textbooks), students…
Model Transformation for a System of Systems Dependability Safety Case
NASA Technical Reports Server (NTRS)
Murphy, Judy; Driskell, Steve
2011-01-01
The presentation reviews the dependability and safety effort of NASA's Independent Verification and Validation Facility. Topics include: safety engineering process, applications to non-space environment, Phase I overview, process creation, sample SRM artifact, Phase I end result, Phase II model transformation, fault management, and applying Phase II to individual projects.
Early Restoration | NOAA Gulf Spill Restoration
Early Restoration Plan. On April 20, 2011 we reached an agreement with BP to start restoration planning draft plan for the third phase of early restoration in December 2013. We are considering your comments : All Phase III information and documents Phase II Useful Links: Phase II Early Restoration Plan &
Effectiveness of safety belt warning and interlock systems
DOT National Transportation Integrated Search
1973-04-01
Rental cars in Fayetteville, N.C., were equipped with four seat belt and warning systems: (Phase I) detachable shoulder and lap belt, no warning system; (Phase II) detachable shoulder and lap belt, warning system (January 1, 1972 standard); (Phase II...
Obayashi, Takeshi; Kinoshita, Kengo
2010-05-01
Gene coexpression analyses are a powerful method to predict the function of genes and/or to identify genes that are functionally related to query genes. The basic idea of gene coexpression analyses is that genes with similar functions should have similar expression patterns under many different conditions. This approach is now widely used by many experimental researchers, especially in the field of plant biology. In this review, we will summarize recent successful examples obtained by using our gene coexpression database, ATTED-II. Specifically, the examples will describe the identification of new genes, such as the subunits of a complex protein, the enzymes in a metabolic pathway and transporters. In addition, we will discuss the discovery of a new intercellular signaling factor and new regulatory relationships between transcription factors and their target genes. In ATTED-II, we provide two basic views of gene coexpression, a gene list view and a gene network view, which can be used as guide gene approach and narrow-down approach, respectively. In addition, we will discuss the coexpression effectiveness for various types of gene sets.
An Ensemble Approach for Expanding Queries
2012-11-01
0.39 pain^0.39 Hospital 15094 0.82 hospital^0.82 Miscarriage 45 3.35 miscarriage ^3.35 Radiotherapy 53 3.28 radiotherapy^3.28 Hypoaldosteronism 3...negated query is the expansion of the original query with negation terms preceding each word. For example, the negated version of “ miscarriage ^3.35...includes “no miscarriage ”^3.35 and “not miscarriage ”^3.35. If a document is the result of both original query and negated query, its score is
Phase-II trials in osteosarcoma recurrences: A systematic review of past experience.
Omer, Natacha; Le Deley, Marie-Cécile; Piperno-Neumann, Sophie; Marec-Berard, Perrine; Italiano, Antoine; Corradini, Nadège; Bellera, Carine; Brugières, Laurence; Gaspar, Nathalie
2017-04-01
The most appropriate design of Phase-II trials evaluating new therapies in osteosarcoma remains poorly defined. To study consistency in phase-II clinical trials evaluating new therapies for osteosarcoma recurrences with respect to eligibility criteria, response assessment, end-points, statistical design and reported results. Systematic review of clinical trials registered on clinicaltrials.gov, clinicaltrialsregister.eu and French National Cancer Institute website or referenced in PubMed and American Society of Clinical Oncology websites, between 2003 and 2016, using the following criteria: (osteosarcoma OR bone sarcoma) AND (Phase-II). Among the 99 trials identified, 80 were Phase-II, 17 I/II and 2 II/III, evaluating mostly targeted therapy (n = 40), and chemotherapy alone (n = 26). Results were fully (n = 28) or partially (abstract, n = 6) published. Twenty-four trials were dedicated to osteosarcoma, 22 had an osteosarcoma stratum. Twenty-eight out of 99 trials refer to the age range observed at recurrence (28%). Overall, 65 trials were run in multicentre settings, including 17 international trials. Only 9 trials were randomised. The primary end-point was tumour response in 71 trials (response rate, n = 40 or best response, n = 31), with various definitions (complete + partial ± minor response and stable disease), mainly evaluated with RECIST criteria (n = 69); it was progression-free survival in 24 trials and OS in 3. In single-arm trials evaluating response rate, the null hypothesis tested (when available, n = 12) varied from 5% to 25%. No robust historical data can currently be derived from past efficacy Phase-II trials. There is an urgent need to develop international randomised Phase-II trials across all age ranges with standardised primary end-point. Copyright © 2017 Elsevier Ltd. All rights reserved.
CFGP: a web-based, comparative fungal genomics platform.
Park, Jongsun; Park, Bongsoo; Jung, Kyongyong; Jang, Suwang; Yu, Kwangyul; Choi, Jaeyoung; Kong, Sunghyung; Park, Jaejin; Kim, Seryun; Kim, Hyojeong; Kim, Soonok; Kim, Jihyun F; Blair, Jaime E; Lee, Kwangwon; Kang, Seogchan; Lee, Yong-Hwan
2008-01-01
Since the completion of the Saccharomyces cerevisiae genome sequencing project in 1996, the genomes of over 80 fungal species have been sequenced or are currently being sequenced. Resulting data provide opportunities for studying and comparing fungal biology and evolution at the genome level. To support such studies, the Comparative Fungal Genomics Platform (CFGP; http://cfgp.snu.ac.kr), a web-based multifunctional informatics workbench, was developed. The CFGP comprises three layers, including the basal layer, middleware and the user interface. The data warehouse in the basal layer contains standardized genome sequences of 65 fungal species. The middleware processes queries via six analysis tools, including BLAST, ClustalW, InterProScan, SignalP 3.0, PSORT II and a newly developed tool named BLASTMatrix. The BLASTMatrix permits the identification and visualization of genes homologous to a query across multiple species. The Data-driven User Interface (DUI) of the CFGP was built on a new concept of pre-collecting data and post-executing analysis instead of the 'fill-in-the-form-and-press-SUBMIT' user interfaces utilized by most bioinformatics sites. A tool termed Favorite, which supports the management of encapsulated sequence data and provides a personalized data repository to users, is another novel feature in the DUI.
DMINDA: an integrated web server for DNA motif identification and analyses.
Ma, Qin; Zhang, Hanyuan; Mao, Xizeng; Zhou, Chuan; Liu, Bingqiang; Chen, Xin; Xu, Ying
2014-07-01
DMINDA (DNA motif identification and analyses) is an integrated web server for DNA motif identification and analyses, which is accessible at http://csbl.bmb.uga.edu/DMINDA/. This web site is freely available to all users and there is no login requirement. This server provides a suite of cis-regulatory motif analysis functions on DNA sequences, which are important to elucidation of the mechanisms of transcriptional regulation: (i) de novo motif finding for a given set of promoter sequences along with statistical scores for the predicted motifs derived based on information extracted from a control set, (ii) scanning motif instances of a query motif in provided genomic sequences, (iii) motif comparison and clustering of identified motifs, and (iv) co-occurrence analyses of query motifs in given promoter sequences. The server is powered by a backend computer cluster with over 150 computing nodes, and is particularly useful for motif prediction and analyses in prokaryotic genomes. We believe that DMINDA, as a new and comprehensive web server for cis-regulatory motif finding and analyses, will benefit the genomic research community in general and prokaryotic genome researchers in particular. © The Author(s) 2014. Published by Oxford University Press on behalf of Nucleic Acids Research.
A novel adaptive Cuckoo search for optimal query plan generation.
Gomathi, Ramalingam; Sharmila, Dhandapani
2014-01-01
The emergence of multiple web pages day by day leads to the development of the semantic web technology. A World Wide Web Consortium (W3C) standard for storing semantic web data is the resource description framework (RDF). To enhance the efficiency in the execution time for querying large RDF graphs, the evolving metaheuristic algorithms become an alternate to the traditional query optimization methods. This paper focuses on the problem of query optimization of semantic web data. An efficient algorithm called adaptive Cuckoo search (ACS) for querying and generating optimal query plan for large RDF graphs is designed in this research. Experiments were conducted on different datasets with varying number of predicates. The experimental results have exposed that the proposed approach has provided significant results in terms of query execution time. The extent to which the algorithm is efficient is tested and the results are documented.
Query-Based Outlier Detection in Heterogeneous Information Networks.
Kuck, Jonathan; Zhuang, Honglei; Yan, Xifeng; Cam, Hasan; Han, Jiawei
2015-03-01
Outlier or anomaly detection in large data sets is a fundamental task in data science, with broad applications. However, in real data sets with high-dimensional space, most outliers are hidden in certain dimensional combinations and are relative to a user's search space and interest. It is often more effective to give power to users and allow them to specify outlier queries flexibly, and the system will then process such mining queries efficiently. In this study, we introduce the concept of query-based outlier in heterogeneous information networks, design a query language to facilitate users to specify such queries flexibly, define a good outlier measure in heterogeneous networks, and study how to process outlier queries efficiently in large data sets. Our experiments on real data sets show that following such a methodology, interesting outliers can be defined and uncovered flexibly and effectively in large heterogeneous networks.
Query-Based Outlier Detection in Heterogeneous Information Networks
Kuck, Jonathan; Zhuang, Honglei; Yan, Xifeng; Cam, Hasan; Han, Jiawei
2015-01-01
Outlier or anomaly detection in large data sets is a fundamental task in data science, with broad applications. However, in real data sets with high-dimensional space, most outliers are hidden in certain dimensional combinations and are relative to a user’s search space and interest. It is often more effective to give power to users and allow them to specify outlier queries flexibly, and the system will then process such mining queries efficiently. In this study, we introduce the concept of query-based outlier in heterogeneous information networks, design a query language to facilitate users to specify such queries flexibly, define a good outlier measure in heterogeneous networks, and study how to process outlier queries efficiently in large data sets. Our experiments on real data sets show that following such a methodology, interesting outliers can be defined and uncovered flexibly and effectively in large heterogeneous networks. PMID:27064397
Querying and Extracting Timeline Information from Road Traffic Sensor Data
Imawan, Ardi; Indikawati, Fitri Indra; Kwon, Joonho; Rao, Praveen
2016-01-01
The escalation of traffic congestion in urban cities has urged many countries to use intelligent transportation system (ITS) centers to collect historical traffic sensor data from multiple heterogeneous sources. By analyzing historical traffic data, we can obtain valuable insights into traffic behavior. Many existing applications have been proposed with limited analysis results because of the inability to cope with several types of analytical queries. In this paper, we propose the QET (querying and extracting timeline information) system—a novel analytical query processing method based on a timeline model for road traffic sensor data. To address query performance, we build a TQ-index (timeline query-index) that exploits spatio-temporal features of timeline modeling. We also propose an intuitive timeline visualization method to display congestion events obtained from specified query parameters. In addition, we demonstrate the benefit of our system through a performance evaluation using a Busan ITS dataset and a Seattle freeway dataset. PMID:27563900
Policy Compliance of Queries for Private Information Retrieval
2010-11-01
SPARQL, unfortunately, is not in RDF and so we had to develop tools to translate SPARQL queries into RDF to be used by our policy compliance prototype...policy-assurance/sparql2n3.py) that accepts SPARQL queries and returns the translated query in our simplified ontology. An example of a translated
Knowledge Query Language (KQL)
2016-02-12
Lexington Massachusetts This page intentionally left blank. iii EXECUTIVE SUMMARY Currently, queries for data ...retrieval from non-Structured Query Language (NoSQL) data stores are tightly coupled to the specific implementation of the data store implementation...independent of the storage content and format for querying NoSQL or relational data stores. This approach uses address expressions (or A-Expressions
Fragger: a protein fragment picker for structural queries.
Berenger, Francois; Simoncini, David; Voet, Arnout; Shrestha, Rojan; Zhang, Kam Y J
2017-01-01
Protein modeling and design activities often require querying the Protein Data Bank (PDB) with a structural fragment, possibly containing gaps. For some applications, it is preferable to work on a specific subset of the PDB or with unpublished structures. These requirements, along with specific user needs, motivated the creation of a new software to manage and query 3D protein fragments. Fragger is a protein fragment picker that allows protein fragment databases to be created and queried. All fragment lengths are supported and any set of PDB files can be used to create a database. Fragger can efficiently search a fragment database with a query fragment and a distance threshold. Matching fragments are ranked by distance to the query. The query fragment can have structural gaps and the allowed amino acid sequences matching a query can be constrained via a regular expression of one-letter amino acid codes. Fragger also incorporates a tool to compute the backbone RMSD of one versus many fragments in high throughput. Fragger should be useful for protein design, loop grafting and related structural bioinformatics tasks.
NASA Astrophysics Data System (ADS)
Skotniczny, Zbigniew
1989-12-01
The Query by Forms (QbF) system is a user-oriented interactive tool for querying large relational database with minimal queries difinition cost. The system was worked out under the assumption that user's time and effort for defining needed queries is the most severe bottleneck. The system may be applied in any Rdb/VMS databases system and is recommended for specific information systems of any project where end-user queries cannot be foreseen. The tool is dedicated to specialist of an application domain who have to analyze data maintained in database from any needed point of view, who do not need to know commercial databases languages. The paper presents the system developed as a compromise between its functionality and usability. User-system communication via a menu-driven "tree-like" structure of screen-forms which produces a query difinition and execution is discussed in detail. Output of query results (printed reports and graphics) is also discussed. Finally the paper shows one application of QbF to a HERA-project.
White, Ryen W; Horvitz, Eric
2014-01-01
Objective To better understand the relationship between online health-seeking behaviors and in-world healthcare utilization (HU) by studies of online search and access activities before and after queries that pursue medical professionals and facilities. Materials and methods We analyzed data collected from logs of online searches gathered from consenting users of a browser toolbar from Microsoft (N=9740). We employed a complementary survey (N=489) to seek a deeper understanding of information-gathering, reflection, and action on the pursuit of professional healthcare. Results We provide insights about HU through the survey, breaking out its findings by different respondent marginalizations as appropriate. Observations made from search logs may be explained by trends observed in our survey responses, even though the user populations differ. Discussion The results provide insights about how users decide if and when to utilize healthcare resources, and how online health information seeking transitions to in-world HU. The findings from both the survey and the logs reveal behavioral patterns and suggest a strong relationship between search behavior and HU. Although the diversity of our survey respondents is limited and we cannot be certain that users visited medical facilities, we demonstrate that it may be possible to infer HU from long-term search behavior by the apparent influence that health concerns and professional advice have on search activity. Conclusions Our findings highlight different phases of online activities around queries pursuing professional healthcare facilities and services. We also show that it may be possible to infer HU from logs without tracking people's physical location, based on the effect of HU on pre- and post-HU search behavior. This allows search providers and others to develop more robust models of interests and preferences by modeling utilization rather than simply the intention to utilize that is expressed in search queries. PMID:23666794
Reflective Cracking of Flexible Pavements Phase I and II Final Recommendations
DOT National Transportation Integrated Search
2008-02-02
This report summarizes all the findings and recommendations from the Phase I and Phase II of the Nevada Department of Transportation (NDOT) study initiated in 2006 to mitigate reflective cracking in hot mix asphalt (HMA) overlays. Based on the analys...
2006-08-01
effective for describing taxonomic categories and properties of things, the structures found in SWRL and SPARQL are better suited to describing conditions...up the query processing time, which may occur many times and furthermore it is time critical. In order to maintain information about the...that time spent during this phase does not depend linearly on the number of concepts present in the data structure , but in the order of log of concepts
Hybrid ontology for semantic information retrieval model using keyword matching indexing system.
Uthayan, K R; Mala, G S Anandha
2015-01-01
Ontology is the process of growth and elucidation of concepts of an information domain being common for a group of users. Establishing ontology into information retrieval is a normal method to develop searching effects of relevant information users require. Keywords matching process with historical or information domain is significant in recent calculations for assisting the best match for specific input queries. This research presents a better querying mechanism for information retrieval which integrates the ontology queries with keyword search. The ontology-based query is changed into a primary order to predicate logic uncertainty which is used for routing the query to the appropriate servers. Matching algorithms characterize warm area of researches in computer science and artificial intelligence. In text matching, it is more dependable to study semantics model and query for conditions of semantic matching. This research develops the semantic matching results between input queries and information in ontology field. The contributed algorithm is a hybrid method that is based on matching extracted instances from the queries and information field. The queries and information domain is focused on semantic matching, to discover the best match and to progress the executive process. In conclusion, the hybrid ontology in semantic web is sufficient to retrieve the documents when compared to standard ontology.
Hybrid Ontology for Semantic Information Retrieval Model Using Keyword Matching Indexing System
Uthayan, K. R.; Anandha Mala, G. S.
2015-01-01
Ontology is the process of growth and elucidation of concepts of an information domain being common for a group of users. Establishing ontology into information retrieval is a normal method to develop searching effects of relevant information users require. Keywords matching process with historical or information domain is significant in recent calculations for assisting the best match for specific input queries. This research presents a better querying mechanism for information retrieval which integrates the ontology queries with keyword search. The ontology-based query is changed into a primary order to predicate logic uncertainty which is used for routing the query to the appropriate servers. Matching algorithms characterize warm area of researches in computer science and artificial intelligence. In text matching, it is more dependable to study semantics model and query for conditions of semantic matching. This research develops the semantic matching results between input queries and information in ontology field. The contributed algorithm is a hybrid method that is based on matching extracted instances from the queries and information field. The queries and information domain is focused on semantic matching, to discover the best match and to progress the executive process. In conclusion, the hybrid ontology in semantic web is sufficient to retrieve the documents when compared to standard ontology. PMID:25922851
Multidimensional indexing structure for use with linear optimization queries
NASA Technical Reports Server (NTRS)
Bergman, Lawrence David (Inventor); Castelli, Vittorio (Inventor); Chang, Yuan-Chi (Inventor); Li, Chung-Sheng (Inventor); Smith, John Richard (Inventor)
2002-01-01
Linear optimization queries, which usually arise in various decision support and resource planning applications, are queries that retrieve top N data records (where N is an integer greater than zero) which satisfy a specific optimization criterion. The optimization criterion is to either maximize or minimize a linear equation. The coefficients of the linear equation are given at query time. Methods and apparatus are disclosed for constructing, maintaining and utilizing a multidimensional indexing structure of database records to improve the execution speed of linear optimization queries. Database records with numerical attributes are organized into a number of layers and each layer represents a geometric structure called convex hull. Such linear optimization queries are processed by searching from the outer-most layer of this multi-layer indexing structure inwards. At least one record per layer will satisfy the query criterion and the number of layers needed to be searched depends on the spatial distribution of records, the query-issued linear coefficients, and N, the number of records to be returned. When N is small compared to the total size of the database, answering the query typically requires searching only a small fraction of all relevant records, resulting in a tremendous speedup as compared to linearly scanning the entire dataset.
The role of economics in the QUERI program: QUERI Series
Smith, Mark W; Barnett, Paul G
2008-01-01
Background The United States (U.S.) Department of Veterans Affairs (VA) Quality Enhancement Research Initiative (QUERI) has implemented economic analyses in single-site and multi-site clinical trials. To date, no one has reviewed whether the QUERI Centers are taking an optimal approach to doing so. Consistent with the continuous learning culture of the QUERI Program, this paper provides such a reflection. Methods We present a case study of QUERI as an example of how economic considerations can and should be integrated into implementation research within both single and multi-site studies. We review theoretical and applied cost research in implementation studies outside and within VA. We also present a critique of the use of economic research within the QUERI program. Results Economic evaluation is a key element of implementation research. QUERI has contributed many developments in the field of implementation but has only recently begun multi-site implementation trials across multiple regions within the national VA healthcare system. These trials are unusual in their emphasis on developing detailed costs of implementation, as well as in the use of business case analyses (budget impact analyses). Conclusion Economics appears to play an important role in QUERI implementation studies, only after implementation has reached the stage of multi-site trials. Economic analysis could better inform the choice of which clinical best practices to implement and the choice of implementation interventions to employ. QUERI economics also would benefit from research on costing methods and development of widely accepted international standards for implementation economics. PMID:18430199
The role of economics in the QUERI program: QUERI Series.
Smith, Mark W; Barnett, Paul G
2008-04-22
The United States (U.S.) Department of Veterans Affairs (VA) Quality Enhancement Research Initiative (QUERI) has implemented economic analyses in single-site and multi-site clinical trials. To date, no one has reviewed whether the QUERI Centers are taking an optimal approach to doing so. Consistent with the continuous learning culture of the QUERI Program, this paper provides such a reflection. We present a case study of QUERI as an example of how economic considerations can and should be integrated into implementation research within both single and multi-site studies. We review theoretical and applied cost research in implementation studies outside and within VA. We also present a critique of the use of economic research within the QUERI program. Economic evaluation is a key element of implementation research. QUERI has contributed many developments in the field of implementation but has only recently begun multi-site implementation trials across multiple regions within the national VA healthcare system. These trials are unusual in their emphasis on developing detailed costs of implementation, as well as in the use of business case analyses (budget impact analyses). Economics appears to play an important role in QUERI implementation studies, only after implementation has reached the stage of multi-site trials. Economic analysis could better inform the choice of which clinical best practices to implement and the choice of implementation interventions to employ. QUERI economics also would benefit from research on costing methods and development of widely accepted international standards for implementation economics.
Prostate Cancer Clinical Trials Group: The University of Michigan Site
2012-04-01
and fusion-negative strata. UM will be the lead site for this trial with the Univ. of Chicago N01 Phase II consortium as the coordinating center. Ten...sensitive prostate cancer: a University of Chicago Phase II Consortium/Department of Defense Prostate Cancer Clinical Trials Consortium study. JE Ward, T...N01 contract with CTEP (University of Chicago – Early Therapeutics Development with Phase II emphasis group). The Program is committed to creating
Fleming, Richard M
2002-01-01
Over 60% of Americans are overweight and a number of popular diets have been advocated, often without evidence, to alleviate this public health hazard. This study was designed to investigate the effects of several diets on weight loss, serum lipids, and other cardiovascular disease risk factors. One hundred men and women followed one of four dietary programs for 1 year: a moderate-fat (MF) program without calorie restriction (28 patients); a low-fat (LF) diet (phase I) (16) ; a MF, calorie-controlled (phase II) diet (38 patients); and a high-fat (HF) diet (18 subjects) [corrected]. Weight, total cholesterol (TC), low-density lipoprotein cholesterol (LDL-C), high-density lipoprotein cholesterol (HDL-C), triglycerides (TG), homocysteine (Ho), and lipoprotein(a) [Lp(a)], were measured every 4th month. The TC/HDL-C ratio was calculated and fibrinogen levels were measured at baseline and after one year. The MF diet resulted in a 2.6% (NS) decrease in weight compared with 18.4% (p=0.045) decrease in patients on phase I, 12.6% (p=0.0085) decrease in patients on phase II, and 13.7% (p=0.025) decrease in those on the HF diet. TC was reduced by 5% (NS) in the MF group, 39.1% (p=0.0005) in the phase I group, and 30.4% (p=0.0001) in the phase II group. HF group had a 4.3% (NS) increase in TC. LDL-C was reduced by 6.1% (NS) on MF, 52.0% (p=0.0001) on phase I, and 38.8% (p=0.0001) on phase II. Patients on HF had a 6.0% (NS) increase in LDL-C. There were nonsignificant reductions in HDL-C in those on MF (-1.5%) and HF (-5.8%). Patients on phase I showed an increase in HDL-C of 9.0% (NS), while those on phase II diet had a 3.6% increase (NS) in HDL-C. TC/HDL-C increased (9.8%) only in patients following the high-fat diets (NS). Patients on MF had a 5.3% (NS) reduction in TC/HDL-C, while those on LF had significant reductions on the phase I ( -45.8%; p=0.0001) diet and phase II diet (-34.7%; p=0.0001). TG levels increased on both the MF (1.0%) and HF (5.5%) diets, although neither was statistically significant. People following the phase I and II diets showed reductions of 37.3% and 36.9%, respectively. Ho levels increased by 9.7% when people followed the MF diet and by 12.4% when they followed the HF diet. Patients following the phase I and phase II diets showed reductions of 13.6% and 14.6%, respectively. Only those following phase II diets showed a tendency toward significant improvement (p=0.061). Lp(a) levels increased by 4.7% following the MF (NS) diet and by 31.0% (NS) on the HF diet. Patients following phase I showed a 7.4% (NS) reduction and a 10.8% reduction (NS) following phase II. Fibrinogen levels increased only in individuals following HF diets (11.9%), while patients following MF (-0.6%), phase I (-11.0%), and phase II (-6.3%) diets showed nonsignificant reductions in fibrinogen. Patients on MF demonstrated nonsignificant reductions in weight, LDL-C, TC, HDL-C, TC/HDL-C ratios, and fibrinogen and nonsignificant increases in TGs, Lp(a), and homocysteine. There was significant weight loss in patients on phase I and II and HF diets after 1 year. Reductions in TC, LDL-C, TGs, and TC/HDL ratios were significant only in patients either following a LF diet or a MF, calorically reduced diet. Only patients following HF diets showed a worsening of each cardiovascular disease risk factor (LDL-C, TG, TC, HDL-C, TC/HDL ratio, Ho, Lp(a), and fibrinogen), despite achieving statistically significant weight loss. Copyright 2002 CHF, Inc.
Research safety vehicle, Phase II. Volume I. Executive summary. Final report jul 75-dec 76
DOE Office of Scientific and Technical Information (OSTI.GOV)
Struble, D.
1976-12-01
Volume I summarizes the results of the Minicars Research Safety Vehicle Phase II program, as detailed in Volumes II and III. Phase I identified trends leading to the desired national social goals of the mid-1980's in vehicle crashworthiness, crash avoidance, damageability, pedestrian safety, fuel economy, emissions and cost, and characterized an RSV to satisfy them. In Phase II an RSV prototype was designed, developed and tested to demonstrate the feasibility of meeting these goals simultaneously. Although further refinement is necessary to assure operational validity, in all categories the results meet or exceed the most advanced performance specified by The Presidentialmore » Task Force on Motor Vehicle Goals beyond 1980.« less
The use of dihexyldithiocarbamate in reverse-phase HPLC of metal chelates
NASA Astrophysics Data System (ADS)
Fatimah, S. S.; Bahti, H. H.; Hastiawan, I.; Permanasari, A.
2018-05-01
Dialkyldithiocarbamates have long been used as chelating agents in reverse-phase HPLC of transition metals. In the previous study, an alkyl homolog of this type of ligand, namely dihexyldithiocarbamate (DHDTC), was synthesized and characterized. The use of this particular ligand in the revese-phase HPLC of some selected transition metal ions is now reported for the first time. The mobile phase comprising of the flow rate and of the detection, in the separation of the metal chelates of Cd (II), Fe (III), Cu (II), and Co (III), were investigated on a C-18 column. The results showed that dihexylditiocarbamate could be used for separating Cd (II), Fe(III), Cu(II), and Co(III). Therefore, it could be used in simultaneous analysis.
Processing SPARQL queries with regular expressions in RDF databases
2011-01-01
Background As the Resource Description Framework (RDF) data model is widely used for modeling and sharing a lot of online bioinformatics resources such as Uniprot (dev.isb-sib.ch/projects/uniprot-rdf) or Bio2RDF (bio2rdf.org), SPARQL - a W3C recommendation query for RDF databases - has become an important query language for querying the bioinformatics knowledge bases. Moreover, due to the diversity of users’ requests for extracting information from the RDF data as well as the lack of users’ knowledge about the exact value of each fact in the RDF databases, it is desirable to use the SPARQL query with regular expression patterns for querying the RDF data. To the best of our knowledge, there is currently no work that efficiently supports regular expression processing in SPARQL over RDF databases. Most of the existing techniques for processing regular expressions are designed for querying a text corpus, or only for supporting the matching over the paths in an RDF graph. Results In this paper, we propose a novel framework for supporting regular expression processing in SPARQL query. Our contributions can be summarized as follows. 1) We propose an efficient framework for processing SPARQL queries with regular expression patterns in RDF databases. 2) We propose a cost model in order to adapt the proposed framework in the existing query optimizers. 3) We build a prototype for the proposed framework in C++ and conduct extensive experiments demonstrating the efficiency and effectiveness of our technique. Conclusions Experiments with a full-blown RDF engine show that our framework outperforms the existing ones by up to two orders of magnitude in processing SPARQL queries with regular expression patterns. PMID:21489225
Processing SPARQL queries with regular expressions in RDF databases.
Lee, Jinsoo; Pham, Minh-Duc; Lee, Jihwan; Han, Wook-Shin; Cho, Hune; Yu, Hwanjo; Lee, Jeong-Hoon
2011-03-29
As the Resource Description Framework (RDF) data model is widely used for modeling and sharing a lot of online bioinformatics resources such as Uniprot (dev.isb-sib.ch/projects/uniprot-rdf) or Bio2RDF (bio2rdf.org), SPARQL - a W3C recommendation query for RDF databases - has become an important query language for querying the bioinformatics knowledge bases. Moreover, due to the diversity of users' requests for extracting information from the RDF data as well as the lack of users' knowledge about the exact value of each fact in the RDF databases, it is desirable to use the SPARQL query with regular expression patterns for querying the RDF data. To the best of our knowledge, there is currently no work that efficiently supports regular expression processing in SPARQL over RDF databases. Most of the existing techniques for processing regular expressions are designed for querying a text corpus, or only for supporting the matching over the paths in an RDF graph. In this paper, we propose a novel framework for supporting regular expression processing in SPARQL query. Our contributions can be summarized as follows. 1) We propose an efficient framework for processing SPARQL queries with regular expression patterns in RDF databases. 2) We propose a cost model in order to adapt the proposed framework in the existing query optimizers. 3) We build a prototype for the proposed framework in C++ and conduct extensive experiments demonstrating the efficiency and effectiveness of our technique. Experiments with a full-blown RDF engine show that our framework outperforms the existing ones by up to two orders of magnitude in processing SPARQL queries with regular expression patterns.
Chen, R S; Nadkarni, P; Marenco, L; Levin, F; Erdos, J; Miller, P L
2000-01-01
The entity-attribute-value representation with classes and relationships (EAV/CR) provides a flexible and simple database schema to store heterogeneous biomedical data. In certain circumstances, however, the EAV/CR model is known to retrieve data less efficiently than conventionally based database schemas. To perform a pilot study that systematically quantifies performance differences for database queries directed at real-world microbiology data modeled with EAV/CR and conventional representations, and to explore the relative merits of different EAV/CR query implementation strategies. Clinical microbiology data obtained over a ten-year period were stored using both database models. Query execution times were compared for four clinically oriented attribute-centered and entity-centered queries operating under varying conditions of database size and system memory. The performance characteristics of three different EAV/CR query strategies were also examined. Performance was similar for entity-centered queries in the two database models. Performance in the EAV/CR model was approximately three to five times less efficient than its conventional counterpart for attribute-centered queries. The differences in query efficiency became slightly greater as database size increased, although they were reduced with the addition of system memory. The authors found that EAV/CR queries formulated using multiple, simple SQL statements executed in batch were more efficient than single, large SQL statements. This paper describes a pilot project to explore issues in and compare query performance for EAV/CR and conventional database representations. Although attribute-centered queries were less efficient in the EAV/CR model, these inefficiencies may be addressable, at least in part, by the use of more powerful hardware or more memory, or both.
Seo, Dong-Woo; Sohn, Chang Hwan; Kim, Sung-Hoon; Ryoo, Seung Mok; Lee, Yoon-Seon; Lee, Jae Ho; Kim, Won Young; Lim, Kyoung Soo
2016-01-01
Background Digital surveillance using internet search queries can improve both the sensitivity and timeliness of the detection of a health event, such as an influenza outbreak. While it has recently been estimated that the mobile search volume surpasses the desktop search volume and mobile search patterns differ from desktop search patterns, the previous digital surveillance systems did not distinguish mobile and desktop search queries. The purpose of this study was to compare the performance of mobile and desktop search queries in terms of digital influenza surveillance. Methods and Results The study period was from September 6, 2010 through August 30, 2014, which consisted of four epidemiological years. Influenza-like illness (ILI) and virologic surveillance data from the Korea Centers for Disease Control and Prevention were used. A total of 210 combined queries from our previous survey work were used for this study. Mobile and desktop weekly search data were extracted from Naver, which is the largest search engine in Korea. Spearman’s correlation analysis was used to examine the correlation of the mobile and desktop data with ILI and virologic data in Korea. We also performed lag correlation analysis. We observed that the influenza surveillance performance of mobile search queries matched or exceeded that of desktop search queries over time. The mean correlation coefficients of mobile search queries and the number of queries with an r-value of ≥ 0.7 equaled or became greater than those of desktop searches over the four epidemiological years. A lag correlation analysis of up to two weeks showed similar trends. Conclusion Our study shows that mobile search queries for influenza surveillance have equaled or even become greater than desktop search queries over time. In the future development of influenza surveillance using search queries, the recognition of changing trend of mobile search data could be necessary. PMID:27391028
Shin, Soo-Yong; Kim, Taerim; Seo, Dong-Woo; Sohn, Chang Hwan; Kim, Sung-Hoon; Ryoo, Seung Mok; Lee, Yoon-Seon; Lee, Jae Ho; Kim, Won Young; Lim, Kyoung Soo
2016-01-01
Digital surveillance using internet search queries can improve both the sensitivity and timeliness of the detection of a health event, such as an influenza outbreak. While it has recently been estimated that the mobile search volume surpasses the desktop search volume and mobile search patterns differ from desktop search patterns, the previous digital surveillance systems did not distinguish mobile and desktop search queries. The purpose of this study was to compare the performance of mobile and desktop search queries in terms of digital influenza surveillance. The study period was from September 6, 2010 through August 30, 2014, which consisted of four epidemiological years. Influenza-like illness (ILI) and virologic surveillance data from the Korea Centers for Disease Control and Prevention were used. A total of 210 combined queries from our previous survey work were used for this study. Mobile and desktop weekly search data were extracted from Naver, which is the largest search engine in Korea. Spearman's correlation analysis was used to examine the correlation of the mobile and desktop data with ILI and virologic data in Korea. We also performed lag correlation analysis. We observed that the influenza surveillance performance of mobile search queries matched or exceeded that of desktop search queries over time. The mean correlation coefficients of mobile search queries and the number of queries with an r-value of ≥ 0.7 equaled or became greater than those of desktop searches over the four epidemiological years. A lag correlation analysis of up to two weeks showed similar trends. Our study shows that mobile search queries for influenza surveillance have equaled or even become greater than desktop search queries over time. In the future development of influenza surveillance using search queries, the recognition of changing trend of mobile search data could be necessary.
Searching for cancer information on the internet: analyzing natural language search queries.
Bader, Judith L; Theofanos, Mary Frances
2003-12-11
Searching for health information is one of the most-common tasks performed by Internet users. Many users begin searching on popular search engines rather than on prominent health information sites. We know that many visitors to our (National Cancer Institute) Web site, cancer.gov, arrive via links in search engine result. To learn more about the specific needs of our general-public users, we wanted to understand what lay users really wanted to know about cancer, how they phrased their questions, and how much detail they used. The National Cancer Institute partnered with AskJeeves, Inc to develop a methodology to capture, sample, and analyze 3 months of cancer-related queries on the Ask.com Web site, a prominent United States consumer search engine, which receives over 35 million queries per week. Using a benchmark set of 500 terms and word roots supplied by the National Cancer Institute, AskJeeves identified a test sample of cancer queries for 1 week in August 2001. From these 500 terms only 37 appeared >or= 5 times/day over the trial test week in 17208 queries. Using these 37 terms, 204165 instances of cancer queries were found in the Ask.com query logs for the actual test period of June-August 2001. Of these, 7500 individual user questions were randomly selected for detailed analysis and assigned to appropriate categories. The exact language of sample queries is presented. Considering multiples of the same questions, the sample of 7500 individual user queries represented 76077 queries (37% of the total 3-month pool). Overall 78.37% of sampled Cancer queries asked about 14 specific cancer types. Within each cancer type, queries were sorted into appropriate subcategories including at least the following: General Information, Symptoms, Diagnosis and Testing, Treatment, Statistics, Definition, and Cause/Risk/Link. The most-common specific cancer types mentioned in queries were Digestive/Gastrointestinal/Bowel (15.0%), Breast (11.7%), Skin (11.3%), and Genitourinary (10.5%). Additional subcategories of queries about specific cancer types varied, depending on user input. Queries that were not specific to a cancer type were also tracked and categorized. Natural-language searching affords users the opportunity to fully express their information needs and can aid users naïve to the content and vocabulary. The specific queries analyzed for this study reflect news and research studies reported during the study dates and would surely change with different study dates. Analyzing queries from search engines represents one way of knowing what kinds of content to provide to users of a given Web site. Users ask questions using whole sentences and keywords, often misspelling words. Providing the option for natural-language searching does not obviate the need for good information architecture, usability engineering, and user testing in order to optimize user experience.
Searching for Cancer Information on the Internet: Analyzing Natural Language Search Queries
Theofanos, Mary Frances
2003-01-01
Background Searching for health information is one of the most-common tasks performed by Internet users. Many users begin searching on popular search engines rather than on prominent health information sites. We know that many visitors to our (National Cancer Institute) Web site, cancer.gov, arrive via links in search engine result. Objective To learn more about the specific needs of our general-public users, we wanted to understand what lay users really wanted to know about cancer, how they phrased their questions, and how much detail they used. Methods The National Cancer Institute partnered with AskJeeves, Inc to develop a methodology to capture, sample, and analyze 3 months of cancer-related queries on the Ask.com Web site, a prominent United States consumer search engine, which receives over 35 million queries per week. Using a benchmark set of 500 terms and word roots supplied by the National Cancer Institute, AskJeeves identified a test sample of cancer queries for 1 week in August 2001. From these 500 terms only 37 appeared ≥ 5 times/day over the trial test week in 17208 queries. Using these 37 terms, 204165 instances of cancer queries were found in the Ask.com query logs for the actual test period of June-August 2001. Of these, 7500 individual user questions were randomly selected for detailed analysis and assigned to appropriate categories. The exact language of sample queries is presented. Results Considering multiples of the same questions, the sample of 7500 individual user queries represented 76077 queries (37% of the total 3-month pool). Overall 78.37% of sampled Cancer queries asked about 14 specific cancer types. Within each cancer type, queries were sorted into appropriate subcategories including at least the following: General Information, Symptoms, Diagnosis and Testing, Treatment, Statistics, Definition, and Cause/Risk/Link. The most-common specific cancer types mentioned in queries were Digestive/Gastrointestinal/Bowel (15.0%), Breast (11.7%), Skin (11.3%), and Genitourinary (10.5%). Additional subcategories of queries about specific cancer types varied, depending on user input. Queries that were not specific to a cancer type were also tracked and categorized. Conclusions Natural-language searching affords users the opportunity to fully express their information needs and can aid users naïve to the content and vocabulary. The specific queries analyzed for this study reflect news and research studies reported during the study dates and would surely change with different study dates. Analyzing queries from search engines represents one way of knowing what kinds of content to provide to users of a given Web site. Users ask questions using whole sentences and keywords, often misspelling words. Providing the option for natural-language searching does not obviate the need for good information architecture, usability engineering, and user testing in order to optimize user experience. PMID:14713659
Optimized Probe Masking for Comparative Transcriptomics of Closely Related Species
Poeschl, Yvonne; Delker, Carolin; Trenner, Jana; Ullrich, Kristian Karsten; Quint, Marcel; Grosse, Ivo
2013-01-01
Microarrays are commonly applied to study the transcriptome of specific species. However, many available microarrays are restricted to model organisms, and the design of custom microarrays for other species is often not feasible. Hence, transcriptomics approaches of non-model organisms as well as comparative transcriptomics studies among two or more species often make use of cost-intensive RNAseq studies or, alternatively, by hybridizing transcripts of a query species to a microarray of a closely related species. When analyzing these cross-species microarray expression data, differences in the transcriptome of the query species can cause problems, such as the following: (i) lower hybridization accuracy of probes due to mismatches or deletions, (ii) probes binding multiple transcripts of different genes, and (iii) probes binding transcripts of non-orthologous genes. So far, methods for (i) exist, but these neglect (ii) and (iii). Here, we propose an approach for comparative transcriptomics addressing problems (i) to (iii), which retains only transcript-specific probes binding transcripts of orthologous genes. We apply this approach to an Arabidopsis lyrata expression data set measured on a microarray designed for Arabidopsis thaliana, and compare it to two alternative approaches, a sequence-based approach and a genomic DNA hybridization-based approach. We investigate the number of retained probe sets, and we validate the resulting expression responses by qRT-PCR. We find that the proposed approach combines the benefit of sequence-based stringency and accuracy while allowing the expression analysis of much more genes than the alternative sequence-based approach. As an added benefit, the proposed approach requires probes to detect transcripts of orthologous genes only, which provides a superior base for biological interpretation of the measured expression responses. PMID:24260119
Cai, Yu-Dong; Chou, Kuo-Chen
2011-01-01
Given a regulatory pathway system consisting of a set of proteins, can we predict which pathway class it belongs to? Such a problem is closely related to the biological function of the pathway in cells and hence is quite fundamental and essential in systems biology and proteomics. This is also an extremely difficult and challenging problem due to its complexity. To address this problem, a novel approach was developed that can be used to predict query pathways among the following six functional categories: (i) “Metabolism”, (ii) “Genetic Information Processing”, (iii) “Environmental Information Processing”, (iv) “Cellular Processes”, (v) “Organismal Systems”, and (vi) “Human Diseases”. The prediction method was established trough the following procedures: (i) according to the general form of pseudo amino acid composition (PseAAC), each of the pathways concerned is formulated as a 5570-D (dimensional) vector; (ii) each of components in the 5570-D vector was derived by a series of feature extractions from the pathway system according to its graphic property, biochemical and physicochemical property, as well as functional property; (iii) the minimum redundancy maximum relevance (mRMR) method was adopted to operate the prediction. A cross-validation by the jackknife test on a benchmark dataset consisting of 146 regulatory pathways indicated that an overall success rate of 78.8% was achieved by our method in identifying query pathways among the above six classes, indicating the outcome is quite promising and encouraging. To the best of our knowledge, the current study represents the first effort in attempting to identity the type of a pathway system or its biological function. It is anticipated that our report may stimulate a series of follow-up investigations in this new and challenging area. PMID:21980418
A two-stage patient enrichment adaptive design in phase II oncology trials.
Song, James X
2014-01-01
Illustrated is the use of a patient enrichment adaptive design in a randomized phase II trial which allows the evaluation of treatment benefits by the biomarker expression level and makes interim adjustment according to the pre-specified rules. The design was applied to an actual phase II metastatic hepatocellular carcinoma (HCC) trial in which progression-free survival (PFS) in two biomarker-defined populations is evaluated at both interim and final analyses. As an extension, a short-term biomarker is used to predict the long-term PFS in a Bayesian model in order to improve the precision of hazard ratio (HR) estimate at the interim analysis. The characteristics of the extended design are examined in a number of scenarios via simulations. The recommended adaptive design is shown to be useful in a phase II setting. When a short-term maker which correlates with the long-term PFS is available, the design can be applied in smaller early phase trials in which PFS requires longer follow-up. In summary, the adaptive design offers flexibility in randomized phase II patient enrichment trials and should be considered in an overall personalized healthcare (PHC) strategy. Copyright © 2013 Elsevier Inc. All rights reserved.
U10 : Trusted Truck(R) II (phase B).
DOT National Transportation Integrated Search
2009-01-01
Phase B of the Trusted Truck II project built on the system developed in Phase A (or Year 1). For the implementation portion of the project, systems were added to the trailer to provide additional diagnostic trailer data that can be sent to the TTM...
Movement Analysis Applied to the Basketball Jump Shot--Part II.
ERIC Educational Resources Information Center
Martin, Thomas P.
1981-01-01
The jump shot is one of the most important shots in the game of basketball. The movement analysis of the jump shot designates four phases: (1) preparatory position; (2) movement phase I (crouch); (3) movement phase II (jump); and (4) follow-through. (JN)
Searching for Images: The Analysis of Users' Queries for Image Retrieval in American History.
ERIC Educational Resources Information Center
Choi, Youngok; Rasmussen, Edie M.
2003-01-01
Studied users' queries for visual information in American history to identify the image attributes important for retrieval and the characteristics of users' queries for digital images, based on queries from 38 faculty and graduate students. Results of pre- and post-test questionnaires and interviews suggest principle categories of search terms.…
Searching and Filtering Tweets: CSIRO at the TREC 2012 Microblog Track
2012-11-01
stages. We first evaluate the effect of tweet corpus pre- processing in vanilla runs (no query expansion), and then assess the effect of query expansion...Effect of a vanilla run on D4 index (both realtime and non-real-time), and query expansion methods based on the submitted runs for two sets of queries
Knowledge Query Language (KQL)
2016-02-01
unlimited. This page intentionally left blank. iii EXECUTIVE SUMMARY Currently, queries for data ...retrieval from non-Structured Query Language (NoSQL) data stores are tightly coupled to the specific implementation of the data store implementation, making...of the storage content and format for querying NoSQL or relational data stores. This approach uses address expressions (or A-Expressions) embedded in
System, method and apparatus for conducting a keyterm search
NASA Technical Reports Server (NTRS)
McGreevy, Michael W. (Inventor)
2004-01-01
A keyterm search is a method of searching a database for subsets of the database that are relevant to an input query. First, a number of relational models of subsets of a database are provided. A query is then input. The query can include one or more keyterms. Next, a gleaning model of the query is created. The gleaning model of the query is then compared to each one of the relational models of subsets of the database. The identifiers of the relevant subsets are then output.
System, method and apparatus for conducting a phrase search
NASA Technical Reports Server (NTRS)
McGreevy, Michael W. (Inventor)
2004-01-01
A phrase search is a method of searching a database for subsets of the database that are relevant to an input query. First, a number of relational models of subsets of a database are provided. A query is then input. The query can include one or more sequences of terms. Next, a relational model of the query is created. The relational model of the query is then compared to each one of the relational models of subsets of the database. The identifiers of the relevant subsets are then output.
Targeted exploration and analysis of large cross-platform human transcriptomic compendia
Zhu, Qian; Wong, Aaron K; Krishnan, Arjun; Aure, Miriam R; Tadych, Alicja; Zhang, Ran; Corney, David C; Greene, Casey S; Bongo, Lars A; Kristensen, Vessela N; Charikar, Moses; Li, Kai; Troyanskaya, Olga G.
2016-01-01
We present SEEK (http://seek.princeton.edu), a query-based search engine across very large transcriptomic data collections, including thousands of human data sets from almost 50 microarray and next-generation sequencing platforms. SEEK uses a novel query-level cross-validation-based algorithm to automatically prioritize data sets relevant to the query and a robust search approach to identify query-coregulated genes, pathways, and processes. SEEK provides cross-platform handling, multi-gene query search, iterative metadata-based search refinement, and extensive visualization-based analysis options. PMID:25581801
Optimization of the Controlled Evaluation of Closed Relational Queries
NASA Astrophysics Data System (ADS)
Biskup, Joachim; Lochner, Jan-Hendrik; Sonntag, Sebastian
For relational databases, controlled query evaluation is an effective inference control mechanism preserving confidentiality regarding a previously declared confidentiality policy. Implementations of controlled query evaluation usually lack efficiency due to costly theorem prover calls. Suitably constrained controlled query evaluation can be implemented efficiently, but is not flexible enough from the perspective of database users and security administrators. In this paper, we propose an optimized framework for controlled query evaluation in relational databases, being efficiently implementable on the one hand and relaxing the constraints of previous approaches on the other hand.
Wollbrett, Julien; Larmande, Pierre; de Lamotte, Frédéric; Ruiz, Manuel
2013-04-15
In recent years, a large amount of "-omics" data have been produced. However, these data are stored in many different species-specific databases that are managed by different institutes and laboratories. Biologists often need to find and assemble data from disparate sources to perform certain analyses. Searching for these data and assembling them is a time-consuming task. The Semantic Web helps to facilitate interoperability across databases. A common approach involves the development of wrapper systems that map a relational database schema onto existing domain ontologies. However, few attempts have been made to automate the creation of such wrappers. We developed a framework, named BioSemantic, for the creation of Semantic Web Services that are applicable to relational biological databases. This framework makes use of both Semantic Web and Web Services technologies and can be divided into two main parts: (i) the generation and semi-automatic annotation of an RDF view; and (ii) the automatic generation of SPARQL queries and their integration into Semantic Web Services backbones. We have used our framework to integrate genomic data from different plant databases. BioSemantic is a framework that was designed to speed integration of relational databases. We present how it can be used to speed the development of Semantic Web Services for existing relational biological databases. Currently, it creates and annotates RDF views that enable the automatic generation of SPARQL queries. Web Services are also created and deployed automatically, and the semantic annotations of our Web Services are added automatically using SAWSDL attributes. BioSemantic is downloadable at http://southgreen.cirad.fr/?q=content/Biosemantic.
2013-01-01
Background In recent years, a large amount of “-omics” data have been produced. However, these data are stored in many different species-specific databases that are managed by different institutes and laboratories. Biologists often need to find and assemble data from disparate sources to perform certain analyses. Searching for these data and assembling them is a time-consuming task. The Semantic Web helps to facilitate interoperability across databases. A common approach involves the development of wrapper systems that map a relational database schema onto existing domain ontologies. However, few attempts have been made to automate the creation of such wrappers. Results We developed a framework, named BioSemantic, for the creation of Semantic Web Services that are applicable to relational biological databases. This framework makes use of both Semantic Web and Web Services technologies and can be divided into two main parts: (i) the generation and semi-automatic annotation of an RDF view; and (ii) the automatic generation of SPARQL queries and their integration into Semantic Web Services backbones. We have used our framework to integrate genomic data from different plant databases. Conclusions BioSemantic is a framework that was designed to speed integration of relational databases. We present how it can be used to speed the development of Semantic Web Services for existing relational biological databases. Currently, it creates and annotates RDF views that enable the automatic generation of SPARQL queries. Web Services are also created and deployed automatically, and the semantic annotations of our Web Services are added automatically using SAWSDL attributes. BioSemantic is downloadable at http://southgreen.cirad.fr/?q=content/Biosemantic. PMID:23586394
Cockle-Hearne, Jane; Barnett, Deborah; Hicks, James; Simpson, Mhairi; White, Isabel; Faithfull, Sara
2018-04-30
Distress after prostate cancer treatment is a substantial burden for up to one-third of men diagnosed. Physical and emotional symptoms and health service use can intensify, yet men are reticent to accept support. To provide accessible support that can be cost effectively integrated into care pathways, we developed a unique, Web-based, self-guided, cognitive-behavior program incorporating filmed and interactive peer support. To assess feasibility of the intervention among men experiencing distress after prostate cancer treatment. Demand, acceptability, change in distress and self-efficacy, and challenges for implementation in clinical practice were measured. A pre-post, within-participant comparison, mixed-methods research design was followed. Phase I and II were conducted in primary care psychological service and secondary care cancer service, respectively. Men received clinician-generated postal invitations: phase I, 432 men diagnosed <5 years; phase II, 606 men diagnosed <3.5 years. Consent was Web-based. Men with mild and moderate distress were enrolled. Web-based assessment included demographic, disease, treatment characteristics; distress (General Health Questionnaire-28); depression (Patient Health Questionnaire-9); anxiety (General Anxiety Disorder Scale-7); self-efficacy (Self-Efficacy for Symptom Control Inventory); satisfaction (author-generated, Likert-type questionnaire). Uptake and adherence were assessed with reference to the persuasive systems design model. Telephone interviews explored participant experience (phase II, n=10); interviews with health care professionals (n=3) explored implementation issues. A total of 135 men consented (phase I, 61/432, 14.1%; phase II, 74/606, 12.2%); from 96 eligible men screened for distress, 32% (30/96) entered the intervention (phase I, n=10; phase II, n=20). Twenty-four completed the Web-based program and assessments (phase I, n=8; phase II, n=16). Adherence for phase I and II was module completion rate 63% (mean 2.5, SD 1.9) versus 92% (mean 3.7, SD 1.0); rate of completing cognitive behavior therapy exercises 77% (mean 16.1, SD 6.2) versus 88% (mean 18.6, SD 3.9). Chat room activity occurred among 63% (5/8) and 75% (12/16) of men, respectively. In phase I, 75% (6/8) of men viewed all the films; in phase II, the total number of unique views weekly was 16, 11, 11, and 10, respectively. The phase II mood diary was completed by 100% (16/16) of men. Satisfaction was high for the program and films. Limited efficacy testing indicated improvement in distress baseline to post intervention: phase I, P=.03, r=-.55; phase II, P=.001, r=-.59. Self-efficacy improved for coping P=.02, r=-.41. Service assessment confirmed ease of assimilation into clinical practice and clarified health care practitioner roles. The Web-based program is acceptable and innovative in clinical practice. It was endorsed by patients and has potential to positively impact the experience of men with distress after prostate cancer treatment. It can potentially be delivered in a stepped model of psychological support in primary or secondary care. Feasibility evidence is compelling, supporting further evaluative research to determine clinical and cost effectiveness. ©Jane Cockle-Hearne, Deborah Barnett, James Hicks, Mhairi Simpson, Isabel White, Sara Faithfull. Originally published in JMIR Cancer (http://cancer.jmir.org), 30.04.2018.
Barnett, Deborah; Hicks, James; Simpson, Mhairi; White, Isabel; Faithfull, Sara
2018-01-01
Background Distress after prostate cancer treatment is a substantial burden for up to one-third of men diagnosed. Physical and emotional symptoms and health service use can intensify, yet men are reticent to accept support. To provide accessible support that can be cost effectively integrated into care pathways, we developed a unique, Web-based, self-guided, cognitive-behavior program incorporating filmed and interactive peer support. Objective To assess feasibility of the intervention among men experiencing distress after prostate cancer treatment. Demand, acceptability, change in distress and self-efficacy, and challenges for implementation in clinical practice were measured. Methods A pre-post, within-participant comparison, mixed-methods research design was followed. Phase I and II were conducted in primary care psychological service and secondary care cancer service, respectively. Men received clinician-generated postal invitations: phase I, 432 men diagnosed <5 years; phase II, 606 men diagnosed <3.5 years. Consent was Web-based. Men with mild and moderate distress were enrolled. Web-based assessment included demographic, disease, treatment characteristics; distress (General Health Questionnaire-28); depression (Patient Health Questionnaire-9); anxiety (General Anxiety Disorder Scale-7); self-efficacy (Self-Efficacy for Symptom Control Inventory); satisfaction (author-generated, Likert-type questionnaire). Uptake and adherence were assessed with reference to the persuasive systems design model. Telephone interviews explored participant experience (phase II, n=10); interviews with health care professionals (n=3) explored implementation issues. Results A total of 135 men consented (phase I, 61/432, 14.1%; phase II, 74/606, 12.2%); from 96 eligible men screened for distress, 32% (30/96) entered the intervention (phase I, n=10; phase II, n=20). Twenty-four completed the Web-based program and assessments (phase I, n=8; phase II, n=16). Adherence for phase I and II was module completion rate 63% (mean 2.5, SD 1.9) versus 92% (mean 3.7, SD 1.0); rate of completing cognitive behavior therapy exercises 77% (mean 16.1, SD 6.2) versus 88% (mean 18.6, SD 3.9). Chat room activity occurred among 63% (5/8) and 75% (12/16) of men, respectively. In phase I, 75% (6/8) of men viewed all the films; in phase II, the total number of unique views weekly was 16, 11, 11, and 10, respectively. The phase II mood diary was completed by 100% (16/16) of men. Satisfaction was high for the program and films. Limited efficacy testing indicated improvement in distress baseline to post intervention: phase I, P=.03, r=−.55; phase II, P=.001, r=−.59. Self-efficacy improved for coping P=.02, r=−.41. Service assessment confirmed ease of assimilation into clinical practice and clarified health care practitioner roles. Conclusions The Web-based program is acceptable and innovative in clinical practice. It was endorsed by patients and has potential to positively impact the experience of men with distress after prostate cancer treatment. It can potentially be delivered in a stepped model of psychological support in primary or secondary care. Feasibility evidence is compelling, supporting further evaluative research to determine clinical and cost effectiveness. PMID:29712628
DOE Office of Scientific and Technical Information (OSTI.GOV)
Oglesby, Kenneth D.; Woskov, Paul; Einstein, Herbert
This report covers the technical work in Phase I of this DOE-Nuclear Program STTR Fast Track project. All key tasks were successfully performed, new tasks were added to utilize DOD-AFRL’s 95 GigaHertz (GHz) gyrotron in Phase II, while other lesser tasks were left for Phase II efforts or were requested to be made optional. This research adds to our understanding of using MMW power to melt and vaporize rocks and steel/ metals and laid plans for future testing in Phase II. This work built upon a prior DOE project DE-EE0005504 that developed the basic waveguide setup, process and instruments. Inmore » this project we were investigating the use of MMW to form rock melt and steel plugs in deep wells to further isolate highly radioactive nuclear waste in ultra-deep basement rocks for long term storage. This technology also has potential for deep well drilling for nuclear storage, geothermal and oil and gas industries. It also has the potential for simultaneously sealing and securing the wellbore with a thick rock melt liner as the wellbore is drilled. This allows for higher levels of safety and protection of the environment during deep drilling operations. The larger purpose of this project was to find answers to key questions in progressing MMW technology for these applications. Phase I of this project continued bench testing using the MIT 10 kilo-Watt (kW), 28 GHz frequency laboratory gyrotron, literature searches, planning and design of equipment for Phase II efforts. Furnace melting and rock testing (Tasks 4 and 5) were deferred to Phase II due to lack of concurrent availability of the furnace and personnel at MIT. That delay and lower temperature furnace (limited to 1650oC) caused rethinking of Task 4 to utilize coordinated rock selection with the DOD testing in Phase II. The high pressure and high power window design work (moved to Phase I Task 3 from Phase II Task 20) and Additive materials and methods (Tasks 7 & 8) performed in Phase I may become patentable and thus little detail can be provided in this public report. A version of that new high pressure, high MMW power window may be built for possible Phase II testing at the DOD site. Most significantly, additional tasks were added for planning the use of the Department of Defense, Air Force Research Laboratory’s (DOD-AFRL’s) System 0 gyrotron in Phase II. Specifically added and accomplished were multiple discussions on DOD and DOE-MIT-Impact goals, timing between ongoing DOD testing, outlining the required equipment and instruments for rock testing, and terms for an agreement. That addition required a visit to Kirtland AFB in Albuquerque, New Mexico to talk to key DOD-AFRL personnel and management. A DOD-Impact-MIT charter (i.e., contract) is now being circulated for signatures. Also added task to Phase I, MIT designed the critical path reflected power isolator screen for Phase II testing. To ensure compatibility, that design was computer simulated for the expected heat load distribution and the resulting temperature increase. Advancing the MMW testing up to the optimum 95 GHz and 100kW (5X higher) power levels was stated in the original proposal to be a key required development step for this technology to achieve prototype drilling, lining, and rock melting/ vaporization for creating sealing plugs.« less
Ritz, Stacey A; Wan, Junxiang; Diaz-Sanchez, David
2007-01-01
Airborne particulate pollutants, such as diesel exhaust particles, are thought to exacerbate lung and cardiovascular diseases through induction of oxidative stress. Sulforaphane, derived from cruciferous vegetables, is the most potent known inducer of phase II enzymes involved in the detoxification of xenobiotics. We postulated that sulforaphane may be able to ameliorate the adverse effects of pollutants by upregulating expression of endogenous antioxidant enzymes. Stimulation of bronchial epithelial cells with the chemical constituents of diesel particles result in the production of proinflammatory cytokines. We first demonstrated a role for phase II enzymes in regulating diesel effects by transfecting the airway epithelial cell line (BEAS-2B) with the sentinel phase II enzyme NAD(P)H: quinine oxidoreductase 1 (NQO1). IL-8 production in response to diesel extract was significantly reduced in these compared with untransfected cells. We then examined whether sulforaphane would stimulate phase II induction and whether this would thereby ablate the effect of diesel extracts on cytokine production. We verified that sulforaphane significantly augmented expression of the phase II enzyme genes GSTM1 and NQO1 and confirmed that sulforaphane treatment increased glutathione S-transferase activity in epithelial cells without inducing cell death or apoptosis. Sulforaphane pretreatment inhibited IL-8 production by BEAS-2B cells upon stimulation with diesel extract. Similarly, whereas diesel extract stimulated production of IL-8, granulocyte-macrophage colony-stimulating factor, and IL-1beta from primary human bronchial epithelial cells, sulforaphane pretreatment inhibited diesel-induced production of all of these cytokines. Our studies show that sulforaphane can mitigate the effect of diesel in respiratory epithelial cells and demonstrate the chemopreventative potential of phase II enzyme enhancement.
1990-01-01
1-20 1-6 Sites Defined and Ranked During IRP Phase I Study. 1-29 1-7 Aerial Photograph of Site 2, April 1988. 1-32 1-8 Site 2 Sampling Locations...Utilized During Phase II Investigations. 1-35 1-9 Aerial Photograph of Site 3, April 1988. 1-38 1-10 Site 3 Sampling Locations Utilized During Phase II...Investigations. 1-47 1-11 Aerial Photograph of Site 4, April 1988. 1-54 1-12 Site 4 Sampling Locations Utilized During Phase II Investigations. 1-57 1-13
Piper betle Induced Cytoprotective Genes and Proteins via the Nrf2/ARE Pathway in Aging Mice.
Aliahmat, Nor Syahida; Abdul Sani, Nur Fathiah; Wan Hasan, Wan Nuraini; Makpol, Suzana; Wan Ngah, Wan Zurinah; Mohd Yusof, Yasmin Anum
2016-01-01
The objective of this study was to elucidate the underlying antioxidant mechanism of aqueous extract of Piper betle (PB) in aging rats. The nuclear factor erythroid 2-related factor 2 (Nrf2)/ARE pathway involving phase II detoxifying and antioxidant enzymes plays an important role in the antioxidant system by reducing electrophiles and reactive oxygen species through induction of phase II enzymes and proteins. Genes and proteins of phase II detoxifying antioxidant enzymes were analyzed by QuantiGenePlex 2.0 Assay and Western blot analysis. PB significantly induced genes and proteins of phase II and antioxidant enzymes, NAD(P)H quinone oxidoreductase 1, and catalase in aging mice (p < 0.05). The expression of these enzymes were stimulated via translocation of Nrf2 into the nucleus, indicating the involvement of ARE, a cis-acting motif located in the promoter region of nearly all phase II genes. PB was testified for the first time to induce cytoprotective genes through the Nrf2/ARE signaling pathway, thus unraveling the antioxidant mechanism of PB during the aging process. © 2016 S. Karger AG, Basel.
ERIC Educational Resources Information Center
Northern Montana Coll., Havre.
The dissemination phase (Phase II) of the Rural Shared Services Project is reported in this document. Efforts of the dissemination phase were concentrated in 5 target states: Vermont, Georgia, Wyoming, Montana, and New Mexico; national dissemination was limited to attendance at national conferences, the U. S. Office of Education PREP materials for…
Site preference of alloying elements in DO22-Ni3V phase: Phase-field and first-principles study
NASA Astrophysics Data System (ADS)
Zhang, Ding-Ni; Shangguan, Qian-Qian; Liu, Fu; Zhang, Ming-Yi
2015-07-01
Site preference of alloying elements in DO22-Ni3V phase was investigated using phase-field and first-principles method. The concentrations of alloying elements on sublattices of DO22-Ni3V phase were quantitatively studied using phase-field model based on microscopic diffusion equations. The phase-field computation results demonstrate that the concentration differences of alloying elements on the NiI and NiII site are attributed to the coordination environment difference. Host atoms Ni and substitutional ternary additions Al prefer to occupy NiI site. Antisite atoms V show site preference on the NiII site. Further reason of site preference of alloying elements on the two different Ni sites were studied using first-principles method to calculate the electronic structure of DO22-Ni3V phase. Calculation of density of states, orbitals population and charge population of the optimized Ni3V structure found that the electronic structures of NiI and NiII sites are different. Electronic structure difference, which is caused by coordination environment difference, is the essential reason for site selectivity behaviors of alloying elements on NiI and NiII sites.
An index-based algorithm for fast on-line query processing of latent semantic analysis
Li, Pohan; Wang, Wei
2017-01-01
Latent Semantic Analysis (LSA) is widely used for finding the documents whose semantic is similar to the query of keywords. Although LSA yield promising similar results, the existing LSA algorithms involve lots of unnecessary operations in similarity computation and candidate check during on-line query processing, which is expensive in terms of time cost and cannot efficiently response the query request especially when the dataset becomes large. In this paper, we study the efficiency problem of on-line query processing for LSA towards efficiently searching the similar documents to a given query. We rewrite the similarity equation of LSA combined with an intermediate value called partial similarity that is stored in a designed index called partial index. For reducing the searching space, we give an approximate form of similarity equation, and then develop an efficient algorithm for building partial index, which skips the partial similarities lower than a given threshold θ. Based on partial index, we develop an efficient algorithm called ILSA for supporting fast on-line query processing. The given query is transformed into a pseudo document vector, and the similarities between query and candidate documents are computed by accumulating the partial similarities obtained from the index nodes corresponds to non-zero entries in the pseudo document vector. Compared to the LSA algorithm, ILSA reduces the time cost of on-line query processing by pruning the candidate documents that are not promising and skipping the operations that make little contribution to similarity scores. Extensive experiments through comparison with LSA have been done, which demonstrate the efficiency and effectiveness of our proposed algorithm. PMID:28520747
Analyzing Medical Image Search Behavior: Semantics and Prediction of Query Results.
De-Arteaga, Maria; Eggel, Ivan; Kahn, Charles E; Müller, Henning
2015-10-01
Log files of information retrieval systems that record user behavior have been used to improve the outcomes of retrieval systems, understand user behavior, and predict events. In this article, a log file of the ARRS GoldMiner search engine containing 222,005 consecutive queries is analyzed. Time stamps are available for each query, as well as masked IP addresses, which enables to identify queries from the same person. This article describes the ways in which physicians (or Internet searchers interested in medical images) search and proposes potential improvements by suggesting query modifications. For example, many queries contain only few terms and therefore are not specific; others contain spelling mistakes or non-medical terms that likely lead to poor or empty results. One of the goals of this report is to predict the number of results a query will have since such a model allows search engines to automatically propose query modifications in order to avoid result lists that are empty or too large. This prediction is made based on characteristics of the query terms themselves. Prediction of empty results has an accuracy above 88%, and thus can be used to automatically modify the query to avoid empty result sets for a user. The semantic analysis and data of reformulations done by users in the past can aid the development of better search systems, particularly to improve results for novice users. Therefore, this paper gives important ideas to better understand how people search and how to use this knowledge to improve the performance of specialized medical search engines.
An index-based algorithm for fast on-line query processing of latent semantic analysis.
Zhang, Mingxi; Li, Pohan; Wang, Wei
2017-01-01
Latent Semantic Analysis (LSA) is widely used for finding the documents whose semantic is similar to the query of keywords. Although LSA yield promising similar results, the existing LSA algorithms involve lots of unnecessary operations in similarity computation and candidate check during on-line query processing, which is expensive in terms of time cost and cannot efficiently response the query request especially when the dataset becomes large. In this paper, we study the efficiency problem of on-line query processing for LSA towards efficiently searching the similar documents to a given query. We rewrite the similarity equation of LSA combined with an intermediate value called partial similarity that is stored in a designed index called partial index. For reducing the searching space, we give an approximate form of similarity equation, and then develop an efficient algorithm for building partial index, which skips the partial similarities lower than a given threshold θ. Based on partial index, we develop an efficient algorithm called ILSA for supporting fast on-line query processing. The given query is transformed into a pseudo document vector, and the similarities between query and candidate documents are computed by accumulating the partial similarities obtained from the index nodes corresponds to non-zero entries in the pseudo document vector. Compared to the LSA algorithm, ILSA reduces the time cost of on-line query processing by pruning the candidate documents that are not promising and skipping the operations that make little contribution to similarity scores. Extensive experiments through comparison with LSA have been done, which demonstrate the efficiency and effectiveness of our proposed algorithm.
Bat-Inspired Algorithm Based Query Expansion for Medical Web Information Retrieval.
Khennak, Ilyes; Drias, Habiba
2017-02-01
With the increasing amount of medical data available on the Web, looking for health information has become one of the most widely searched topics on the Internet. Patients and people of several backgrounds are now using Web search engines to acquire medical information, including information about a specific disease, medical treatment or professional advice. Nonetheless, due to a lack of medical knowledge, many laypeople have difficulties in forming appropriate queries to articulate their inquiries, which deem their search queries to be imprecise due the use of unclear keywords. The use of these ambiguous and vague queries to describe the patients' needs has resulted in a failure of Web search engines to retrieve accurate and relevant information. One of the most natural and promising method to overcome this drawback is Query Expansion. In this paper, an original approach based on Bat Algorithm is proposed to improve the retrieval effectiveness of query expansion in medical field. In contrast to the existing literature, the proposed approach uses Bat Algorithm to find the best expanded query among a set of expanded query candidates, while maintaining low computational complexity. Moreover, this new approach allows the determination of the length of the expanded query empirically. Numerical results on MEDLINE, the on-line medical information database, show that the proposed approach is more effective and efficient compared to the baseline.
Luo, Yuan; Szolovits, Peter
2016-01-01
In natural language processing, stand-off annotation uses the starting and ending positions of an annotation to anchor it to the text and stores the annotation content separately from the text. We address the fundamental problem of efficiently storing stand-off annotations when applying natural language processing on narrative clinical notes in electronic medical records (EMRs) and efficiently retrieving such annotations that satisfy position constraints. Efficient storage and retrieval of stand-off annotations can facilitate tasks such as mapping unstructured text to electronic medical record ontologies. We first formulate this problem into the interval query problem, for which optimal query/update time is in general logarithm. We next perform a tight time complexity analysis on the basic interval tree query algorithm and show its nonoptimality when being applied to a collection of 13 query types from Allen's interval algebra. We then study two closely related state-of-the-art interval query algorithms, proposed query reformulations, and augmentations to the second algorithm. Our proposed algorithm achieves logarithmic time stabbing-max query time complexity and solves the stabbing-interval query tasks on all of Allen's relations in logarithmic time, attaining the theoretic lower bound. Updating time is kept logarithmic and the space requirement is kept linear at the same time. We also discuss interval management in external memory models and higher dimensions.
Luo, Yuan; Szolovits, Peter
2016-01-01
In natural language processing, stand-off annotation uses the starting and ending positions of an annotation to anchor it to the text and stores the annotation content separately from the text. We address the fundamental problem of efficiently storing stand-off annotations when applying natural language processing on narrative clinical notes in electronic medical records (EMRs) and efficiently retrieving such annotations that satisfy position constraints. Efficient storage and retrieval of stand-off annotations can facilitate tasks such as mapping unstructured text to electronic medical record ontologies. We first formulate this problem into the interval query problem, for which optimal query/update time is in general logarithm. We next perform a tight time complexity analysis on the basic interval tree query algorithm and show its nonoptimality when being applied to a collection of 13 query types from Allen’s interval algebra. We then study two closely related state-of-the-art interval query algorithms, proposed query reformulations, and augmentations to the second algorithm. Our proposed algorithm achieves logarithmic time stabbing-max query time complexity and solves the stabbing-interval query tasks on all of Allen’s relations in logarithmic time, attaining the theoretic lower bound. Updating time is kept logarithmic and the space requirement is kept linear at the same time. We also discuss interval management in external memory models and higher dimensions. PMID:27478379
Executing SPARQL Queries over the Web of Linked Data
NASA Astrophysics Data System (ADS)
Hartig, Olaf; Bizer, Christian; Freytag, Johann-Christoph
The Web of Linked Data forms a single, globally distributed dataspace. Due to the openness of this dataspace, it is not possible to know in advance all data sources that might be relevant for query answering. This openness poses a new challenge that is not addressed by traditional research on federated query processing. In this paper we present an approach to execute SPARQL queries over the Web of Linked Data. The main idea of our approach is to discover data that might be relevant for answering a query during the query execution itself. This discovery is driven by following RDF links between data sources based on URIs in the query and in partial results. The URIs are resolved over the HTTP protocol into RDF data which is continuously added to the queried dataset. This paper describes concepts and algorithms to implement our approach using an iterator-based pipeline. We introduce a formalization of the pipelining approach and show that classical iterators may cause blocking due to the latency of HTTP requests. To avoid blocking, we propose an extension of the iterator paradigm. The evaluation of our approach shows its strengths as well as the still existing challenges.
A Natural Language Interface Concordant with a Knowledge Base.
Han, Yong-Jin; Park, Seong-Bae; Park, Se-Young
2016-01-01
The discordance between expressions interpretable by a natural language interface (NLI) system and those answerable by a knowledge base is a critical problem in the field of NLIs. In order to solve this discordance problem, this paper proposes a method to translate natural language questions into formal queries that can be generated from a graph-based knowledge base. The proposed method considers a subgraph of a knowledge base as a formal query. Thus, all formal queries corresponding to a concept or a predicate in the knowledge base can be generated prior to query time and all possible natural language expressions corresponding to each formal query can also be collected in advance. A natural language expression has a one-to-one mapping with a formal query. Hence, a natural language question is translated into a formal query by matching the question with the most appropriate natural language expression. If the confidence of this matching is not sufficiently high the proposed method rejects the question and does not answer it. Multipredicate queries are processed by regarding them as a set of collected expressions. The experimental results show that the proposed method thoroughly handles answerable questions from the knowledge base and rejects unanswerable ones effectively.
Saying What You're Looking For: Linguistics Meets Video Search.
Barrett, Daniel Paul; Barbu, Andrei; Siddharth, N; Siskind, Jeffrey Mark
2016-10-01
We present an approach to searching large video corpora for clips which depict a natural-language query in the form of a sentence. Compositional semantics is used to encode subtle meaning differences lost in other approaches, such as the difference between two sentences which have identical words but entirely different meaning: The person rode the horse versus The horse rode the person. Given a sentential query and a natural-language parser, we produce a score indicating how well a video clip depicts that sentence for each clip in a corpus and return a ranked list of clips. Two fundamental problems are addressed simultaneously: detecting and tracking objects, and recognizing whether those tracks depict the query. Because both tracking and object detection are unreliable, our approach uses the sentential query to focus the tracker on the relevant participants and ensures that the resulting tracks are described by the sentential query. While most earlier work was limited to single-word queries which correspond to either verbs or nouns, we search for complex queries which contain multiple phrases, such as prepositional phrases, and modifiers, such as adverbs. We demonstrate this approach by searching for 2,627 naturally elicited sentential queries in 10 Hollywood movies.
Context-Aware Online Commercial Intention Detection
NASA Astrophysics Data System (ADS)
Hu, Derek Hao; Shen, Dou; Sun, Jian-Tao; Yang, Qiang; Chen, Zheng
With more and more commercial activities moving onto the Internet, people tend to purchase what they need through Internet or conduct some online research before the actual transactions happen. For many Web users, their online commercial activities start from submitting a search query to search engines. Just like the common Web search queries, the queries with commercial intention are usually very short. Recognizing the queries with commercial intention against the common queries will help search engines provide proper search results and advertisements, help Web users obtain the right information they desire and help the advertisers benefit from the potential transactions. However, the intentions behind a query vary a lot for users with different background and interest. The intentions can even be different for the same user, when the query is issued in different contexts. In this paper, we present a new algorithm framework based on skip-chain conditional random field (SCCRF) for automatically classifying Web queries according to context-based online commercial intention. We analyze our algorithm performance both theoretically and empirically. Extensive experiments on several real search engine log datasets show that our algorithm can improve more than 10% on F1 score than previous algorithms on commercial intention detection.
Pharmacological Treatment of Cannabis-Related Disorders: A Narrative Review.
Gorelick, David A
2016-01-01
Cannabis is the most widely used illicit psychoactive substance world-wide, yet no medication is approved for the treatment of intoxication, withdrawal, or cannabis use disorder (CUD). To comprehensively review the current state of knowledge. Search of the PubMed electronic data base and review of reference lists of relevant articles to identify controlled clinical trials of pharmacological treatment. The search identified 4 trials for specific intoxication symptoms (none for global intoxication), 7 trials for withdrawal, and 12 phase II trials for CUD. One or two trials each suggest that propranolol is effective for some intoxication symptoms, antipsychotics for cannabis-induced psychosis, and dronabinol (synthetic THC) and gabapentin for cannabis withdrawal. Of 10 medications and one medication combination studied in 12 trials for CUD, only two medications were effective (in single trials): gabapentin and Nacetylcysteine (in adolescents). Not effective were dronabinol and several antidepressants, anticonvulsants, and antianxiety medications. Three trials of antidepressants for CUD with comorbid depression gave inconsistent results. A trial of atomoxetine for CUD with comorbid ADHD showed no efficacy. Five trials of second-generation antipsychotics for CUD with comorbid schizophrenia showed none better than any other. Further research is needed to confirm the efficacy of gabapentin for withdrawal and gabapentin and N-acetylcysteine for CUD and to develop new medications for all 3 cannabis-related disorders. Copyright© Bentham Science Publishers; For any queries, please email at epub@benthamscience.org.
Incremental Query Rewriting with Resolution
NASA Astrophysics Data System (ADS)
Riazanov, Alexandre; Aragão, Marcelo A. T.
We address the problem of semantic querying of relational databases (RDB) modulo knowledge bases using very expressive knowledge representation formalisms, such as full first-order logic or its various fragments. We propose to use a resolution-based first-order logic (FOL) reasoner for computing schematic answers to deductive queries, with the subsequent translation of these schematic answers to SQL queries which are evaluated using a conventional relational DBMS. We call our method incremental query rewriting, because an original semantic query is rewritten into a (potentially infinite) series of SQL queries. In this chapter, we outline the main idea of our technique - using abstractions of databases and constrained clauses for deriving schematic answers, and provide completeness and soundness proofs to justify the applicability of this technique to the case of resolution for FOL without equality. The proposed method can be directly used with regular RDBs, including legacy databases. Moreover, we propose it as a potential basis for an efficient Web-scale semantic search technology.
Research on presentation and query service of geo-spatial data based on ontology
NASA Astrophysics Data System (ADS)
Li, Hong-wei; Li, Qin-chao; Cai, Chang
2008-10-01
The paper analyzed the deficiency on presentation and query of geo-spatial data existed in current GIS, discussed the advantages that ontology possessed in formalization of geo-spatial data and the presentation of semantic granularity, taken land-use classification system as an example to construct domain ontology, and described it by OWL; realized the grade level and category presentation of land-use data benefited from the thoughts of vertical and horizontal navigation; and then discussed query mode of geo-spatial data based on ontology, including data query based on types and grade levels, instances and spatial relation, and synthetic query based on types and instances; these methods enriched query mode of current GIS, and is a useful attempt; point out that the key point of the presentation and query of spatial data based on ontology is to construct domain ontology that can correctly reflect geo-concept and its spatial relation and realize its fine formalization description.
Computer Simulations of Polytetrafluoroethylene in the Solid State
NASA Astrophysics Data System (ADS)
Holt, D. B.; Farmer, B. L.; Eby, R. K.; Macturk, K. S.
1996-03-01
Force field parameters (Set I) for fluoropolymers were previously derived from MOPAC AM1 semiempirical data on model molecules. A second set (Set II) was derived from the AM1 results augmented by ab initio calculations. Both sets yield reasonable helical and phase II packing structures for polytetrafluoroethylene (PTFE) chains. However, Set I and Set II differ in the strength of van der Waals interactions, with Set II having deeper potential wells (order of magnitude). To differentiate which parameter set provides a better description of PTFE behavior, molecular dynamics simulations have been performed with Biosym Discover on clusters of PTFE chains which begin in a phase II packing environment. Added to the model are artificial constraints which allow the simulation of thermal expansion without having to define periodic boundary conditions for each specific temperature of interest. The preliminary dynamics simulations indicate that the intra- and intermolecular interactions provided by Set I are too weak. The degree of helical disorder and chain motion are high even at temperatures well below the phase II-phase IV transition temperature (19 C). Set II appears to yield a better description of PTFE in the solid state.
Design, analysis, and test verification of advanced encapsulation systems
NASA Technical Reports Server (NTRS)
Mardesich, N.; Minning, C.
1982-01-01
Design sensitivities are established for the development of photovoltaic module criteria and the definition of needed research tasks. The program consists of three phases. In Phase I, analytical models were developed to perform optical, thermal, electrical, and structural analyses on candidate encapsulation systems. From these analyses several candidate systems will be selected for qualification testing during Phase II. Additionally, during Phase II, test specimens of various types will be constructed and tested to determine the validity of the analysis methodology developed in Phase I. In Phse III, a finalized optimum design based on knowledge gained in Phase I and II will be developed. All verification testing was completed during this period. Preliminary results and observations are discussed. Descriptions of the thermal, thermal structural, and structural deflection test setups are included.
Technology Demonstration of the Zero Emissions Chromium Electroplating System
2008-02-01
Phase I trivalent chromium results ................................................................... 23 18 Phase II total chromium in PRD fluid results...0 xa B D F H J L Sam pies Figure 16. Phase II iron results. ERDC/CERL TR-05-35, Vol. 1 23 Trivalent Chromium Phase I Analysis for Phase I was...with the samples. Each sample was analyzed twice, and an average was computed. Figure 17 shows the results. ANAD has specified that Trivalent Chromium
Laboratory modeling of energy dissipation in broken-back culverts - phase II.
DOT National Transportation Integrated Search
2011-05-01
This report represents Phase II of broken-back culverts with a drop of 6 feet. The first phase of this research was performed for a drop of 24 feet. This research investigates the reduction in scour downstream of a broken-back culvert by forming a hy...
This Small Business Innovation Research (SBIR) Phase II project will employ the large scale; highly reliable boron-doped ultrananocrystalline diamond (BD-UNCD®) electrodes developed during Phase I project to build and test Electrochemical Anodic Oxidation process (EAOP)...
78 FR 18305 - Notice of Request for Extension of a Currently Approved Information Collection
Federal Register 2010, 2011, 2012, 2013, 2014
2013-03-26
... Identity Verification (PIV) Request for Credential, the USDA Homeland Security Presidential Directive 12... consists of two phases of implementation: Personal Identity Verification phase I (PIV I) and Personal Identity Verification phase II (PIV II). The information requested must be provided by Federal employees...
A two-step spin crossover mononuclear iron(II) complex with a [HS-LS-LS] intermediate phase.
Bonnet, Sylvestre; Siegler, Maxime A; Costa, José Sánchez; Molnár, Gábor; Bousseksou, Azzedine; Spek, Anthony L; Gamez, Patrick; Reedijk, Jan
2008-11-21
The two-step spin crossover of a new mononuclear iron(ii) complex is studied by magnetic, crystallographic and calorimetric methods revealing two successive first-order phase transitions and an ordered intermediate phase built by the repetition of the unprecedented [HS-LS-LS] motif.
DOT National Transportation Integrated Search
2008-08-01
In Phase II of this investigation, we used a fully interactive PC-based STISIM driving simulator, to conduct two : experiments which were similar to experiments in Phase I. The participants were 120 licensed drivers from three : age groups18-24, 3...
Tipster Text Phase 2 Architecture Design
1996-06-19
TIPSTER Text Phase II Architecture Design Version 2.1p 19 June 1996 Ralph Grishman New York University grishman @cs.nyu.edu and the TIPSTER...1996 2. REPORT TYPE 3. DATES COVERED 00-00-1996 to 00-00-1996 4. TITLE AND SUBTITLE TIPSTER Text Phase II Architecture Design 5a. CONTRACT
Hironaka, Shuichi; Tsubosa, Yasuhiro; Mizusawa, Junki; Kii, Takayuki; Kato, Ken; Tsushima, Takahiro; Chin, Keisho; Tomori, Akihisa; Okuno, Tatsuya; Taniki, Toshikatsu; Ura, Takashi; Matsushita, Hisayuki; Kojima, Takashi; Doki, Yuichiro; Kusaba, Hitoshi; Fujitani, Kazumasa; Taira, Koichi; Seki, Shiko; Nakamura, Tsutomu; Kitagawa, Yuko
2014-01-01
We carried out a phase I/II trial of adding 2-weekly docetaxel to cisplatin plus fluorouracil (CF) therapy (2-weekly DCF regimen) in esophageal cancer patients to investigate its safety and antimetastatic activity. Patients received 2-weekly docetaxel (30 mg/m2 [dose level (DL)1] or 40 mg/m2 [DL2] with a 3 + 3 design in phase I, on days 1 and 15) in combination with fixed-dose CF (80 mg/m2 cisplatin, day 1; 800 mg/m2 fluorouracil, days 1–5) repeated every 4 weeks. The primary endpoint was dose-limiting toxicity (DLT) in phase I and central peer review-based response rate in phase II. At least 22 responders among 50 patients were required to satisfy the primary endpoint with a threshold of 35%. Sixty-two patients were enrolled in phase I and II. In phase I, 10 patients were enrolled with DLT of 0/3 at DL1 and 2/7 in DL2. Considering DLT and treatment compliance, the recommended phase II dose was determined as DL1. In phase II, the response rate was 62% (P < 0.0001; 95% confidence interval, 48–75%); median overall survival and progression-free survival were 11.1 and 5.8 months, respectively. Common grade 3/4 adverse events were neutropenia (25%), anemia (36%), hyponatremia (29%), anorexia (24%), and nausea (11%). No febrile neutropenia was observed. Pneumonitis caused treatment-related death in one patient. The 2-weekly DCF regimen showed promising antimetastatic activity and tolerability. A phase III study comparing this regimen with CF therapy is planned by the Japan Clinical Oncology Group. This study was registered at the UMIN Clinical Trials Registry as UMIN 000001737. PMID:25041052
Unusual Enhancement of Magnetization by Pressure in the Antiferro-Quadrupole-Ordered Phase in CeB6
NASA Astrophysics Data System (ADS)
Ikeda, Suguru; Sera, Masafumi; Hane, Shingo; Uwatoko, Yoshiya; Kosaka, Masashi; Kunii, Satoru
2007-06-01
The effect of pressure on CeB6 was investigated by the measurement of the magnetization (M) under pressure, and we obtained the following results. The effect of pressure on M in phase I is very small. By applying pressure, TQ is enhanced, but TN and the critical field from the antiferromagnetic (AFM) phase III to the antiferro-quadrupole (AFQ) phase II (HcIII--II) are suppressed, as previously reported. The magnetization curve in phase III shows the characteristic shoulder at H˜ HcIII--II/2 at ambient pressure. This shoulder becomes much more pronounced by applying pressure. Both HcIII--II and the magnetic field, where a shoulder is seen in the magnetization curve in phase III, are largely suppressed by pressure. In phase II, the M-T curve at a low magnetic field exhibits an unusual concave temperature dependence below TQ down to TN. Thus, we found that the lower the magnetic field, the larger the enhancement of M in both phases III and II. To clarify the origin of the unusual pressure effect of M, we performed a mean-field calculation for the 4-sublattice model using the experimental results of dTQ/dP>0 and dTN/dP<0 and assuming the positive pressure dependence of the Txyz-antiferro-octupole (AFO) interaction. The characteristic features of the pressure effect of M obtained by the experiments could be reproduced well by the mean-field calculation. We found that the origin of the characteristic effect of pressure on CeB6 is the change in the subtle balance between the AFM interaction and the magnetic field-induced-effective FM interaction induced by the coexistence of the Oxy-AFQ and Txyz-AFO interactions under pressure.
In-context query reformulation for failing SPARQL queries
NASA Astrophysics Data System (ADS)
Viswanathan, Amar; Michaelis, James R.; Cassidy, Taylor; de Mel, Geeth; Hendler, James
2017-05-01
Knowledge bases for decision support systems are growing increasingly complex, through continued advances in data ingest and management approaches. However, humans do not possess the cognitive capabilities to retain a bird's-eyeview of such knowledge bases, and may end up issuing unsatisfiable queries to such systems. This work focuses on the implementation of a query reformulation approach for graph-based knowledge bases, specifically designed to support the Resource Description Framework (RDF). The reformulation approach presented is instance-and schema-aware. Thus, in contrast to relaxation techniques found in the state-of-the-art, the presented approach produces in-context query reformulation.
Model-based query language for analyzing clinical processes.
Barzdins, Janis; Barzdins, Juris; Rencis, Edgars; Sostaks, Agris
2013-01-01
Nowadays large databases of clinical process data exist in hospitals. However, these data are rarely used in full scope. In order to perform queries on hospital processes, one must either choose from the predefined queries or develop queries using MS Excel-type software system, which is not always a trivial task. In this paper we propose a new query language for analyzing clinical processes that is easily perceptible also by non-IT professionals. We develop this language based on a process modeling language which is also described in this paper. Prototypes of both languages have already been verified using real examples from hospitals.
AQBE — QBE Style Queries for Archetyped Data
NASA Astrophysics Data System (ADS)
Sachdeva, Shelly; Yaginuma, Daigo; Chu, Wanming; Bhalla, Subhash
Large-scale adoption of electronic healthcare applications requires semantic interoperability. The new proposals propose an advanced (multi-level) DBMS architecture for repository services for health records of patients. These also require query interfaces at multiple levels and at the level of semi-skilled users. In this regard, a high-level user interface for querying the new form of standardized Electronic Health Records system has been examined in this study. It proposes a step-by-step graphical query interface to allow semi-skilled users to write queries. Its aim is to decrease user effort and communication ambiguities, and increase user friendliness.
StarView: The object oriented design of the ST DADS user interface
NASA Technical Reports Server (NTRS)
Williams, J. D.; Pollizzi, J. A.
1992-01-01
StarView is the user interface being developed for the Hubble Space Telescope Data Archive and Distribution Service (ST DADS). ST DADS is the data archive for HST observations and a relational database catalog describing the archived data. Users will use StarView to query the catalog and select appropriate datasets for study. StarView sends requests for archived datasets to ST DADS which processes the requests and returns the database to the user. StarView is designed to be a powerful and extensible user interface. Unique features include an internal relational database to navigate query results, a form definition language that will work with both CRT and X interfaces, a data definition language that will allow StarView to work with any relational database, and the ability to generate adhoc queries without requiring the user to understand the structure of the ST DADS catalog. Ultimately, StarView will allow the user to refine queries in the local database for improved performance and merge in data from external sources for correlation with other query results. The user will be able to create a query from single or multiple forms, merging the selected attributes into a single query. Arbitrary selection of attributes for querying is supported. The user will be able to select how query results are viewed. A standard form or table-row format may be used. Navigation capabilities are provided to aid the user in viewing query results. Object oriented analysis and design techniques were used in the design of StarView to support the mechanisms and concepts required to implement these features. One such mechanism is the Model-View-Controller (MVC) paradigm. The MVC allows the user to have multiple views of the underlying database, while providing a consistent mechanism for interaction regardless of the view. This approach supports both CRT and X interfaces while providing a common mode of user interaction. Another powerful abstraction is the concept of a Query Model. This concept allows a single query to be built form a single or multiple forms before it is submitted to ST DADS. Supporting this concept is the adhoc query generator which allows the user to select and qualify an indeterminate number attributes from the database. The user does not need any knowledge of how the joins across various tables are to be resolved. The adhoc generator calculates the joins automatically and generates the correct SQL query.
Data Processing Factory for the Sloan Digital Sky Survey
NASA Astrophysics Data System (ADS)
Stoughton, Christopher; Adelman, Jennifer; Annis, James T.; Hendry, John; Inkmann, John; Jester, Sebastian; Kent, Steven M.; Kuropatkin, Nickolai; Lee, Brian; Lin, Huan; Peoples, John, Jr.; Sparks, Robert; Tucker, Douglas; Vanden Berk, Dan; Yanny, Brian; Yocum, Dan
2002-12-01
The Sloan Digital Sky Survey (SDSS) data handling presents two challenges: large data volume and timely production of spectroscopic plates from imaging data. A data processing factory, using technologies both old and new, handles this flow. Distribution to end users is via disk farms, to serve corrected images and calibrated spectra, and a database, to efficiently process catalog queries. For distribution of modest amounts of data from Apache Point Observatory to Fermilab, scripts use rsync to update files, while larger data transfers are accomplished by shipping magnetic tapes commercially. All data processing pipelines are wrapped in scripts to address consecutive phases: preparation, submission, checking, and quality control. We constructed the factory by chaining these pipelines together while using an operational database to hold processed imaging catalogs. The science database catalogs all imaging and spectroscopic object, with pointers to the various external files associated with them. Diverse computing systems address particular processing phases. UNIX computers handle tape reading and writing, as well as calibration steps that require access to a large amount of data with relatively modest computational demands. Commodity CPUs process steps that require access to a limited amount of data with more demanding computations requirements. Disk servers optimized for cost per Gbyte serve terabytes of processed data, while servers optimized for disk read speed run SQLServer software to process queries on the catalogs. This factory produced data for the SDSS Early Data Release in June 2001, and it is currently producing Data Release One, scheduled for January 2003.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Jackson, Matthew W., E-mail: matthew.jackson@ucdenver.edu; Rusthoven, Chad G.; Jones, Bernard L.
Background: Primary mediastinal B cell lymphoma (PMBCL) is an uncommon lymphoma for which trials are few with small patient numbers. The role of radiation therapy (RT) after standard immunochemotherapy for early-stage disease has never been studied prospectively. We used the Surveillance, Epidemiology, and End Results (SEER) database to evaluate PMBCL and the impact of RT on outcomes. Methods and Materials: We queried the SEER database for patients with stage I-II PMBCL diagnosed from 2001 to 2011. Retrievable data included age, gender, race (white/nonwhite), stage, extranodal disease, year of diagnosis, and use of RT as a component of definitive therapy. Kaplan-Meier overallmore » survival (OS) estimates, univariate (UVA) log-rank and multivariate (MVA) Cox proportional hazards regression analyses were performed. Results: Two hundred fifty patients with stage I-II disease were identified, with a median follow-up time of 39 months (range, 3-125 months). The median age was 36 years (range, 18-89 years); 61% were female; 76% were white; 45% had stage I disease, 60% had extranodal disease, and 55% were given RT. The 5-year OS for the entire cohort was 86%. On UVA, OS was improved with RT (hazard ratio [HR] 0.446, P=.029) and decreased in association with nonwhite race (HR 2.70, P=.006). The 5-year OS was 79% (no RT) and 90% (RT). On MVA, white race and RT remained significantly associated with improved OS (P=.007 and .018, respectively). The use of RT decreased over time: 61% for the 67 patients whose disease was diagnosed from 2001 to 2005 and 53% in the 138 patients treated from 2006 to 2010. Conclusion: This retrospective population-based analysis is the largest PMBCL dataset to date and demonstrates a significant survival benefit associated with RT. Nearly half of patients treated in the United States do not receive RT, and its use appears to be declining. In the absence of phase 3 data, the use of RT should be strongly considered for its survival benefit in early-stage disease.« less
NASA Technical Reports Server (NTRS)
Aspinall, David; Denney, Ewen; Lueth, Christoph
2012-01-01
We motivate and introduce a query language PrQL designed for inspecting machine representations of proofs. PrQL natively supports hiproofs which express proof structure using hierarchical nested labelled trees. The core language presented in this paper is locally structured (first-order), with queries built using recursion and patterns over proof structure and rule names. We define the syntax and semantics of locally structured queries, demonstrate their power, and sketch some implementation experiments.
Smith, E M; Wandtke, J; Robinson, A
1999-05-01
The Medical Information, Communication and Archive System (MICAS) is a multivendor incremental approach to picture archiving and communications system (PACS). It is a multimodality integrated image management system that is seamlessly integrated with the radiology information system (RIS). Phase II enhancements of MICAS include a permanent archive, automated workflow, study caches, Microsoft (Redmond, WA) Windows NT diagnostic workstations with all components adhering to Digital Information Communications in Medicine (DICOM) standards. MICAS is designed as an enterprise-wide PACS to provide images and reports throughout the Strong Health healthcare network. Phase II includes the addition of a Cemax-Icon (Fremont, CA) archive, PACS broker (Mitra, Waterloo, Canada), an interface (IDX PACSlink, Burlington, VT) to the RIS (IDXrad) plus the conversion of the UNIX-based redundant array of inexpensive disks (RAID) 5 temporary archives in phase I to NT-based RAID 0 DICOM modality-specific study caches (ImageLabs, Bedford, MA). The phase I acquisition engines and workflow management software was uninstalled and the Cemax archive manager (AM) assumed these functions. The existing ImageLabs UNIX-based viewing software was enhanced and converted to an NT-based DICOM viewer. Installation of phase II hardware and software and integration with existing components began in July 1998. Phase II of MICAS demonstrates that a multivendor open-system incremental approach to PACS is feasible, cost-effective, and has significant advantages over a single-vendor implementation.
Servagi-Vernat, Stéphanie; Créhange, Gilles; Bonnetain, Franck; Mertens, Cécile; Brain, Etienne; Bosset, Jean François
2017-07-13
The management of elderly patients with cancer is a therapeutic challenge and a public health problem. Definitive chemoradiotherapy (CRT) is an accepted standard treatment for patients with locally advanced esophageal cancer who cannot undergo surgery. However, there are few reports regarding tolerance to CRT in elderly patients. We previously reported results for CRT in patients aged ≥75 years. Following this first phase II trial, we propose to conduct a phase I/II study to evaluate the combination of carboplatin and paclitaxel, with concurrent RT in unresectable esophageal cancer patients aged 75 years or older. This prospective multicenter phase I/II study will include esophageal cancer in patients aged 75 years or older. Study procedures will consist to determinate the tolerated dose of chemotherapy (Carboplatin, paclitaxel) and of radiotherapy (41.4-45 and 50.4 Gy) in the phase I. Efficacy will be assessed using a co-primary endpoint encompassing health related quality of life and the progression-free survival in the phase II with the dose recommended of CRT in the phase I. This geriatric evaluation was defined by the French geriatric oncology group (GERICO). This trial has been designed to assess the tolerated dose of CRT in selected patient aged 75 years or older. Clinicaltrials.gov ID: NCT02735057 . Registered on 18 March 2016.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Herrmann, W.; von Laven, G.M.; Parker, T.
1993-09-01
The Bibliographic Retrieval System (BARS) is a data base management system specially designed to retrieve bibliographic references. Two databases are available, (i) the Sandia Shock Compression (SSC) database which contains over 5700 references to the literature related to stress waves in solids and their applications, and (ii) the Shock Physics Index (SPHINX) which includes over 8000 further references to stress waves in solids, material properties at intermediate and low rates, ballistic and hypervelocity impact, and explosive or shock fabrication methods. There is some overlap in the information in the two data bases.
Effective Multi-Query Expansions: Collaborative Deep Networks for Robust Landmark Retrieval.
Wang, Yang; Lin, Xuemin; Wu, Lin; Zhang, Wenjie
2017-03-01
Given a query photo issued by a user (q-user), the landmark retrieval is to return a set of photos with their landmarks similar to those of the query, while the existing studies on the landmark retrieval focus on exploiting geometries of landmarks for similarity matches between candidate photos and a query photo. We observe that the same landmarks provided by different users over social media community may convey different geometry information depending on the viewpoints and/or angles, and may, subsequently, yield very different results. In fact, dealing with the landmarks with low quality shapes caused by the photography of q-users is often nontrivial and has seldom been studied. In this paper, we propose a novel framework, namely, multi-query expansions, to retrieve semantically robust landmarks by two steps. First, we identify the top- k photos regarding the latent topics of a query landmark to construct multi-query set so as to remedy its possible low quality shape. For this purpose, we significantly extend the techniques of Latent Dirichlet Allocation. Then, motivated by the typical collaborative filtering methods, we propose to learn a collaborative deep networks-based semantically, nonlinear, and high-level features over the latent factor for landmark photo as the training set, which is formed by matrix factorization over collaborative user-photo matrix regarding the multi-query set. The learned deep network is further applied to generate the features for all the other photos, meanwhile resulting into a compact multi-query set within such space. Then, the final ranking scores are calculated over the high-level feature space between the multi-query set and all other photos, which are ranked to serve as the final ranking list of landmark retrieval. Extensive experiments are conducted on real-world social media data with both landmark photos together with their user information to show the superior performance over the existing methods, especially our recently proposed multi-query based mid-level pattern representation method [1].
Benchmarking distributed data warehouse solutions for storing genomic variant information
Wiewiórka, Marek S.; Wysakowicz, Dawid P.; Okoniewski, Michał J.
2017-01-01
Abstract Genomic-based personalized medicine encompasses storing, analysing and interpreting genomic variants as its central issues. At a time when thousands of patientss sequenced exomes and genomes are becoming available, there is a growing need for efficient database storage and querying. The answer could be the application of modern distributed storage systems and query engines. However, the application of large genomic variant databases to this problem has not been sufficiently far explored so far in the literature. To investigate the effectiveness of modern columnar storage [column-oriented Database Management System (DBMS)] and query engines, we have developed a prototypic genomic variant data warehouse, populated with large generated content of genomic variants and phenotypic data. Next, we have benchmarked performance of a number of combinations of distributed storages and query engines on a set of SQL queries that address biological questions essential for both research and medical applications. In addition, a non-distributed, analytical database (MonetDB) has been used as a baseline. Comparison of query execution times confirms that distributed data warehousing solutions outperform classic relational DBMSs. Moreover, pre-aggregation and further denormalization of data, which reduce the number of distributed join operations, significantly improve query performance by several orders of magnitude. Most of distributed back-ends offer a good performance for complex analytical queries, while the Optimized Row Columnar (ORC) format paired with Presto and Parquet with Spark 2 query engines provide, on average, the lowest execution times. Apache Kudu on the other hand, is the only solution that guarantees a sub-second performance for simple genome range queries returning a small subset of data, where low-latency response is expected, while still offering decent performance for running analytical queries. In summary, research and clinical applications that require the storage and analysis of variants from thousands of samples can benefit from the scalability and performance of distributed data warehouse solutions. Database URL: https://github.com/ZSI-Bio/variantsdwh PMID:29220442
Madankumar, Arumugam; Jayakumar, Subramaniyan; Gokuladhas, Krishnan; Rajan, Balan; Raghunandhakumar, Subramanian; Asokkumar, Selvamani; Devaki, Thiruvengadam
2013-04-05
Xenobiotic metabolizing enzymes are chief determinants in both the susceptibility to mutagenic effect of chemical carcinogens and in the response of tumors to chemotherapy. The present study was aimed to analyze the effect of geraniol administration on the activity of phase I and phase II carcinogen metabolizing enzymes through the nuclear factor erythroid 2-related factor-2 (Nrf2) activation against 4-niroquinoline-1-oxide (4NQO) induced oral carcinogenesis. The well-known chemical carcinogen 4NQO (50 ppm) was used to induce oral carcinogenesis through drinking water for 4, 12, and 20 weeks. The degree of cancer progression at each stage was confirmed by histological examination. At the end of the experimental period, 100% tumor formation was observed in the oral cavity of 4NQO induced animals with significant (P<0.05) alteration in the status of tumor markers, tongue and liver phase I and phase II drug metabolizing enzymes indicating progression of disease. Oral administration of geraniol at the dose of 200 mg/kg b.wt., thrice a week to 4NQO induced animals was able to inhibit tumor formation and thereby delayed the progression of oral carcinogenesis by modulating tongue and liver phase I and phase II drug metabolizing enzymes, as substantiated further by the histological and transmission electron microscopic studies. Our results demonstrate that geraniol exerts its chemopreventive potential by altering activities of phases I and II drug metabolizing enzymes to achieve minimum bioactivation of carcinogen and maximum detoxification. Copyright © 2013 Elsevier B.V. All rights reserved.
Querying graphs in protein-protein interactions networks using feedback vertex set.
Blin, Guillaume; Sikora, Florian; Vialette, Stéphane
2010-01-01
Recent techniques increase rapidly the amount of our knowledge on interactions between proteins. The interpretation of these new information depends on our ability to retrieve known substructures in the data, the Protein-Protein Interactions (PPIs) networks. In an algorithmic point of view, it is an hard task since it often leads to NP-hard problems. To overcome this difficulty, many authors have provided tools for querying patterns with a restricted topology, i.e., paths or trees in PPI networks. Such restriction leads to the development of fixed parameter tractable (FPT) algorithms, which can be practicable for restricted sizes of queries. Unfortunately, Graph Homomorphism is a W[1]-hard problem, and hence, no FPT algorithm can be found when patterns are in the shape of general graphs. However, Dost et al. gave an algorithm (which is not implemented) to query graphs with a bounded treewidth in PPI networks (the treewidth of the query being involved in the time complexity). In this paper, we propose another algorithm for querying pattern in the shape of graphs, also based on dynamic programming and the color-coding technique. To transform graphs queries into trees without loss of informations, we use feedback vertex set coupled to a node duplication mechanism. Hence, our algorithm is FPT for querying graphs with a bounded size of their feedback vertex set. It gives an alternative to the treewidth parameter, which can be better or worst for a given query. We provide a python implementation which allows us to validate our implementation on real data. Especially, we retrieve some human queries in the shape of graphs into the fly PPI network.
Hanauer, David A; Wu, Danny T Y; Yang, Lei; Mei, Qiaozhu; Murkowski-Steffy, Katherine B; Vydiswaran, V G Vinod; Zheng, Kai
2017-03-01
The utility of biomedical information retrieval environments can be severely limited when users lack expertise in constructing effective search queries. To address this issue, we developed a computer-based query recommendation algorithm that suggests semantically interchangeable terms based on an initial user-entered query. In this study, we assessed the value of this approach, which has broad applicability in biomedical information retrieval, by demonstrating its application as part of a search engine that facilitates retrieval of information from electronic health records (EHRs). The query recommendation algorithm utilizes MetaMap to identify medical concepts from search queries and indexed EHR documents. Synonym variants from UMLS are used to expand the concepts along with a synonym set curated from historical EHR search logs. The empirical study involved 33 clinicians and staff who evaluated the system through a set of simulated EHR search tasks. User acceptance was assessed using the widely used technology acceptance model. The search engine's performance was rated consistently higher with the query recommendation feature turned on vs. off. The relevance of computer-recommended search terms was also rated high, and in most cases the participants had not thought of these terms on their own. The questions on perceived usefulness and perceived ease of use received overwhelmingly positive responses. A vast majority of the participants wanted the query recommendation feature to be available to assist in their day-to-day EHR search tasks. Challenges persist for users to construct effective search queries when retrieving information from biomedical documents including those from EHRs. This study demonstrates that semantically-based query recommendation is a viable solution to addressing this challenge. Published by Elsevier Inc.
Machiela, Mitchell J; Chanock, Stephen J
2015-11-01
Assessing linkage disequilibrium (LD) across ancestral populations is a powerful approach for investigating population-specific genetic structure as well as functionally mapping regions of disease susceptibility. Here, we present LDlink, a web-based collection of bioinformatic modules that query single nucleotide polymorphisms (SNPs) in population groups of interest to generate haplotype tables and interactive plots. Modules are designed with an emphasis on ease of use, query flexibility, and interactive visualization of results. Phase 3 haplotype data from the 1000 Genomes Project are referenced for calculating pairwise metrics of LD, searching for proxies in high LD, and enumerating all observed haplotypes. LDlink is tailored for investigators interested in mapping common and uncommon disease susceptibility loci by focusing on output linking correlated alleles and highlighting putative functional variants. LDlink is a free and publically available web tool which can be accessed at http://analysistools.nci.nih.gov/LDlink/. mitchell.machiela@nih.gov. Published by Oxford University Press 2015. This work is written by US Government employees and is in the public domain in the US.
Espeau, Philippe; Céolin, René; Tamarit, Josep-Lluis; Perrin, Marc-Antoine; Gauchi, Jean-Pierre; Leveiller, Franck
2005-03-01
The thermodynamic relationships between the two known polymorphs of paracetamol have been investigated, and the subsequent pressure-temperature and temperature-volume phase diagrams were constructed using data from crystallographic and calorimetric measurements as a function of the temperature. Irrespective of temperature, monoclinic Form I and orthorhombic Form II are stable phases at ordinary and high pressures, respectively. The I and II phase regions in the pressure-temperature diagram are bordered by the I-II equilibrium curve, for which a negative slope (dp/dT approximately -0.3 MPa x K(-1)) was determined although it was not observed experimentally. This curve goes through the I-II-liquid triple point whose coordinates (p approximately 234 MPa, T approximately 505 K) correspond to the crossing point of the melting curves, for which dp/dT values of +3.75 MPa x K(-1) (I) and +3.14 MPa x K(-1) (II) were calculated from enthalpy and volume changes upon fusion. More generally, this case exemplifies how the stability hierarchy of polymorphs may be inferred from the difference in their sublimation curves, as topologically positioned with respect to each other, using the phase rule and simple inferences resorting to Gibbs equilibrium thermodynamics. Copyright 2004 Wiley-Liss, Inc. and the American Pharmacists Association.
Occam's razor: supporting visual query expression for content-based image queries
NASA Astrophysics Data System (ADS)
Venters, Colin C.; Hartley, Richard J.; Hewitt, William T.
2005-01-01
This paper reports the results of a usability experiment that investigated visual query formulation on three dimensions: effectiveness, efficiency, and user satisfaction. Twenty eight evaluation sessions were conducted in order to assess the extent to which query by visual example supports visual query formulation in a content-based image retrieval environment. In order to provide a context and focus for the investigation, the study was segmented by image type, user group, and use function. The image type consisted of a set of abstract geometric device marks supplied by the UK Trademark Registry. Users were selected from the 14 UK Patent Information Network offices. The use function was limited to the retrieval of images by shape similarity. Two client interfaces were developed for comparison purposes: Trademark Image Browser Engine (TRIBE) and Shape Query Image Retrieval Systems Engine (SQUIRE).
Geometric Representations of Condition Queries on Three-Dimensional Vector Fields
NASA Technical Reports Server (NTRS)
Henze, Chris
1999-01-01
Condition queries on distributed data ask where particular conditions are satisfied. It is possible to represent condition queries as geometric objects by plotting field data in various spaces derived from the data, and by selecting loci within these derived spaces which signify the desired conditions. Rather simple geometric partitions of derived spaces can represent complex condition queries because much complexity can be encapsulated in the derived space mapping itself A geometric view of condition queries provides a useful conceptual unification, allowing one to intuitively understand many existing vector field feature detection algorithms -- and to design new ones -- as variations on a common theme. A geometric representation of condition queries also provides a simple and coherent basis for computer implementation, reducing a wide variety of existing and potential vector field feature detection techniques to a few simple geometric operations.
Occam"s razor: supporting visual query expression for content-based image queries
NASA Astrophysics Data System (ADS)
Venters, Colin C.; Hartley, Richard J.; Hewitt, William T.
2004-12-01
This paper reports the results of a usability experiment that investigated visual query formulation on three dimensions: effectiveness, efficiency, and user satisfaction. Twenty eight evaluation sessions were conducted in order to assess the extent to which query by visual example supports visual query formulation in a content-based image retrieval environment. In order to provide a context and focus for the investigation, the study was segmented by image type, user group, and use function. The image type consisted of a set of abstract geometric device marks supplied by the UK Trademark Registry. Users were selected from the 14 UK Patent Information Network offices. The use function was limited to the retrieval of images by shape similarity. Two client interfaces were developed for comparison purposes: Trademark Image Browser Engine (TRIBE) and Shape Query Image Retrieval Systems Engine (SQUIRE).
Retrieval feedback in MEDLINE.
Srinivasan, P
1996-01-01
OBJECTIVE: To investigate a new approach for query expansion based on retrieval feedback. The first objective in this study was to examine alternative query-expansion methods within the same retrieval-feedback framework. The three alternatives proposed are: expansion on the MeSH query field alone, expansion on the free-text field alone, and expansion on both the MeSH and the free-text fields. The second objective was to gain further understanding of retrieval feedback by examining possible dependencies on relevant documents during the feedback cycle. DESIGN: Comparative study of retrieval effectiveness using the original unexpanded and the alternative expanded user queries on a MEDLINE test collection of 75 queries and 2,334 MEDLINE citations. MEASUREMENTS: Retrieval effectivenesses of the original unexpanded and the alternative expanded queries were compared using 11-point-average precision scores (11-AvgP). These are averages of precision scores obtained at 11 standard recall points. RESULTS: All three expansion strategies significantly improved the original queries in terms of retrieval effectiveness. Expansion on MeSH alone was equivalent to expansion on both MeSH and the free-text fields. Expansion on the free-text field alone improved the queries significantly less than did the other two strategies. The second part of the study indicated that retrieval-feedback-based expansion yields significant performance improvements independent of the availability of relevant documents for feedback information. CONCLUSIONS: Retrieval feedback offers a robust procedure for query expansion that is most effective for MEDLINE when applied to the MeSH field. PMID:8653452
DOT National Transportation Integrated Search
2009-12-01
A new web-based geotechnical Geographic Information System (GeoGIS) was developed and tested for the Alabama Department of Transportation (ALDOT) during Phase II of this research project. This web-based system stores geotechnical information about tr...
DOT National Transportation Integrated Search
1985-10-01
This report summarizes the findings from the second phase of a two-part analysis of hazardous materials truck routes in the Dallas-Fort Worth area. Phase II of this study analyzes the risk of transporting hazardous materials on freeways and arterial ...
Federal Register 2010, 2011, 2012, 2013, 2014
2010-01-29
... activity.\\5\\ There will also be an increase in the monthly fee for the mutual fund Profile Phase II Service.... Profile Phase I transmits mutual fund price and rate information. Profile Phase II stores data elements such as accumulation, breakpoints, and commission eligibility that relate to mutual fund processing...
The report describes Phase II of a demonstration of the utilization of commercial phosphoric acid fuel cells to recover energy from landfill gas. This phase consisted primarily of the construction and testing of a Gas Pretreatment Unit (GPU) whose function is to remove those impu...
DOT National Transportation Integrated Search
2016-03-07
Building on the success of developing a UAV based unpaved road assessment system in Phase I, the project team was awarded a Phase II project by the USDOT to focus on outreach and implementation. The project team added Valerie Lefler of Integrated Glo...
A Reliability Simulator for Radiation-Hard Microelectronics Development
1991-07-01
1 3.0 PHASE II WORK PLANS ................................................................ 2... plan . The correlation experimental details including the devices utilized, the hot-carrier stressing and the wafer-level radiation correlation procedure...channel devices, and a new lifetime extrapolation method is demonstrated for p-channel devices. 3.0 PHASE II WORK PLANS The Phase 1I program consisted of
Query Expansion Using SNOMED-CT and Weighing Schemes
2014-11-01
For this research, we have used SNOMED-CT along with UMLS Methathesaurus as our ontology in medical domain to expand the queries. General Terms...CT along with UMLS Methathesaurus as our ontology in medical domain to expand the queries. 15. SUBJECT TERMS 16. SECURITY CLASSIFICATION OF: 17...University of the Basque country discuss their finding on query expansion using external sources headlined by Unified Medical Language System ( UMLS
ERIC Educational Resources Information Center
Chung, EunKyung; Yoon, JungWon
2009-01-01
Introduction: The purpose of this study is to compare characteristics and features of user supplied tags and search query terms for images on the "Flickr" Website in terms of categories of pictorial meanings and level of term specificity. Method: This study focuses on comparisons between tags and search queries using Shatford's categorization…
Design Recommendations for Query Languages
1980-09-01
DESIGN RECOMMENDATIONS FOR QUERY LANGUAGES S.L. Ehrenreich Submitted by: Stanley M. Halpin, Acting Chief HUMAN FACTORS TECHNICAL AREA Approved by: Edgar ...respond to que- ries that it recognizes as faulty. Codd (1974) states that in designing a nat- ural query language, attention must be given to dealing...impaired. Codd (1974) also regarded the user’s perception of the data base to be of critical importance in properly designing a query language system
Agent-Based Framework for Discrete Entity Simulations
2006-11-01
Postgres database server for environment queries of neighbors and continuum data. As expected for raw database queries (no database optimizations in...form. Eventually the code was ported to GNU C++ on the same single Intel Pentium 4 CPU running RedHat Linux 9.0 and Postgres database server...Again Postgres was used for environmental queries, and the tool remained relatively slow because of the immense number of queries necessary to assess
Akce, Abdullah; Norton, James J S; Bretl, Timothy
2015-09-01
This paper presents a brain-computer interface for text entry using steady-state visually evoked potentials (SSVEP). Like other SSVEP-based spellers, ours identifies the desired input character by posing questions (or queries) to users through a visual interface. Each query defines a mapping from possible characters to steady-state stimuli. The user responds by attending to one of these stimuli. Unlike other SSVEP-based spellers, ours chooses from a much larger pool of possible queries-on the order of ten thousand instead of ten. The larger query pool allows our speller to adapt more effectively to the inherent structure of what is being typed and to the input performance of the user, both of which make certain queries provide more information than others. In particular, our speller chooses queries from this pool that maximize the amount of information to be received per unit of time, a measure of mutual information that we call information gain rate. To validate our interface, we compared it with two other state-of-the-art SSVEP-based spellers, which were re-implemented to use the same input mechanism. Results showed that our interface, with the larger query pool, allowed users to spell multiple-word texts nearly twice as fast as they could with the compared spellers.
Query construction, entropy, and generalization in neural-network models
NASA Astrophysics Data System (ADS)
Sollich, Peter
1994-05-01
We study query construction algorithms, which aim at improving the generalization ability of systems that learn from examples by choosing optimal, nonredundant training sets. We set up a general probabilistic framework for deriving such algorithms from the requirement of optimizing a suitable objective function; specifically, we consider the objective functions entropy (or information gain) and generalization error. For two learning scenarios, the high-low game and the linear perceptron, we evaluate the generalization performance obtained by applying the corresponding query construction algorithms and compare it to training on random examples. We find qualitative differences between the two scenarios due to the different structure of the underlying rules (nonlinear and ``noninvertible'' versus linear); in particular, for the linear perceptron, random examples lead to the same generalization ability as a sequence of queries in the limit of an infinite number of examples. We also investigate learning algorithms which are ill matched to the learning environment and find that, in this case, minimum entropy queries can in fact yield a lower generalization ability than random examples. Finally, we study the efficiency of single queries and its dependence on the learning history, i.e., on whether the previous training examples were generated randomly or by querying, and the difference between globally and locally optimal query construction.
Spatial information semantic query based on SPARQL
NASA Astrophysics Data System (ADS)
Xiao, Zhifeng; Huang, Lei; Zhai, Xiaofang
2009-10-01
How can the efficiency of spatial information inquiries be enhanced in today's fast-growing information age? We are rich in geospatial data but poor in up-to-date geospatial information and knowledge that are ready to be accessed by public users. This paper adopts an approach for querying spatial semantic by building an Web Ontology language(OWL) format ontology and introducing SPARQL Protocol and RDF Query Language(SPARQL) to search spatial semantic relations. It is important to establish spatial semantics that support for effective spatial reasoning for performing semantic query. Compared to earlier keyword-based and information retrieval techniques that rely on syntax, we use semantic approaches in our spatial queries system. Semantic approaches need to be developed by ontology, so we use OWL to describe spatial information extracted by the large-scale map of Wuhan. Spatial information expressed by ontology with formal semantics is available to machines for processing and to people for understanding. The approach is illustrated by introducing a case study for using SPARQL to query geo-spatial ontology instances of Wuhan. The paper shows that making use of SPARQL to search OWL ontology instances can ensure the result's accuracy and applicability. The result also indicates constructing a geo-spatial semantic query system has positive efforts on forming spatial query and retrieval.
DISPAQ: Distributed Profitable-Area Query from Big Taxi Trip Data.
Putri, Fadhilah Kurnia; Song, Giltae; Kwon, Joonho; Rao, Praveen
2017-09-25
One of the crucial problems for taxi drivers is to efficiently locate passengers in order to increase profits. The rapid advancement and ubiquitous penetration of Internet of Things (IoT) technology into transportation industries enables us to provide taxi drivers with locations that have more potential passengers (more profitable areas) by analyzing and querying taxi trip data. In this paper, we propose a query processing system, called Distributed Profitable-Area Query ( DISPAQ ) which efficiently identifies profitable areas by exploiting the Apache Software Foundation's Spark framework and a MongoDB database. DISPAQ first maintains a profitable-area query index (PQ-index) by extracting area summaries and route summaries from raw taxi trip data. It then identifies candidate profitable areas by searching the PQ-index during query processing. Then, it exploits a Z-Skyline algorithm, which is an extension of skyline processing with a Z-order space filling curve, to quickly refine the candidate profitable areas. To improve the performance of distributed query processing, we also propose local Z-Skyline optimization, which reduces the number of dominant tests by distributing killer profitable areas to each cluster node. Through extensive evaluation with real datasets, we demonstrate that our DISPAQ system provides a scalable and efficient solution for processing profitable-area queries from huge amounts of big taxi trip data.
Zhou, ZhangBing; Zhao, Deng; Shu, Lei; Tsang, Kim-Fung
2015-01-01
Wireless sensor networks, serving as an important interface between physical environments and computational systems, have been used extensively for supporting domain applications, where multiple-attribute sensory data are queried from the network continuously and periodically. Usually, certain sensory data may not vary significantly within a certain time duration for certain applications. In this setting, sensory data gathered at a certain time slot can be used for answering concurrent queries and may be reused for answering the forthcoming queries when the variation of these data is within a certain threshold. To address this challenge, a popularity-based cooperative caching mechanism is proposed in this article, where the popularity of sensory data is calculated according to the queries issued in recent time slots. This popularity reflects the possibility that sensory data are interested in the forthcoming queries. Generally, sensory data with the highest popularity are cached at the sink node, while sensory data that may not be interested in the forthcoming queries are cached in the head nodes of divided grid cells. Leveraging these cooperatively cached sensory data, queries are answered through composing these two-tier cached data. Experimental evaluation shows that this approach can reduce the network communication cost significantly and increase the network capability. PMID:26131665
DISPAQ: Distributed Profitable-Area Query from Big Taxi Trip Data †
Putri, Fadhilah Kurnia; Song, Giltae; Rao, Praveen
2017-01-01
One of the crucial problems for taxi drivers is to efficiently locate passengers in order to increase profits. The rapid advancement and ubiquitous penetration of Internet of Things (IoT) technology into transportation industries enables us to provide taxi drivers with locations that have more potential passengers (more profitable areas) by analyzing and querying taxi trip data. In this paper, we propose a query processing system, called Distributed Profitable-Area Query (DISPAQ) which efficiently identifies profitable areas by exploiting the Apache Software Foundation’s Spark framework and a MongoDB database. DISPAQ first maintains a profitable-area query index (PQ-index) by extracting area summaries and route summaries from raw taxi trip data. It then identifies candidate profitable areas by searching the PQ-index during query processing. Then, it exploits a Z-Skyline algorithm, which is an extension of skyline processing with a Z-order space filling curve, to quickly refine the candidate profitable areas. To improve the performance of distributed query processing, we also propose local Z-Skyline optimization, which reduces the number of dominant tests by distributing killer profitable areas to each cluster node. Through extensive evaluation with real datasets, we demonstrate that our DISPAQ system provides a scalable and efficient solution for processing profitable-area queries from huge amounts of big taxi trip data. PMID:28946679
ERIC Educational Resources Information Center
Price Waterhouse and Co., New York, NY.
This volume on Phase II of the New York State Educational Information System (NYSEIS) describes the Gross Systems Analysis and Design, which includes the general flow diagram and processing chart for each of the student, personnel, and financial subsystems. Volume II, Functional Specifications, includes input/output requirements and file…
NASA Technical Reports Server (NTRS)
Crawford, Winifred C.
2010-01-01
The AMU created new logistic regression equations in an effort to increase the skill of the Objective Lightning Forecast Tool developed in Phase II (Lambert 2007). One equation was created for each of five sub-seasons based on the daily lightning climatology instead of by month as was done in Phase II. The assumption was that these equations would capture the physical attributes that contribute to thunderstorm formation more so than monthly equations. However, the SS values in Section 5.3.2 showed that the Phase III equations had worse skill than the Phase II equations and, therefore, will not be transitioned into operations. The current Objective Lightning Forecast Tool developed in Phase II will continue to be used operationally in MIDDS. Three warm seasons were added to the Phase II dataset to increase the POR from 17 to 20 years (1989-2008), and data for October were included since the daily climatology showed lightning occurrence extending into that month. None of the three methods tested to determine the start of the subseason in each individual year were able to discern the start dates with consistent accuracy. Therefore, the start dates were determined by the daily climatology shown in Figure 10 and were the same in every year. The procedures used to create the predictors and develop the equations were identical to those in Phase II. The equations were made up of one to three predictors. TI and the flow regime probabilities were the top predictors followed by 1-day persistence, then VT and Ll. Each equation outperformed four other forecast methods by 7-57% using the verification dataset, but the new equations were outperformed by the Phase II equations in every sub-season. The reason for the degradation may be due to the fact that the same sub-season start dates were used in every year. It is likely there was overlap of sub-season days at the beginning and end of each defined sub-season in each individual year, which could very well affect equation performance.
Bellmunt, J; Kerst, J M; Vázquez, F; Morales-Barrera, R; Grande, E; Medina, A; González Graguera, M B; Rubio, G; Anido, U; Fernández Calvo, O; González-Billalabeitia, E; Van den Eertwegh, A J M; Pujol, E; Perez-Gracia, J L; González Larriba, J L; Collado, R; Los, M; Maciá, S; De Wit, R
2017-07-01
Despite the advent of immunotherapy in urothelial cancer, there is still a need to find effective cytotoxic agents beyond first and second lines. Vinflunine is the only treatment approved in this setting by the European Medicines Agency and taxanes are also widely used in second line. Cabazitaxel is a taxane with activity in docetaxel-refractory cancers. A randomized study was conducted to compare its efficacy versus vinflunine. This is a multicenter, randomized, open-label, phase II/III study, following a Simon's optimal method with stopping rules based on an interim futility analysis and a formal efficacy analysis at the end of the phase II. ECOG Performance Status, anaemia and liver metastases were stratification factors. Primary objectives were overall response rate for the phase II and overall survival for the phase III. Seventy patients were included in the phase II across 19 institutions in Europe. Baseline characteristics were well balanced between the two arms. Three patients (13%) obtained a partial response on cabazitaxel (95% CI 2.7-32.4) and six patients (30%) in the vinflunine arm (95% CI 11.9-54.3). Median progression-free survival for cabazitaxel was 1.9 versus 2.9 months for vinflunine (P = 0.039). The study did not proceed to phase III since the futility analysis showed a lack of efficacy of cabazitaxel. A trend for overall survival benefit was found favouring vinflunine (median 7.6 versus 5.5 months). Grade 3- to 4-related adverse events were seen in 41% patients with no difference between the two arms. This phase II/III second line bladder study comparing cabazitaxel with vinflunine was closed when the phase II showed a lack of efficacy of the cabazitaxel arm. Vinflunine results were consistent with those known previously. NCT01830231. © The Author 2017. Published by Oxford University Press on behalf of the European Society for Medical Oncology. All rights reserved. For Permissions, please email: journals.permissions@oup.com.
Zhang, Qiang; Pi, Jingbo; Woods, Courtney G; Andersen, Melvin E
2009-06-15
Hormetic responses to xenobiotic exposure likely occur as a result of overcompensation by the homeostatic control systems operating in biological organisms. However, the mechanisms underlying overcompensation that leads to hormesis are still unclear. A well-known homeostatic circuit in the cell is the gene induction network comprising phase I, II and III metabolizing enzymes, which are responsible for xenobiotic detoxification, and in many cases, bioactivation. By formulating a differential equation-based computational model, we investigated in this study whether hormesis can arise from the operation of this gene/enzyme network. The model consists of two feedback and one feedforward controls. With the phase I negative feedback control, xenobiotic X activates nuclear receptors to induce cytochrome P450 enzyme, which bioactivates X into a reactive metabolite X'. With the phase II negative feedback control, X' activates transcription factor Nrf2 to induce phase II enzymes such as glutathione S-transferase and glutamate cysteine ligase, etc., which participate in a set of reactions that lead to the metabolism of X' into a less toxic conjugate X''. The feedforward control involves phase I to II cross-induction, in which the parent chemical X can also induce phase II enzymes directly through the nuclear receptor and indirectly through transcriptionally upregulating Nrf2. As a result of the active feedforward control, a steady-state hormetic relationship readily arises between the concentrations of the reactive metabolite X' and the extracellular parent chemical X to which the cell is exposed. The shape of dose-response evolves over time from initially monotonically increasing to J-shaped at the final steady state-a temporal sequence consistent with adaptation-mediated hormesis. The magnitude of the hormetic response is enhanced by increases in the feedforward gain, but attenuated by increases in the bioactivation or phase II feedback loop gains. Our study suggests a possibly common mechanism for the hormetic responses observed with many mutagens/carcinogens whose activities require bioactivation by phase I enzymes. Feedforward control, often operating in combination with negative feedback regulation in a homeostatic system, may be a general control theme responsible for steady-state hormesis.
VPipe: Virtual Pipelining for Scheduling of DAG Stream Query Plans
NASA Astrophysics Data System (ADS)
Wang, Song; Gupta, Chetan; Mehta, Abhay
There are data streams all around us that can be harnessed for tremendous business and personal advantage. For an enterprise-level stream processing system such as CHAOS [1] (Continuous, Heterogeneous Analytic Over Streams), handling of complex query plans with resource constraints is challenging. While several scheduling strategies exist for stream processing, efficient scheduling of complex DAG query plans is still largely unsolved. In this paper, we propose a novel execution scheme for scheduling complex directed acyclic graph (DAG) query plans with meta-data enriched stream tuples. Our solution, called Virtual Pipelined Chain (or VPipe Chain for short), effectively extends the "Chain" pipelining scheduling approach to complex DAG query plans.
NASA Astrophysics Data System (ADS)
Warren, Z.; Shahriar, M. S.; Tripathi, R.; Pati, G. S.
2018-02-01
A repeated query technique has been demonstrated as a new interrogation method in pulsed coherent population trapping for producing single-peaked Ramsey interference with high contrast. This technique enhances the contrast of the central Ramsey fringe by nearly 1.5 times and significantly suppresses the side fringes by using more query pulses ( >10) in the pulse cycle. Theoretical models have been developed to simulate Ramsey interference and analyze the characteristics of the Ramsey spectrum produced by the repeated query technique. Experiments have also been carried out employing a repeated query technique in a prototype rubidium clock to study its frequency stability performance.
Nadkarni, P M
1997-08-01
Concept Locator (CL) is a client-server application that accesses a Sybase relational database server containing a subset of the UMLS Metathesaurus for the purpose of retrieval of concepts corresponding to one or more query expressions supplied to it. CL's query grammar permits complex Boolean expressions, wildcard patterns, and parenthesized (nested) subexpressions. CL translates the query expressions supplied to it into one or more SQL statements that actually perform the retrieval. The generated SQL is optimized by the client to take advantage of the strengths of the server's query optimizer, and sidesteps its weaknesses, so that execution is reasonably efficient.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Starcher, Autumn N.; Li, Wei; Kukkadapu, Ravi K.
Fe(II)-Al(III)-LDH (layered double hydroxide) phases have been shown to form from reactions of aqueous Fe(II) with Fe-free Al-bearing minerals (phyllosilicate/clays and Al-oxides). To our knowledge, the effect of small amounts of structural Fe(III) impurities in “neutral” clays on such reactions, however, were not studied. In this study to understand the role of structural Fe(III) impurity in clays, laboratory batch studies with pyrophyllite (10 g/L), an Al-bearing phyllosilicate, containing small amounts of structural Fe(III) impurities and 0.8 mM and 3 mM Fe(II) (both natural and enriched in 57Fe) were carried out at pH 7.5 under anaerobic conditions (4% H2 – 96%more » N2 atmosphere). Samples were taken up to 4 weeks for analysis by Fe-X-ray absorption spectroscopy and 57Fe Mössbauer spectroscopy. In addition to the precipitation of Fe(II)-Al(III)-LDH phases as observed in earlier studies with pure minerals (no Fe(III) impurities in the minerals), the analyses indicated formation of small amounts of Fe(III) containing solid(s), most probably hybrid a Fe(II)-Al(III)/Fe(III)-LDH phase. The mechanism of Fe(II) oxidation was not apparent but most likely was due to interfacial electron transfer from the sorbed Fe(II) to the structural Fe(III) and/or surface-sorption-induced electron-transfer from the sorbed Fe(II) to the clay lattice. Increase in the Fe(II)/Al ratio of the LDH with reaction time further indicated the complex nature of the samples. This research provides evidence for the formation of both Fe(II)-Al(III)-LDH and Fe(II)-Fe(III)/Al(III)-LDH-like phases during reactions of Fe(II) in systems that mimic the natural environments. Better understanding Fe phase formation in complex laboratory studies will improve models of natural redox systems.« less
An alternative database approach for management of SNOMED CT and improved patient data queries.
Campbell, W Scott; Pedersen, Jay; McClay, James C; Rao, Praveen; Bastola, Dhundy; Campbell, James R
2015-10-01
SNOMED CT is the international lingua franca of terminologies for human health. Based in Description Logics (DL), the terminology enables data queries that incorporate inferences between data elements, as well as, those relationships that are explicitly stated. However, the ontologic and polyhierarchical nature of the SNOMED CT concept model make it difficult to implement in its entirety within electronic health record systems that largely employ object oriented or relational database architectures. The result is a reduction of data richness, limitations of query capability and increased systems overhead. The hypothesis of this research was that a graph database (graph DB) architecture using SNOMED CT as the basis for the data model and subsequently modeling patient data upon the semantic core of SNOMED CT could exploit the full value of the terminology to enrich and support advanced data querying capability of patient data sets. The hypothesis was tested by instantiating a graph DB with the fully classified SNOMED CT concept model. The graph DB instance was tested for integrity by calculating the transitive closure table for the SNOMED CT hierarchy and comparing the results with transitive closure tables created using current, validated methods. The graph DB was then populated with 461,171 anonymized patient record fragments and over 2.1 million associated SNOMED CT clinical findings. Queries, including concept negation and disjunction, were then run against the graph database and an enterprise Oracle relational database (RDBMS) of the same patient data sets. The graph DB was then populated with laboratory data encoded using LOINC, as well as, medication data encoded with RxNorm and complex queries performed using LOINC, RxNorm and SNOMED CT to identify uniquely described patient populations. A graph database instance was successfully created for two international releases of SNOMED CT and two US SNOMED CT editions. Transitive closure tables and descriptive statistics generated using the graph database were identical to those using validated methods. Patient queries produced identical patient count results to the Oracle RDBMS with comparable times. Database queries involving defining attributes of SNOMED CT concepts were possible with the graph DB. The same queries could not be directly performed with the Oracle RDBMS representation of the patient data and required the creation and use of external terminology services. Further, queries of undefined depth were successful in identifying unknown relationships between patient cohorts. The results of this study supported the hypothesis that a patient database built upon and around the semantic model of SNOMED CT was possible. The model supported queries that leveraged all aspects of the SNOMED CT logical model to produce clinically relevant query results. Logical disjunction and negation queries were possible using the data model, as well as, queries that extended beyond the structural IS_A hierarchy of SNOMED CT to include queries that employed defining attribute-values of SNOMED CT concepts as search parameters. As medical terminologies, such as SNOMED CT, continue to expand, they will become more complex and model consistency will be more difficult to assure. Simultaneously, consumers of data will increasingly demand improvements to query functionality to accommodate additional granularity of clinical concepts without sacrificing speed. This new line of research provides an alternative approach to instantiating and querying patient data represented using advanced computable clinical terminologies. Copyright © 2015 Elsevier Inc. All rights reserved.
Content-Aware DataGuide with Incremental Index Update using Frequently Used Paths
NASA Astrophysics Data System (ADS)
Sharma, A. K.; Duhan, Neelam; Khattar, Priyanka
2010-11-01
Size of the WWW is increasing day by day. Due to the absence of structured data on the Web, it becomes very difficult for information retrieval tools to fully utilize the Web information. As a solution to this problem, XML pages come into play, which provide structural information to the users to some extent. Without efficient indexes, query processing can be quite inefficient due to an exhaustive traversal on XML data. In this paper an improved content-centric approach of Content-Aware DataGuide, which is an indexing technique for XML databases, is being proposed that uses frequently used paths from historical query logs to improve query performance. The index can be updated incrementally according to the changes in query workload and thus, the overhead of reconstruction can be minimized. Frequently used paths are extracted using any Sequential Pattern mining algorithm on subsequent queries in the query workload. After this, the data structures are incrementally updated. This indexing technique proves to be efficient as partial matching queries can be executed efficiently and users can now get the more relevant documents in results.
Anal Cancer: An Examination of Radiotherapy Strategies
DOE Office of Scientific and Technical Information (OSTI.GOV)
Glynne-Jones, Rob; Lim, Faye
2011-04-01
The Radiation Therapy Oncology Group 9811, ACCORD-03, and ACT II Phase III trials in anal cancer showed no benefit for cisplatin-based induction and maintenance chemotherapy, or radiation dose-escalation >59 Gy. This review examines the efficacy and toxicity of chemoradiation (CRT) in anal cancer, and discusses potential alternative radiotherapy strategies. The evidence for the review was compiled from randomized and nonrandomized trials of radiation therapy and CRT. A total of 103 retrospective/observational studies, 4 Phase I/II studies, 16 Phase II prospective studies, 2 randomized Phase II studies, and 6 Phase III trials of radiotherapy or chemoradiation were identified. There are nomore » meta-analyses based on individual patient data. A 'one-size-fits-all' approach for all stages of anal cancer is inappropriate. Early T1 tumors are probably currently overtreated, whereas T3/T4 lesions might merit escalation of treatment. Intensity-modulated radiotherapy or the integration of biological therapy may play a role in future.« less
Małuszyńska, Hanna; Czarnecki, Piotr; Czarnecka, Anna; Pająk, Zdzisław
2012-04-01
Pyridinium chlorochromate, [C(5)H(5)NH](+)[ClCrO(3)](-) (hereafter referred to as PyClCrO(3)), was studied by X-ray diffraction, differential scanning calorimetry (DSC) and dielectric methods. Studies reveal three reversible phase transitions at 346, 316 and 170 K with the following phase sequence: R ̅3m (I) → R3m (II) → Cm (III) → Cc (IV), c' = 2c. PyClCrO(3) is the first pyridinium salt in which all four phases have been successfully characterized by a single-crystal X-ray diffraction method. Structural results together with dielectric and calorimetric studies allow the classification of the two intermediate phases (II) and (III) as ferroelectric with the Curie point at 346 K, and the lowest phase (IV) as most probably ferroelectric. The ferroelectric hysteresis loop was observed only in phase (III). The high ionic conductivity hindered its observation in phase (II).
Levin, Bruce; Thompson, John L P; Chakraborty, Bibhas; Levy, Gilberto; MacArthur, Robert; Haley, E Clarke
2011-08-01
TNK-S2B, an innovative, randomized, seamless phase II/III trial of tenecteplase versus rt-PA for acute ischemic stroke, terminated for slow enrollment before regulatory approval of use of phase II patients in phase III. (1) To review the trial design and comprehensive type I error rate simulations and (2) to discuss issues raised during regulatory review, to facilitate future approval of similar designs. In phase II, an early (24-h) outcome and adaptive sequential procedure selected one of three tenecteplase doses for phase III comparison with rt-PA. Decision rules comparing this dose to rt-PA would cause stopping for futility at phase II end, or continuation to phase III. Phase III incorporated two co-primary hypotheses, allowing for a treatment effect at either end of the trichotomized Rankin scale. Assuming no early termination, four interim analyses and one final analysis of 1908 patients provided an experiment-wise type I error rate of <0.05. Over 1,000 distribution scenarios, each involving 40,000 replications, the maximum type I error in phase III was 0.038. Inflation from the dose selection was more than offset by the one-half continuity correction in the test statistics. Inflation from repeated interim analyses was more than offset by the reduction from the clinical stopping rules for futility at the first interim analysis. Design complexity and evolving regulatory requirements lengthened the review process. (1) The design was innovative and efficient. Per protocol, type I error was well controlled for the co-primary phase III hypothesis tests, and experiment-wise. (2a) Time must be allowed for communications with regulatory reviewers from first design stages. (2b) Adequate type I error control must be demonstrated. (2c) Greater clarity is needed on (i) whether this includes demonstration of type I error control if the protocol is violated and (ii) whether simulations of type I error control are acceptable. (2d) Regulatory agency concerns that protocols for futility stopping may not be followed may be allayed by submitting interim analysis results to them as these analyses occur.
ERIC Educational Resources Information Center
Wake Forest Univ., Winston Salem, NC. Bowman Gray School of Medicine.
This publication contains a curriculum developed through functional job analyses for a 24-month physician's assistant training program. Phase 1 of the 3-phase program is a 6-month basic course program in clinical and bioscience principles and is required of all students regardless of their specialty interest. Phase 2 is a 6 to 10 month period of…
47 CFR 90.765 - Licenses term for Phase II licenses.
Code of Federal Regulations, 2010 CFR
2010-10-01
... 220-222 MHz Band Policies Governing the Licensing and Use of Phase II Ea, Regional and Nationwide...(a), EA and Regional licenses authorized pursuant to § 90.761, and non-nationwide licenses authorized...
47 CFR 90.765 - Licenses term for Phase II licenses.
Code of Federal Regulations, 2011 CFR
2011-10-01
... 220-222 MHz Band Policies Governing the Licensing and Use of Phase II Ea, Regional and Nationwide...(a), EA and Regional licenses authorized pursuant to § 90.761, and non-nationwide licenses authorized...
Planning Targets for Phase II Watershed Implementation Plans
On August 1, 2011, EPA provided planning targets for nitrogen, phosphorus and sediment for the Phase II Watershed Implementation Plans (WIPs) of the Chesapeake Bay TMDL. This page provides the letters containing those planning targets.
Sears Point Tidal Marsh Restoration Project: Phase II
Information about the SFBWQP Sears Point Tidal Marsh Restoration Project: Phase II, part of an EPA competitive grant program to improve SF Bay water quality focused on restoring impaired waters and enhancing aquatic resources.
Three-dimensional spatiotemporal features for fast content-based retrieval of focal liver lesions.
Roy, Sharmili; Chi, Yanling; Liu, Jimin; Venkatesh, Sudhakar K; Brown, Michael S
2014-11-01
Content-based image retrieval systems for 3-D medical datasets still largely rely on 2-D image-based features extracted from a few representative slices of the image stack. Most 2 -D features that are currently used in the literature not only model a 3-D tumor incompletely but are also highly expensive in terms of computation time, especially for high-resolution datasets. Radiologist-specified semantic labels are sometimes used along with image-based 2-D features to improve the retrieval performance. Since radiological labels show large interuser variability, are often unstructured, and require user interaction, their use as lesion characterizing features is highly subjective, tedious, and slow. In this paper, we propose a 3-D image-based spatiotemporal feature extraction framework for fast content-based retrieval of focal liver lesions. All the features are computer generated and are extracted from four-phase abdominal CT images. Retrieval performance and query processing times for the proposed framework is evaluated on a database of 44 hepatic lesions comprising of five pathological types. Bull's eye percentage score above 85% is achieved for three out of the five lesion pathologies and for 98% of query lesions, at least one same type of lesion is ranked among the top two retrieved results. Experiments show that the proposed system's query processing is more than 20 times faster than other already published systems that use 2-D features. With fast computation time and high retrieval accuracy, the proposed system has the potential to be used as an assistant to radiologists for routine hepatic tumor diagnosis.
Autocorrelation and Regularization of Query-Based Information Retrieval Scores
2008-02-01
of the most general information retrieval models [ Salton , 1968]. By treating a query as a very short document, documents and queries can be rep... Salton , 1971]. In the context of single link hierarchical clustering, Jardine and van Rijsbergen showed that ranking all k clusters and retrieving a...a document about “dogs”, then the system will always miss this document when a user queries “dog”. Salton recognized that a document’s representation
Query Log Analysis of an Electronic Health Record Search Engine
Yang, Lei; Mei, Qiaozhu; Zheng, Kai; Hanauer, David A.
2011-01-01
We analyzed a longitudinal collection of query logs of a full-text search engine designed to facilitate information retrieval in electronic health records (EHR). The collection, 202,905 queries and 35,928 user sessions recorded over a course of 4 years, represents the information-seeking behavior of 533 medical professionals, including frontline practitioners, coding personnel, patient safety officers, and biomedical researchers for patient data stored in EHR systems. In this paper, we present descriptive statistics of the queries, a categorization of information needs manifested through the queries, as well as temporal patterns of the users’ information-seeking behavior. The results suggest that information needs in medical domain are substantially more sophisticated than those that general-purpose web search engines need to accommodate. Therefore, we envision there exists a significant challenge, along with significant opportunities, to provide intelligent query recommendations to facilitate information retrieval in EHR. PMID:22195150
Efficient hemodynamic event detection utilizing relational databases and wavelet analysis
NASA Technical Reports Server (NTRS)
Saeed, M.; Mark, R. G.
2001-01-01
Development of a temporal query framework for time-oriented medical databases has hitherto been a challenging problem. We describe a novel method for the detection of hemodynamic events in multiparameter trends utilizing wavelet coefficients in a MySQL relational database. Storage of the wavelet coefficients allowed for a compact representation of the trends, and provided robust descriptors for the dynamics of the parameter time series. A data model was developed to allow for simplified queries along several dimensions and time scales. Of particular importance, the data model and wavelet framework allowed for queries to be processed with minimal table-join operations. A web-based search engine was developed to allow for user-defined queries. Typical queries required between 0.01 and 0.02 seconds, with at least two orders of magnitude improvement in speed over conventional queries. This powerful and innovative structure will facilitate research on large-scale time-oriented medical databases.
A Fuzzy Query Mechanism for Human Resource Websites
NASA Astrophysics Data System (ADS)
Lai, Lien-Fu; Wu, Chao-Chin; Huang, Liang-Tsung; Kuo, Jung-Chih
Users' preferences often contain imprecision and uncertainty that are difficult for traditional human resource websites to deal with. In this paper, we apply the fuzzy logic theory to develop a fuzzy query mechanism for human resource websites. First, a storing mechanism is proposed to store fuzzy data into conventional database management systems without modifying DBMS models. Second, a fuzzy query language is proposed for users to make fuzzy queries on fuzzy databases. User's fuzzy requirement can be expressed by a fuzzy query which consists of a set of fuzzy conditions. Third, each fuzzy condition associates with a fuzzy importance to differentiate between fuzzy conditions according to their degrees of importance. Fourth, the fuzzy weighted average is utilized to aggregate all fuzzy conditions based on their degrees of importance and degrees of matching. Through the mutual compensation of all fuzzy conditions, the ordering of query results can be obtained according to user's preference.
Sampri, Alexia; Sypsa, Karla; Tsagarakis, Konstantinos P
2018-01-01
Background With the internet’s penetration and use constantly expanding, this vast amount of information can be employed in order to better assess issues in the US health care system. Google Trends, a popular tool in big data analytics, has been widely used in the past to examine interest in various medical and health-related topics and has shown great potential in forecastings, predictions, and nowcastings. As empirical relationships between online queries and human behavior have been shown to exist, a new opportunity to explore the behavior toward asthma—a common respiratory disease—is present. Objective This study aimed at forecasting the online behavior toward asthma and examined the correlations between queries and reported cases in order to explore the possibility of nowcasting asthma prevalence in the United States using online search traffic data. Methods Applying Holt-Winters exponential smoothing to Google Trends time series from 2004 to 2015 for the term “asthma,” forecasts for online queries at state and national levels are estimated from 2016 to 2020 and validated against available Google query data from January 2016 to June 2017. Correlations among yearly Google queries and between Google queries and reported asthma cases are examined. Results Our analysis shows that search queries exhibit seasonality within each year and the relationships between each 2 years’ queries are statistically significant (P<.05). Estimated forecasting models for a 5-year period (2016 through 2020) for Google queries are robust and validated against available data from January 2016 to June 2017. Significant correlations were found between (1) online queries and National Health Interview Survey lifetime asthma (r=–.82, P=.001) and current asthma (r=–.77, P=.004) rates from 2004 to 2015 and (2) between online queries and Behavioral Risk Factor Surveillance System lifetime (r=–.78, P=.003) and current asthma (r=–.79, P=.002) rates from 2004 to 2014. The correlations are negative, but lag analysis to identify the period of response cannot be employed until short-interval data on asthma prevalence are made available. Conclusions Online behavior toward asthma can be accurately predicted, and significant correlations between online queries and reported cases exist. This method of forecasting Google queries can be used by health care officials to nowcast asthma prevalence by city, state, or nationally, subject to future availability of daily, weekly, or monthly data on reported cases. This method could therefore be used for improved monitoring and assessment of the needs surrounding the current population of patients with asthma. PMID:29530839
Vigorito, Fabio de Abreu; Dominguez, Gladys Cristina; Aidar, Luís Antônio de Arruda
2014-01-01
Objective To assess the dentoskeletal changes observed in treatment of Class II, division 1 malocclusion patients with mandibular retrognathism. Treatment was performed with the Herbst orthopedic appliance during 13 months (phase I) and pre-adjusted orthodontic fixed appliance (phase II). Methods Lateral cephalograms of 17 adolescents were taken in phase I onset (T1) and completion (T2); in the first thirteen months of phase II (T3) and in phase II completion (T4). Differences among the cephalometric variables were statistically analyzed (Bonferroni variance and multiple comparisons). Results From T1 to T4, 42% of overall maxillary growth was observed between T1 and T2 (P < 0.01), 40.3% between T2 and T3 (P < 0.05) and 17.7% between T3 and T4 (n.s.). As for overall mandibular movement, 48.2% was observed between T1 and T2 (P < 0.001) and 51.8% between T2 and T4 (P < 0.01) of which 15.1% was observed between T2 and T3 (n.s.) and 36.7% between T3 and T4 (P < 0.01). Class II molar relationship and overjet were properly corrected. The occlusal plane which rotated clockwise between T1 and T2, returned to its initial position between T2 and T3 remaining stable until T4. The mandibular plane inclination did not change at any time during treatment. Conclusion Mandibular growth was significantly greater in comparison to maxillary, allowing sagittal maxillomandibular adjustment. The dentoalveolar changes (upper molar) that overcorrected the malocclusion in phase I, partially recurred in phase II, but did not hinder correction of the malocclusion. Facial type was preserved. PMID:24713559
Tian, Hua; Hu, Zheng; He, Qun; Liu, Xueliang; Zhang, Li; Chang, Xijun
2012-07-01
Two solid-phase adsorbents (phase I and phase II) were synthesized successfully that o-Anisic acid derivatives were evenly functionalized on the surface of activated carbon. It was certified that the two adsorbents were applied to preconcentrate and separate trace levels of Pb(II) and Fe(III) from natural liquid samples with satisfactory results. It can be found that the adsorption capacity of the ions adsorbed on phase I and phase II was 48.3 and 85.7 mg g(-1) for Pb(II), 39.5 and 72.5 mg g(-1) for Fe(III), respectively. The detection limit (3σ) of the method separated on phase I and phase II was 0.12 and 0.09 ng mL(-1) for Pb(II), 0.23 and 0.17 ng mL(-1) for Fe(III), respectively. The relative standard deviation (R.S.D.) of the method was lower than 3.0%. The adsorption and desorption property of two kinds of adsorbents was comparatively studied, respectively. The adsorption selectivity of heavy metal ions at certain pH, the adsorption kinetics, the condition of complete elution, the effect of coexisting ions, the adsorption capacity and adsorption isotherm modes were examined. Based on the experimental datum determined by inductively coupled plasma optical emission spectrometry (ICP-OES), it was certified that the adsorption on the surface of adsorbents was in strict accordance with the monolayer adsorption principle. The structural features of series of multidentate ligand modified on adsorption matrix had been obtained. These conclusions can provide reference for synthesizing an efficient adsorbent which is specific to remove a particular kind of contaminant. Copyright © 2012 Elsevier B.V. All rights reserved.
A review of EO image information mining
NASA Astrophysics Data System (ADS)
Quartulli, Marco; Olaizola, Igor G.
2013-01-01
We analyze the state of the art of content-based retrieval in Earth observation image archives focusing on complete systems showing promise for operational implementation. The different paradigms at the basis of the main system families are introduced. The approaches taken are considered, focusing in particular on the phases after primitive feature extraction. The solutions envisaged for the issues related to feature simplification and synthesis, indexing, semantic labeling are reviewed. The methodologies for query specification and execution are evaluated. Conclusions are drawn on the state of published research in Earth observation (EO) mining.
Menopause and big data: Word Adjacency Graph modeling of menopause-related ChaCha data.
Carpenter, Janet S; Groves, Doyle; Chen, Chen X; Otte, Julie L; Miller, Wendy R
2017-07-01
To detect and visualize salient queries about menopause using Big Data from ChaCha. We used Word Adjacency Graph (WAG) modeling to detect clusters and visualize the range of menopause-related topics and their mutual proximity. The subset of relevant queries was fully modeled. We split each query into token words (ie, meaningful words and phrases) and removed stopwords (ie, not meaningful functional words). The remaining words were considered in sequence to build summary tables of words and two and three-word phrases. Phrases occurring at least 10 times were used to build a network graph model that was iteratively refined by observing and removing clusters of unrelated content. We identified two menopause-related subsets of queries by searching for questions containing menopause and menopause-related terms (eg, climacteric, hot flashes, night sweats, hormone replacement). The first contained 263,363 queries from individuals aged 13 and older and the second contained 5,892 queries from women aged 40 to 62 years. In the first set, we identified 12 topic clusters: 6 relevant to menopause and 6 less relevant. In the second set, we identified 15 topic clusters: 11 relevant to menopause and 4 less relevant. Queries about hormones were pervasive within both WAG models. Many of the queries reflected low literacy levels and/or feelings of embarrassment. We modeled menopause-related queries posed by ChaCha users between 2009 and 2012. ChaCha data may be used on its own or in combination with other Big Data sources to identify patient-driven educational needs and create patient-centered interventions.
Fast Inbound Top-K Query for Random Walk with Restart.
Zhang, Chao; Jiang, Shan; Chen, Yucheng; Sun, Yidan; Han, Jiawei
2015-09-01
Random walk with restart (RWR) is widely recognized as one of the most important node proximity measures for graphs, as it captures the holistic graph structure and is robust to noise in the graph. In this paper, we study a novel query based on the RWR measure, called the inbound top-k (Ink) query. Given a query node q and a number k , the Ink query aims at retrieving k nodes in the graph that have the largest weighted RWR scores to q . Ink queries can be highly useful for various applications such as traffic scheduling, disease treatment, and targeted advertising. Nevertheless, none of the existing RWR computation techniques can accurately and efficiently process the Ink query in large graphs. We propose two algorithms, namely Squeeze and Ripple, both of which can accurately answer the Ink query in a fast and incremental manner. To identify the top- k nodes, Squeeze iteratively performs matrix-vector multiplication and estimates the lower and upper bounds for all the nodes in the graph. Ripple employs a more aggressive strategy by only estimating the RWR scores for the nodes falling in the vicinity of q , the nodes outside the vicinity do not need to be evaluated because their RWR scores are propagated from the boundary of the vicinity and thus upper bounded. Ripple incrementally expands the vicinity until the top- k result set can be obtained. Our extensive experiments on real-life graph data sets show that Ink queries can retrieve interesting results, and the proposed algorithms are orders of magnitude faster than state-of-the-art method.
Sleep-wake time perception varies by direct or indirect query.
Alameddine, Y; Ellenbogen, J M; Bianchi, M T
2015-01-15
The diagnosis of insomnia rests on self-report of difficulty initiating or maintaining sleep. However, subjective reports may be unreliable, and possibly may vary by the method of inquiry. We investigated this possibility by comparing within-individual response to direct versus indirect time queries after overnight polysomnography. We obtained self-reported sleep-wake times via morning questionnaires in 879 consecutive adult diagnostic polysomnograms. Responses were compared within subjects (direct versus indirect query) and across groups defined by apnea-hypopnea index and by self-reported insomnia symptoms in pre-sleep questionnaires. Direct queries required a time duration response, while indirect queries required clock times from which we calculated time durations. Direct and indirect queries of sleep latency were the same in only 41% of cases, and total sleep time queries matched in only 5.4%. For both latency and total sleep, the most common discrepancy involved the indirect value being larger than the direct response. The discrepancy between direct and indirect queries was not related to objective sleep metrics. The degree of discrepancy was not related to the presence of insomnia symptoms, although patients reporting insomnia symptoms showed underestimation of total sleep duration by direct response. Self-reported sleep latency and total sleep time are often internally inconsistent when comparing direct and indirect survey queries of each measure. These discrepancies represent substantive challenges to effective clinical practice, particularly when diagnosis and management depends on self-reported sleep patterns, as with insomnia. Although self-reported sleep-wake times remains fundamental to clinical practice, objective measures provide clinically relevant adjunctive information. © 2015 American Academy of Sleep Medicine.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Pugliese, S.M.
1977-02-01
In Phase I of the Research Safety Vehicle Program (RSV), preliminary design and performance specifications were developed for a mid-1980's vehicle that integrates crashworthiness and occupant safety features with material resource conservation, economy, and producibility. Phase II of the program focused on development of the total vehicle design via systems engineering and integration analyses. As part of this effort, it was necessary to continuously review the Phase I recommended performance specification in relation to ongoing design/test activities. This document contains the results of analyses of the Phase I specifications. The RSV is expected to satisfy all of the producibility andmore » safety related specifications, i.e., handling and stability systems, crashworthiness, occupant protection, pedestrian/cyclist protection, etc.« less
Lee, Cindy; Vather, Ryash; O'Callaghan, Anne; Robinson, Jackie; McLeod, Briar; Findlay, Michael; Bissett, Ian
2013-12-01
Malignant bowel obstruction (MBO) is common in patients with advanced cancer. To perform a phase II study to assess the feasibility of conducting a phase III trial investigating the therapeutic value of gastrografin in MBO. Randomized double-blinded placebo-controlled feasibility study. Participants received 100 mL of either gastrografin or placebo. Over 8 months, 57 patients were screened and 9 enrolled (15.8% recruitment rate). Of the 9 enrolled, 4 received gastrografin (with 2 completing assessment) and 5 received placebo (with 4 completing assessment). It is not feasible to conduct a phase III trial using the same study protocol. This study validates the use of the phase II feasibility study to assess protocol viability in a palliative population prior to embarking on a larger trial.
Chronic Iron Limitation Confers Transient Resistance to Oxidative Stress in Marine Diatoms.
Graff van Creveld, Shiri; Rosenwasser, Shilo; Levin, Yishai; Vardi, Assaf
2016-10-01
Diatoms are single-celled, photosynthetic, bloom-forming algae that are responsible for at least 20% of global primary production. Nevertheless, more than 30% of the oceans are considered "ocean deserts" due to iron limitation. We used the diatom Phaeodactylum tricornutum as a model system to explore diatom's response to iron limitation and its interplay with susceptibility to oxidative stress. By analyzing physiological parameters and proteome profiling, we defined two distinct phases: short-term (<3 d, phase I) and chronic (>5 d, phase II) iron limitation. While at phase I no significant changes in physiological parameters were observed, molecular markers for iron starvation, such as Iron Starvation Induced Protein and flavodoxin, were highly up-regulated. At phase II, down-regulation of numerous iron-containing proteins was detected in parallel to reduction in growth rate, chlorophyll content, photosynthetic activity, respiration rate, and antioxidant capacity. Intriguingly, while application of oxidative stress to phase I and II iron-limited cells similarly oxidized the reduced glutathione (GSH) pool, phase II iron limitation exhibited transient resistance to oxidative stress, despite the down regulation of many antioxidant proteins. By comparing proteomic profiles of P. tricornutum under iron limitation and metatranscriptomic data of an iron enrichment experiment conducted in the Pacific Ocean, we propose that iron-limited cells in the natural environment resemble the phase II metabolic state. These results provide insights into the trade-off between optimal growth rate and susceptibility to oxidative stress in the response of diatoms to iron quota in the marine environment. © 2016 American Society of Plant Biologists. All Rights Reserved.
Phillips, Bryn M; Anderson, Brian S; Hunt, John W; Clark, Sara L; Voorhees, Jennifer P; Tjeerdema, Ron S; Casteline, Jane; Stewart, Margaret
2009-02-01
Phase I whole sediment toxicity identification evaluation (TIE) methods have been developed to characterize the cause of toxicity as organic chemicals, metals, or ammonia. In Phase II identification treatments, resins added to whole sediment to reduce toxicity caused by metals and organics can be separated and eluted much like solid-phase extraction (SPE) columns are eluted for interstitial water. In this study, formulated reference sediments spiked with toxic concentrations of copper, fluoranthene, and nonylphenol were subjected to whole sediment and interstitial water TIE treatments to evaluate Phase I and II TIE procedures for identifying the cause of toxicity to Hyalella azteca. Phase I TIE treatments consisted of adding adsorbent resins to whole sediment, and using SPE columns to remove spiked chemicals from interstitial water. Phase II treatments consisted of eluting resins and SPE columns and the preparation and testing of eluates for toxicity and chemistry. Whole sediment resins and SPE columns significantly reduced toxicity, and the eluates from all treatments contained toxic concentrations of the spiked chemical except for interstitial water fluoranthene. Toxic unit analysis based on median lethal concentrations (LC50s) allowed for the comparison of chemical concentrations among treatments, and demonstrated that the bioavailability of some chemicals was reduced in some samples and treatments. The concentration of fluoranthene in the resin eluate closely approximated the original interstitial water concentration, but the resin eluate concentrations of copper and nonylphenol were much higher than the original interstitial water concentrations. Phase II whole sediment TIE treatments provided complementary lines of evidence to the interstitial water TIE results.
Chronic Iron Limitation Confers Transient Resistance to Oxidative Stress in Marine Diatoms1
Graff van Creveld, Shiri; Rosenwasser, Shilo; Vardi, Assaf
2016-01-01
Diatoms are single-celled, photosynthetic, bloom-forming algae that are responsible for at least 20% of global primary production. Nevertheless, more than 30% of the oceans are considered “ocean deserts” due to iron limitation. We used the diatom Phaeodactylum tricornutum as a model system to explore diatom’s response to iron limitation and its interplay with susceptibility to oxidative stress. By analyzing physiological parameters and proteome profiling, we defined two distinct phases: short-term (<3 d, phase I) and chronic (>5 d, phase II) iron limitation. While at phase I no significant changes in physiological parameters were observed, molecular markers for iron starvation, such as Iron Starvation Induced Protein and flavodoxin, were highly up-regulated. At phase II, down-regulation of numerous iron-containing proteins was detected in parallel to reduction in growth rate, chlorophyll content, photosynthetic activity, respiration rate, and antioxidant capacity. Intriguingly, while application of oxidative stress to phase I and II iron-limited cells similarly oxidized the reduced glutathione (GSH) pool, phase II iron limitation exhibited transient resistance to oxidative stress, despite the down regulation of many antioxidant proteins. By comparing proteomic profiles of P. tricornutum under iron limitation and metatranscriptomic data of an iron enrichment experiment conducted in the Pacific Ocean, we propose that iron-limited cells in the natural environment resemble the phase II metabolic state. These results provide insights into the trade-off between optimal growth rate and susceptibility to oxidative stress in the response of diatoms to iron quota in the marine environment. PMID:27503604
SH-2F LAMPS Instructional Systems Development: Phase II. Final Report.
ERIC Educational Resources Information Center
Gibbons, Andrew S.; Hymes, Jonah P.
This project was one of four aircrew training development projects in a continuing study of the methodology, effectiveness, and resource requirements of the Instructional Systems Development (ISD) process. This report covers the Phase II activities of a two-phase project for the development of aircrew training for SH-2F anti-submarine warfare…
40 CFR 63.163 - Standards: Pumps in light liquid service.
Code of Federal Regulations, 2013 CFR
2013-07-01
... later than 1 year after the compliance date; and (C) Phase III, beginning no later than 21/2 years after... requirements; and (B) Beginning no later than 1 year after initial start-up, comply with the Phase III... parts per million or greater. (ii) For Phase II, an instrument reading of 5,000 parts per million or...
40 CFR 63.163 - Standards: Pumps in light liquid service.
Code of Federal Regulations, 2012 CFR
2012-07-01
... later than 1 year after the compliance date; and (C) Phase III, beginning no later than 21/2 years after... requirements; and (B) Beginning no later than 1 year after initial start-up, comply with the Phase III... parts per million or greater. (ii) For Phase II, an instrument reading of 5,000 parts per million or...
40 CFR 63.163 - Standards: Pumps in light liquid service.
Code of Federal Regulations, 2014 CFR
2014-07-01
... later than 1 year after the compliance date; and (C) Phase III, beginning no later than 21/2 years after... requirements; and (B) Beginning no later than 1 year after initial start-up, comply with the Phase III... parts per million or greater. (ii) For Phase II, an instrument reading of 5,000 parts per million or...
An Experimental Evaluation of Hyperactivity and Food Additives. 1977-Phase II.
ERIC Educational Resources Information Center
Harley, J. Preston; And Others
Phase II of a study on the effectiveness of B. Feingold's recommended diet for hyperactive children involved the nine children (mean age 9 years) who had shown the "best" response to diet manipulation in Phase I. Each child served as his own control and was challenged with specified amounts of placebo and artificial color containing food…
ERIC Educational Resources Information Center
Schulz, Russel E.; Farrell, Jean R.
This resource guide for the use of job aids ("how-to-do-it" guidance) for activities identified in the second phase of the Instructional Systems Development Model (ISD) contains an introduction to the use of job aids, as well as descriptive authoring flowcharts for Blocks II.1 through II.4. The introduction includes definitions;…
NASA Astrophysics Data System (ADS)
Panicker, Lata
2018-05-01
Polycrystalline samples of 4-hydroxybenzaldehyde (4-HOBAL) were investigated using differential scanning calorimeter (DSC), Raman spectroscopy and X-ray powder diffraction. The DSC data indicated that 4-HOBAL on heating undergoes a polymorphic transformation from polymorph I to polymorph II. The polymorph II formed remains metastable at ambient condition and transforms to polymorph I when annealed at ambient temperature for more than seven days. The structural information of polymorphs I and II obtained using its X-ray powder diffraction patterns indicated that 4-HOBAL undergoes an isostructural phase transition from polymorph I (monoclinic, P21/c) to polymorph II (monoclinic, P21/c). Raman data suggest that this structural change is associated with some change in its molecular interactions. Thus, in 4-HOBAL the polymorphic phase transformation (II to I) even though energetically favoured is kinetically hindered.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Herchenhorn, Daniel, E-mail: herchenhorn@hotmail.co; Dias, Fernando L.; Viegas, Celia M.P.
Purpose: Erlotinib, an oral tyrosine kinase inhibitor, is active against head-and-neck squamous cell carcinoma (HNSCC) and possibly has a synergistic interaction with chemotherapy and radiotherapy. We investigated the safety and efficacy of erlotinib added to cisplatin and radiotherapy in locally advanced HNSCC. Methods and Materials: In this Phase I/II trial 100 mg/m{sup 2} of cisplatin was administered on Days 8, 29, and 50, and radiotherapy at 70 Gy was started on Day 8. During Phase I, the erlotinib dose was escalated (50 mg, 100 mg, and 150 mg) in consecutive cohorts of 3 patients, starting on Day 1 and continuingmore » during radiotherapy. Dose-limiting toxicity was defined as any Grade 4 event requiring radiotherapy interruptions. Phase II was initiated 8 weeks after the last Phase I enrollment. Results: The study accrued 9 patients in Phase I and 28 in Phase II; all were evaluable for efficacy and safety. No dose-limiting toxicity occurred in Phase I, and the recommended Phase II dose was 150 mg. The most frequent nonhematologic toxicities were nausea/vomiting, dysphagia, stomatitis, xerostomia and in-field dermatitis, acneiform rash, and diarrhea. Of the 31 patients receiving a 150-mg daily dose of erlotinib, 23 (74%; 95% confidence interval, 56.8%-86.3%) had a complete response, 3 were disease free after salvage surgery, 4 had inoperable residual disease, and 1 died of sepsis during treatment. With a median 37 months' follow-up, the 3-year progression-free and overall survival rates were 61% and 72%, respectively. Conclusions: This combination appears safe, has encouraging activity, and deserves further studies in locally advanced HNSCC.« less
Bagheri, Hasan; Afkhami, Abbas; Saber-Tehrani, Mohammad; Khoshsafar, Hosein
2012-08-15
A versatile and robust solid phase with both magnetic property and a very high adsorption capacity is presented on the basis of modification of iron oxide-silica magnetic particles with a newly synthesized Schiff base (Fe(3)O(4)/SiO(2)/L). The structure of the resulting product was confirmed by Fourier transform infrared (FT-IR) spectra, X-ray diffraction (XRD) spectrometry and transmission electron microscopy (TEM). We developed an efficient and cost-effective method for the preconcentration of trace amounts of Pb(II), Cd(II) and Cu(II) in environmental and biological samples using this novel magnetic solid phase. Prepared magnetic solid phase is an ideal support because it has a large surface area, good selectivity and can be easily retrieved from large volumes of aqueous solutions. The possible parameters affecting the enrichment were optimized. Under the optimal conditions, the method detection limit was 0.14, 0.19 and 0.12 μg L(-1) for Pb(II), Cd(II) and Cu(II) ions, respectively. The established method has been successfully applied to analyze real samples, and satisfactory results were obtained. All these indicated that this magnetic phase had a great potential in environmental and biological fields. Copyright © 2012 Elsevier B.V. All rights reserved.
Lee, Keun-Wook; Lee, Kyung Hee; Zang, Dae Young; Park, Young Iee; Shin, Dong Bok; Kim, Jin Won; Im, Seock-Ah; Koh, Sung Ae; Yu, Kyung-Sang; Cho, Joo-Youn; Jung, Jin-A; Bang, Yung-Jue
2015-08-01
Oraxol consists of paclitaxel and HM30181A, a P-glycoprotein inhibitor, to increase the oral bioavailability of paclitaxel. This phase I/II study (HM-OXL-201) was conducted to determine the maximum tolerated dose (MTD) and recommended phase II dose (RP2D) of Oraxol. In addition, we investigated the efficacy and safety of Oraxol as second-line chemotherapy for metastatic or recurrent gastric cancer (GC). In the phase I component, paclitaxel was orally administered at escalating doses (90, 120, or 150 mg/m(2) per day) with a fixed dose (15 mg/day) of HM30181A. Oraxol was administrated 6 times per cycle (days 1, 2, 8, 9, 15, and 16) every 4 weeks. In the phase II component, the efficacy and safety of Oraxol were evaluated. In the phase I component, the MTD could not be determined. Based on toxicity and pharmacokinetic data, the RP2D of oral paclitaxel was determined to be 150 mg/m(2). In the phase II component, 4 of 43 patients (9.3%) achieved partial responses. Median progression-free survival and overall survival were 2.6 and 10.7 months, respectively. Toxicity profiles were favorable, and the most common drug-related adverse events (grade ≥3) were neutropenia and diarrhea. Oraxol exhibited modest efficacy and favorable toxicity profiles as second-line chemotherapy for GC. ©AlphaMed Press; the data published online to support this summary is the property of the authors.
Tchou, Isabelle; Margeli, Alexandra; Tsironi, Maria; Skenderi, Katerina; Barnet, Marc; Kanaka-Gantenbein, Christina; Papassotiriou, Ioannis; Beris, Photis
2009-09-01
We investigated the actions of growth-differentiation factor (GDF)-15, endoglin and N-terminal pro-brain natriuretic peptide (NT-pro-BNP) in 15 male athletes who participated in the ultradistance foot race of the 246 km 'Sparthathlon'. Measurements were performed before (phase I), at the end of the race (phase II) and 48 h post-race (phase III). GDF-15 and endoglin serum concentrations were determined with enzyme-linked immunosorbent assay and NT-pro-BNP plasma levels by electrochemiluminescence. GDF-15 levels were increased from phase I (563.9 +/- 57.1 pg ml(-1)) to phase II (2311.1 +/- 462.3 pg ml(-1)) and decreased at phase III (862.0 +/- 158.0 pg ml(-1)) (p < 0.0002). NT-pro-BNP levels followed a similar pattern to that of GDF-15 from 38.1 +/- 4.8 pg ml(-1) at phase I to 1280.6 +/- 259.0 pg ml(-1) at phase II and 89.8 +/- 13.6 pg ml(-1) at phase III (p < 0.0001) and at the same time points, endoglin levels were 4.7 +/- 0.2 ng ml(-1) at phase I, 5.8 +/- 0.2 ng ml(-1) at phase II and 4.3 +/- 0.2 ng ml(-1) at phase III (p < 0.002). These findings indicate that circulating GDF-15, endoglin and NT-pro-BNP levels reflect a transient endothelial dysfunction in these athletes who participated in a foot race consisting of continuous, prolonged and brisk exercise.
Secure and Efficient k-NN Queries⋆
Asif, Hafiz; Vaidya, Jaideep; Shafiq, Basit; Adam, Nabil
2017-01-01
Given the morass of available data, ranking and best match queries are often used to find records of interest. As such, k-NN queries, which give the k closest matches to a query point, are of particular interest, and have many applications. We study this problem in the context of the financial sector, wherein an investment portfolio database is queried for matching portfolios. Given the sensitivity of the information involved, our key contribution is to develop a secure k-NN computation protocol that can enable the computation k-NN queries in a distributed multi-party environment while taking domain semantics into account. The experimental results show that the proposed protocols are extremely efficient. PMID:29218333
Nearest private query based on quantum oblivious key distribution
NASA Astrophysics Data System (ADS)
Xu, Min; Shi, Run-hua; Luo, Zhen-yu; Peng, Zhen-wan
2017-12-01
Nearest private query is a special private query which involves two parties, a user and a data owner, where the user has a private input (e.g., an integer) and the data owner has a private data set, and the user wants to query which element in the owner's private data set is the nearest to his input without revealing their respective private information. In this paper, we first present a quantum protocol for nearest private query, which is based on quantum oblivious key distribution (QOKD). Compared to the classical related protocols, our protocol has the advantages of the higher security and the better feasibility, so it has a better prospect of applications.
Cognitive issues in searching images with visual queries
NASA Astrophysics Data System (ADS)
Yu, ByungGu; Evens, Martha W.
1999-01-01
In this paper, we propose our image indexing technique and visual query processing technique. Our mental images are different from the actual retinal images and many things, such as personal interests, personal experiences, perceptual context, the characteristics of spatial objects, and so on, affect our spatial perception. These private differences are propagated into our mental images and so our visual queries become different from the real images that we want to find. This is a hard problem and few people have tried to work on it. In this paper, we survey the human mental imagery system, the human spatial perception, and discuss several kinds of visual queries. Also, we propose our own approach to visual query interpretation and processing.
South Bay Salt Pond Restoration, Phase II at Ravenswood
Information about the South Bay Salt Pond Restoration Project: Phase II Construction at Ravenswood, part of an EPA competitive grant program to improve SF Bay water quality focused on restoring impaired waters and enhancing aquatic resources.
Carbon footprint estimator, phase II : volume II - technical appendices.
DOT National Transportation Integrated Search
2014-03-01
The GASCAP model was developed to provide a software tool for analysis of the life-cycle GHG : emissions associated with the construction and maintenance of transportation projects. This phase : of development included techniques for estimating emiss...
Morristown Alternative Transportation Study Phase II.
DOT National Transportation Integrated Search
2005-10-14
This report summarizes the Phase II planning effort conducted by the park and the US Department of Transportation's Volpe Center (the Volpe Center) to articulate a viable park-community pilot transit service for Morristown National Historical Park. M...
Céolin, R; Rietveld, I B
2015-01-01
A topological pressure-temperature phase diagram involving the phase relationships of ritonavir forms I and II has been constructed using experimental calorimetric and volumetric data available from the literature. The triple point I-II-liquid is located at a temperature of about 407 K and a pressure as extraordinarily small as 17.5 MPa (175 bar). Thus, the less soluble solid phase (form II) will become metastable on increasing pressure. At room temperature, form I becomes stable around 100 MPa indicating that form II may turn into form I at a relatively low pressure of 1000 bar, which may occur under processing conditions such as mixing or grinding. This case is a good example for which a proper thermodynamic evaluation trumps "rules of thumb" such as the density rule. Copyright © 2014 Elsevier Masson SAS. All rights reserved.
Detailed validation of the bidirectional effect in various Case I and Case II waters.
Gleason, Arthur C R; Voss, Kenneth J; Gordon, Howard R; Twardowski, Michael; Sullivan, James; Trees, Charles; Weidemann, Alan; Berthon, Jean-François; Clark, Dennis; Lee, Zhong-Ping
2012-03-26
Simulated bidirectional reflectance distribution functions (BRDF) were compared with measurements made just beneath the water's surface. In Case I water, the set of simulations that varied the particle scattering phase function depending on chlorophyll concentration agreed more closely with the data than other models. In Case II water, however, the simulations using fixed phase functions agreed well with the data and were nearly indistinguishable from each other, on average. The results suggest that BRDF corrections in Case II water are feasible using single, average, particle scattering phase functions, but that the existing approach using variable particle scattering phase functions is still warranted in Case I water.
Mechanochemical induced structural changes in sucrose using the rotational diamond anvil cell
NASA Astrophysics Data System (ADS)
Ciezak-Jenkins, Jennifer A.; Jenkins, Timothy A.
2018-02-01
The response of sucrose to high-pressure and shear conditions has been studied in a rotational diamond anvil cell. Previous experiments conducted by Bridgman and Teller showed divergent behavior in regard to the existence of a rheological explosion under mechanochemical stimuli. Raman spectroscopy confirmed the existence of the isostructural Phase I to Phase II transition near 5 GPa. When subjected to high-pressure and shear, Raman spectra of Phase I showed evidence that while the sucrose molecule underwent significant molecular deformation, there was no evidence of a complete chemical reaction. In contrast, Phase II showed a near-total loss of the in-situ Raman signal in response to shear, suggesting the onset of amorphization or decomposition. The divergent behaviors of Phase I and Phase II are examined in light of the differences in the hydrogen bonding and plasticity of the material.
Blind Seer: A Scalable Private DBMS
2014-05-01
searchable index terms per DB row, in time comparable to (insecure) MySQL (many practical queries can be privately executed with work 1.2-3 times slower...than MySQL , although some queries are costlier). We support a rich query set, including searching on arbitrary boolean formulas on keywords and ranges...index terms per DB row, in time comparable to (insecure) MySQL (many practical queries can be privately executed with work 1.2-3 times slower than MySQL
Applying Wave (registered trademark) to Build an Air Force Community of Interest Shared Space
2007-08-01
Performance. It is essential that an inverse transform be defined for every transform, or else the query mediator must be smart enough to figure out how...to invert it. Without an inverse transform , if an incoming query constrains on the transformed attribute, the query mediator might generate a query...plan that is horribly inefficient. If you must code a custom transformation function, you must also code the inverse transform . Putting the
Concept Based Tie-breaking and Maximal Marginal Relevance Retrieval in Microblog Retrieval
2014-11-01
the same score, another singal will be used to rank these documents to break the ties , but the relative orders of other documents against these...documents remain the same. The tie- breaking step above is repeatedly applied to further break ties until all candidate signals are applied and the ranking...searched it on the Yahoo! search engine, which returned some query sug- gestions for the query. The original queries as well as their query suggestions
Multi-field query expansion is effective for biomedical dataset retrieval.
Bouadjenek, Mohamed Reda; Verspoor, Karin
2017-01-01
In the context of the bioCADDIE challenge addressing information retrieval of biomedical datasets, we propose a method for retrieval of biomedical data sets with heterogenous schemas through query reformulation. In particular, the method proposed transforms the initial query into a multi-field query that is then enriched with terms that are likely to occur in the relevant datasets. We compare and evaluate two query expansion strategies, one based on the Rocchio method and another based on a biomedical lexicon. We then perform a comprehensive comparative evaluation of our method on the bioCADDIE dataset collection for biomedical retrieval. We demonstrate the effectiveness of our multi-field query method compared to two baselines, with MAP improved from 0.2171 and 0.2669 to 0.2996. We also show the benefits of query expansion, where the Rocchio expanstion method improves the MAP for our two baselines from 0.2171 and 0.2669 to 0.335. We show that the Rocchio query expansion method slightly outperforms the one based on the biomedical lexicon as a source of terms, with an improvement of roughly 3% for MAP. However, the query expansion method based on the biomedical lexicon is much less resource intensive since it does not require computation of any relevance feedback set or any initial execution of the query. Hence, in term of trade-off between efficiency, execution time and retrieval accuracy, we argue that the query expansion method based on the biomedical lexicon offers the best performance for a prototype biomedical data search engine intended to be used at a large scale. In the official bioCADDIE challenge results, although our approach is ranked seventh in terms of the infNDCG evaluation metric, it ranks second in term of P@10 and NDCG. Hence, the method proposed here provides overall good retrieval performance in relation to the approaches of other competitors. Consequently, the observations made in this paper should benefit the development of a Data Discovery Index prototype or the improvement of the existing one. © The Author(s) 2017. Published by Oxford University Press.
Multi-field query expansion is effective for biomedical dataset retrieval
2017-01-01
Abstract In the context of the bioCADDIE challenge addressing information retrieval of biomedical datasets, we propose a method for retrieval of biomedical data sets with heterogenous schemas through query reformulation. In particular, the method proposed transforms the initial query into a multi-field query that is then enriched with terms that are likely to occur in the relevant datasets. We compare and evaluate two query expansion strategies, one based on the Rocchio method and another based on a biomedical lexicon. We then perform a comprehensive comparative evaluation of our method on the bioCADDIE dataset collection for biomedical retrieval. We demonstrate the effectiveness of our multi-field query method compared to two baselines, with MAP improved from 0.2171 and 0.2669 to 0.2996. We also show the benefits of query expansion, where the Rocchio expanstion method improves the MAP for our two baselines from 0.2171 and 0.2669 to 0.335. We show that the Rocchio query expansion method slightly outperforms the one based on the biomedical lexicon as a source of terms, with an improvement of roughly 3% for MAP. However, the query expansion method based on the biomedical lexicon is much less resource intensive since it does not require computation of any relevance feedback set or any initial execution of the query. Hence, in term of trade-off between efficiency, execution time and retrieval accuracy, we argue that the query expansion method based on the biomedical lexicon offers the best performance for a prototype biomedical data search engine intended to be used at a large scale. In the official bioCADDIE challenge results, although our approach is ranked seventh in terms of the infNDCG evaluation metric, it ranks second in term of P@10 and NDCG. Hence, the method proposed here provides overall good retrieval performance in relation to the approaches of other competitors. Consequently, the observations made in this paper should benefit the development of a Data Discovery Index prototype or the improvement of the existing one. PMID:29220457
Xiao, Fuyuan; Aritsugi, Masayoshi; Wang, Qing; Zhang, Rong
2016-09-01
For efficient and sophisticated analysis of complex event patterns that appear in streams of big data from health care information systems and support for decision-making, a triaxial hierarchical model is proposed in this paper. Our triaxial hierarchical model is developed by focusing on hierarchies among nested event pattern queries with an event concept hierarchy, thereby allowing us to identify the relationships among the expressions and sub-expressions of the queries extensively. We devise a cost-based heuristic by means of the triaxial hierarchical model to find an optimised query execution plan in terms of the costs of both the operators and the communications between them. According to the triaxial hierarchical model, we can also calculate how to reuse the results of the common sub-expressions in multiple queries. By integrating the optimised query execution plan with the reuse schemes, a multi-query optimisation strategy is developed to accomplish efficient processing of multiple nested event pattern queries. We present empirical studies in which the performance of multi-query optimisation strategy was examined under various stream input rates and workloads. Specifically, the workloads of pattern queries can be used for supporting monitoring patients' conditions. On the other hand, experiments with varying input rates of streams can correspond to changes of the numbers of patients that a system should manage, whereas burst input rates can correspond to changes of rushes of patients to be taken care of. The experimental results have shown that, in Workload 1, our proposal can improve about 4 and 2 times throughput comparing with the relative works, respectively; in Workload 2, our proposal can improve about 3 and 2 times throughput comparing with the relative works, respectively; in Workload 3, our proposal can improve about 6 times throughput comparing with the relative work. The experimental results demonstrated that our proposal was able to process complex queries efficiently which can support health information systems and further decision-making. Copyright © 2016 Elsevier B.V. All rights reserved.
Huang, Liang-Chin; Ross, Karen E; Baffi, Timothy R; Drabkin, Harold; Kochut, Krzysztof J; Ruan, Zheng; D'Eustachio, Peter; McSkimming, Daniel; Arighi, Cecilia; Chen, Chuming; Natale, Darren A; Smith, Cynthia; Gaudet, Pascale; Newton, Alexandra C; Wu, Cathy; Kannan, Natarajan
2018-04-25
Many bioinformatics resources with unique perspectives on the protein landscape are currently available. However, generating new knowledge from these resources requires interoperable workflows that support cross-resource queries. In this study, we employ federated queries linking information from the Protein Kinase Ontology, iPTMnet, Protein Ontology, neXtProt, and the Mouse Genome Informatics to identify key knowledge gaps in the functional coverage of the human kinome and prioritize understudied kinases, cancer variants and post-translational modifications (PTMs) for functional studies. We identify 32 functional domains enriched in cancer variants and PTMs and generate mechanistic hypotheses on overlapping variant and PTM sites by aggregating information at the residue, protein, pathway and species level from these resources. We experimentally test the hypothesis that S768 phosphorylation in the C-helix of EGFR is inhibitory by showing that oncogenic variants altering S768 phosphorylation increase basal EGFR activity. In contrast, oncogenic variants altering conserved phosphorylation sites in the 'hydrophobic motif' of PKCβII (S660F and S660C) are loss-of-function in that they reduce kinase activity and enhance membrane translocation. Our studies provide a framework for integrative, consistent, and reproducible annotation of the cancer kinomes.
CFGP: a web-based, comparative fungal genomics platform
Park, Jongsun; Park, Bongsoo; Jung, Kyongyong; Jang, Suwang; Yu, Kwangyul; Choi, Jaeyoung; Kong, Sunghyung; Park, Jaejin; Kim, Seryun; Kim, Hyojeong; Kim, Soonok; Kim, Jihyun F.; Blair, Jaime E.; Lee, Kwangwon; Kang, Seogchan; Lee, Yong-Hwan
2008-01-01
Since the completion of the Saccharomyces cerevisiae genome sequencing project in 1996, the genomes of over 80 fungal species have been sequenced or are currently being sequenced. Resulting data provide opportunities for studying and comparing fungal biology and evolution at the genome level. To support such studies, the Comparative Fungal Genomics Platform (CFGP; http://cfgp.snu.ac.kr), a web-based multifunctional informatics workbench, was developed. The CFGP comprises three layers, including the basal layer, middleware and the user interface. The data warehouse in the basal layer contains standardized genome sequences of 65 fungal species. The middleware processes queries via six analysis tools, including BLAST, ClustalW, InterProScan, SignalP 3.0, PSORT II and a newly developed tool named BLASTMatrix. The BLASTMatrix permits the identification and visualization of genes homologous to a query across multiple species. The Data-driven User Interface (DUI) of the CFGP was built on a new concept of pre-collecting data and post-executing analysis instead of the ‘fill-in-the-form-and-press-SUBMIT’ user interfaces utilized by most bioinformatics sites. A tool termed Favorite, which supports the management of encapsulated sequence data and provides a personalized data repository to users, is another novel feature in the DUI. PMID:17947331
78 FR 20473 - National Practitioner Data Bank
Federal Register 2010, 2011, 2012, 2013, 2014
2013-04-05
... may self-query. Information under the HCQIA is reported by medical malpractice payers, state medical... Organizations (QIOs). Individual health care practitioners and entities may self-query. Information under... have access to this information. Individual practitioners, providers, and suppliers may self-query the...
Query by example video based on fuzzy c-means initialized by fixed clustering center
NASA Astrophysics Data System (ADS)
Hou, Sujuan; Zhou, Shangbo; Siddique, Muhammad Abubakar
2012-04-01
Currently, the high complexity of video contents has posed the following major challenges for fast retrieval: (1) efficient similarity measurements, and (2) efficient indexing on the compact representations. A video-retrieval strategy based on fuzzy c-means (FCM) is presented for querying by example. Initially, the query video is segmented and represented by a set of shots, each shot can be represented by a key frame, and then we used video processing techniques to find visual cues to represent the key frame. Next, because the FCM algorithm is sensitive to the initializations, here we initialized the cluster center by the shots of query video so that users could achieve appropriate convergence. After an FCM cluster was initialized by the query video, each shot of query video was considered a benchmark point in the aforesaid cluster, and each shot in the database possessed a class label. The similarity between the shots in the database with the same class label and benchmark point can be transformed into the distance between them. Finally, the similarity between the query video and the video in database was transformed into the number of similar shots. Our experimental results demonstrated the performance of this proposed approach.
NASA Technical Reports Server (NTRS)
Friedman, S. Z.; Walker, R. E.; Aitken, R. B.
1986-01-01
The Image Based Information System (IBIS) has been under development at the Jet Propulsion Laboratory (JPL) since 1975. It is a collection of more than 90 programs that enable processing of image, graphical, tabular data for spatial analysis. IBIS can be utilized to create comprehensive geographic data bases. From these data, an analyst can study various attributes describing characteristics of a given study area. Even complex combinations of disparate data types can be synthesized to obtain a new perspective on spatial phenomena. In 1984, new query software was developed enabling direct Boolean queries of IBIS data bases through the submission of easily understood expressions. An improved syntax methodology, a data dictionary, and display software simplified the analysts' tasks associated with building, executing, and subsequently displaying the results of a query. The primary purpose of this report is to describe the features and capabilities of the new query software. A secondary purpose of this report is to compare this new query software to the query software developed previously (Friedman, 1982). With respect to this topic, the relative merits and drawbacks of both approaches are covered.
NCBI2RDF: enabling full RDF-based access to NCBI databases.
Anguita, Alberto; García-Remesal, Miguel; de la Iglesia, Diana; Maojo, Victor
2013-01-01
RDF has become the standard technology for enabling interoperability among heterogeneous biomedical databases. The NCBI provides access to a large set of life sciences databases through a common interface called Entrez. However, the latter does not provide RDF-based access to such databases, and, therefore, they cannot be integrated with other RDF-compliant databases and accessed via SPARQL query interfaces. This paper presents the NCBI2RDF system, aimed at providing RDF-based access to the complete NCBI data repository. This API creates a virtual endpoint for servicing SPARQL queries over different NCBI repositories and presenting to users the query results in SPARQL results format, thus enabling this data to be integrated and/or stored with other RDF-compliant repositories. SPARQL queries are dynamically resolved, decomposed, and forwarded to the NCBI-provided E-utilities programmatic interface to access the NCBI data. Furthermore, we show how our approach increases the expressiveness of the native NCBI querying system, allowing several databases to be accessed simultaneously. This feature significantly boosts productivity when working with complex queries and saves time and effort to biomedical researchers. Our approach has been validated with a large number of SPARQL queries, thus proving its reliability and enhanced capabilities in biomedical environments.
Köhler, M J; Springer, S; Kaatz, M
2014-09-01
The volume of search engine queries about disease-relevant items reflects public interest and correlates with disease prevalence as proven by the example of flu (influenza). Other influences include media attention or holidays. The present work investigates if the seasonality of prevalence or symptom severity of dermatoses correlates with search engine query data. The relative weekly volume of dermatological relevant search terms was assessed by the online tool Google Trends for the years 2009-2013. For each item, the degree of seasonality was calculated via frequency analysis and a geometric approach. Many dermatoses show a marked seasonality, reflected by search engine query volumes. Unexpected seasonal variations of these queries suggest a previously unknown variability of the respective disease prevalence. Furthermore, using the example of allergic rhinitis, a close correlation of search engine query data with actual pollen count can be demonstrated. In many cases, search engine query data are appropriate to estimate seasonal variability in prevalence of common dermatoses. This finding may be useful for real-time analysis and formation of hypotheses concerning pathogenetic or symptom aggravating mechanisms and may thus contribute to improvement of diagnostics and prevention of skin diseases.
HodDB: Design and Analysis of a Query Processor for Brick.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Fierro, Gabriel; Culler, David
Brick is a recently proposed metadata schema and ontology for describing building components and the relationships between them. It represents buildings as directed labeled graphs using the RDF data model. Using the SPARQL query language, building-agnostic applications query a Brick graph to discover the set of resources and relationships they require to operate. Latency-sensitive applications, such as user interfaces, demand response and modelpredictive control, require fast queries — conventionally less than 100ms. We benchmark a set of popular open-source and commercial SPARQL databases against three real Brick models using seven application queries and find that none of them meet thismore » performance target. This lack of performance can be attributed to design decisions that optimize for queries over large graphs consisting of billions of triples, but give poor spatial locality and join performance on the small dense graphs typical of Brick. We present the design and evaluation of HodDB, a RDF/SPARQL database for Brick built over a node-based index structure. HodDB performs Brick queries 3-700x faster than leading SPARQL databases and consistently meets the 100ms threshold, enabling the portability of important latency-sensitive building applications.« less
NASA Astrophysics Data System (ADS)
Kase, Sue E.; Vanni, Michelle; Knight, Joanne A.; Su, Yu; Yan, Xifeng
2016-05-01
Within operational environments decisions must be made quickly based on the information available. Identifying an appropriate knowledge base and accurately formulating a search query are critical tasks for decision-making effectiveness in dynamic situations. The spreading of graph data management tools to access large graph databases is a rapidly emerging research area of potential benefit to the intelligence community. A graph representation provides a natural way of modeling data in a wide variety of domains. Graph structures use nodes, edges, and properties to represent and store data. This research investigates the advantages of information search by graph query initiated by the analyst and interactively refined within the contextual dimensions of the answer space toward a solution. The paper introduces SLQ, a user-friendly graph querying system enabling the visual formulation of schemaless and structureless graph queries. SLQ is demonstrated with an intelligence analyst information search scenario focused on identifying individuals responsible for manufacturing a mosquito-hosted deadly virus. The scenario highlights the interactive construction of graph queries without prior training in complex query languages or graph databases, intuitive navigation through the problem space, and visualization of results in graphical format.
Representation and alignment of sung queries for music information retrieval
NASA Astrophysics Data System (ADS)
Adams, Norman H.; Wakefield, Gregory H.
2005-09-01
The pursuit of robust and rapid query-by-humming systems, which search melodic databases using sung queries, is a common theme in music information retrieval. The retrieval aspect of this database problem has received considerable attention, whereas the front-end processing of sung queries and the data structure to represent melodies has been based on musical intuition and historical momentum. The present work explores three time series representations for sung queries: a sequence of notes, a ``smooth'' pitch contour, and a sequence of pitch histograms. The performance of the three representations is compared using a collection of naturally sung queries. It is found that the most robust performance is achieved by the representation with highest dimension, the smooth pitch contour, but that this representation presents a formidable computational burden. For all three representations, it is necessary to align the query and target in order to achieve robust performance. The computational cost of the alignment is quadratic, hence it is necessary to keep the dimension small for rapid retrieval. Accordingly, iterative deepening is employed to achieve both robust performance and rapid retrieval. Finally, the conventional iterative framework is expanded to adapt the alignment constraints based on previous iterations, further expediting retrieval without degrading performance.
Concept-based query language approach to enterprise information systems
NASA Astrophysics Data System (ADS)
Niemi, Timo; Junkkari, Marko; Järvelin, Kalervo
2014-01-01
In enterprise information systems (EISs) it is necessary to model, integrate and compute very diverse data. In advanced EISs the stored data often are based both on structured (e.g. relational) and semi-structured (e.g. XML) data models. In addition, the ad hoc information needs of end-users may require the manipulation of data-oriented (structural), behavioural and deductive aspects of data. Contemporary languages capable of treating this kind of diversity suit only persons with good programming skills. In this paper we present a concept-oriented query language approach to manipulate this diversity so that the programming skill requirements are considerably reduced. In our query language, the features which need technical knowledge are hidden in application-specific concepts and structures. Therefore, users need not be aware of the underlying technology. Application-specific concepts and structures are represented by the modelling primitives of the extended RDOOM (relational deductive object-oriented modelling) which contains primitives for all crucial real world relationships (is-a relationship, part-of relationship, association), XML documents and views. Our query language also supports intensional and extensional-intensional queries, in addition to conventional extensional queries. In its query formulation, the end-user combines available application-specific concepts and structures through shared variables.
Pavement performance evaluation, phase II : data collection.
DOT National Transportation Integrated Search
2008-12-01
Phase I and II of this study tested approximately 1500 rehabilitated pavements (asphalt and PCC) : throughout the State. These pavements ranged from 5 to 15 years old and were intended to develop a : snapshot of how various rehabilitations were perfo...
Improving traffic safety culture in Iowa : phase II.
DOT National Transportation Integrated Search
2013-07-01
Phase II of Improving Traffic Safety Culture in Iowa focuses on producing actions that will improve the traffic safety culture across the state, and involves collaboration among the three large public universities in Iowa: Iowa State University, Univ...
South Bay Salt Pond Tidal Wetland Restoration Phase II Planning
Information about the SFBWQP South Bay Salt Pond Tidal Wetland Restoration Phase II Planning project, part of an EPA competitive grant program to improve SF Bay water quality focused on restoring impaired waters and enhancing aquatic re
NASA Astrophysics Data System (ADS)
Puertas, Ricardo; Rute, Maria A.; Salud, Josep; López, David O.; Diez, Sergio; van Miltenburg, J. Kees; Pardo, Luis C.; Tamarit, Josep Ll.; Barrio, Maria; Pérez-Jubindo, Miguel A.; de La Fuente, Maria R.
2004-06-01
The stable solid polymorphism of cyclooctanol (C8H16O, for short C8 OH) is revealed to be a complex problem and only two stable solid phases, denoted on cooling from the liquid as phases I and II, are found using static (thermodynamic and x-ray diffraction) as well as dynamic (dielectric spectroscopy) experimental techniques. Both solid phases are known to exhibit glass transitions if they are cooled down fast enough to prevent transition to ordered crystalline states. Although glass transitions corresponding to both phases had been well documented by means of specific heat measurements, x-ray measurements constitute, as far as we know, the first evidence from the structural point of view. In addition, a great amount of dielectric works devoted to phase I and its glass transition, were published in the past but next to nothing relating to the dielectric properties of phase II and its glass transition. The nature of the disorder of phase II will be discussed.
Quality of reporting in oncology phase II trials: A 5-year assessment through systematic review.
Langrand-Escure, Julien; Rivoirard, Romain; Oriol, Mathieu; Tinquaut, Fabien; Rancoule, Chloé; Chauvin, Frank; Magné, Nicolas; Bourmaud, Aurélie
2017-01-01
Phase II clinical trials are a cornerstone of the development in experimental treatments They work as a "filter" for phase III trials confirmation. Surprisingly the attrition ratio in Phase III trials in oncology is significantly higher than in any other medical specialty. This suggests phase II trials in oncology fail to achieve their goal. Objective The present study aims at estimating the quality of reporting in published oncology phase II clinical trials. A literature review was conducted among all phase II and phase II/III clinical trials published during a 5-year period (2010-2015). All articles electronically published by three randomly-selected oncology journals with Impact-Factors>4 were included: Journal of Clinical Oncology, Annals of Oncology and British Journal of Cancer. Quality of reporting was assessed using the Key Methodological Score. 557 articles were included. 315 trials were single-arm studies (56.6%), 193 (34.6%) were randomized and 49 (8.8%) were non-randomized multiple-arm studies. The Methodological Score was equal to 0 (lowest level), 1, 2, 3 (highest level) respectively for 22 (3.9%), 119 (21.4%), 270 (48.5%) and 146 (26.2%) articles. The primary end point is almost systematically reported (90.5%), while sample size calculation is missing in 66% of the articles. 3 variables were independently associated with reporting of a high standard: presence of statistical design (p-value <0.001), multicenter trial (p-value = 0.012), per-protocol analysis (p-value <0.001). Screening was mainly performed by a sole author. The Key Methodological Score was based on only 3 items, making grey zones difficult to translate. This literature review highlights the existence of gaps concerning the quality of reporting. It therefore raised the question of the suitability of the methodology as well as the quality of these trials, reporting being incomplete in the corresponding articles.
Quality of reporting in oncology phase II trials: A 5-year assessment through systematic review
Langrand-Escure, Julien; Rivoirard, Romain; Oriol, Mathieu; Tinquaut, Fabien; Rancoule, Chloé; Chauvin, Frank; Magné, Nicolas; Bourmaud, Aurélie
2017-01-01
Background Phase II clinical trials are a cornerstone of the development in experimental treatments They work as a "filter" for phase III trials confirmation. Surprisingly the attrition ratio in Phase III trials in oncology is significantly higher than in any other medical specialty. This suggests phase II trials in oncology fail to achieve their goal. Objective The present study aims at estimating the quality of reporting in published oncology phase II clinical trials. Data sources A literature review was conducted among all phase II and phase II/III clinical trials published during a 5-year period (2010–2015). Study eligibility criteria All articles electronically published by three randomly-selected oncology journals with Impact-Factors>4 were included: Journal of Clinical Oncology, Annals of Oncology and British Journal of Cancer. Intervention Quality of reporting was assessed using the Key Methodological Score. Results 557 articles were included. 315 trials were single-arm studies (56.6%), 193 (34.6%) were randomized and 49 (8.8%) were non-randomized multiple-arm studies. The Methodological Score was equal to 0 (lowest level), 1, 2, 3 (highest level) respectively for 22 (3.9%), 119 (21.4%), 270 (48.5%) and 146 (26.2%) articles. The primary end point is almost systematically reported (90.5%), while sample size calculation is missing in 66% of the articles. 3 variables were independently associated with reporting of a high standard: presence of statistical design (p-value <0.001), multicenter trial (p-value = 0.012), per-protocol analysis (p-value <0.001). Limitations Screening was mainly performed by a sole author. The Key Methodological Score was based on only 3 items, making grey zones difficult to translate. Conclusions & implications of key findings This literature review highlights the existence of gaps concerning the quality of reporting. It therefore raised the question of the suitability of the methodology as well as the quality of these trials, reporting being incomplete in the corresponding articles. PMID:29216190
Bohu, Tsing; Santelli, Cara M; Akob, Denise M.; Neu, Thomas R; Ciobota, Valerian; Rösch, Petra; Popp, Jürgen; Nietzsche, Sándor; Küsel, Kirsten
2015-01-01
Despite the ubiquity of Mn oxides in natural environments, there are only a few observations of biological Mn(II) oxidation at pH < 6. The lack of low pH Mn-oxidizing bacteria (MOB) isolates limits our understanding of how pH influences biological Mn(II) oxidation in extreme environments. Here, we report that a novel MOB isolate, Mesorhizobium australicum strain T-G1, isolated from an acidic and metalliferous uranium mining area, can oxidize Mn(II) at both acidic and neutral pH using different enzymatic pathways. X-ray diffraction, Raman spectroscopy, and scanning electron microscopy with energy dispersive X-ray spectroscopy revealed that T-G1 initiated bixbyite-like Mn oxide formation at pH 5.5 which coincided with multi-copper oxidase expression from early exponential phase to late stationary phase. In contrast, reactive oxygen species (ROS), particularly superoxide, appeared to be more important for T-G1 mediated Mn(II) oxidation at neutral pH. ROS was produced in parallel with the occurrence of Mn(II) oxidation at pH 7.2 from early stationary phase. Solid phase Mn oxides did not precipitate, which is consistent with the presence of a high amount of H2O2 and lower activity of catalase in the liquid culture at pH 7.2. Our results show that M. australicum T-G1, an acid tolerant MOB, can initiate Mn(II) oxidation by varying its oxidation mechanisms depending on the pH and may play an important role in low pH manganese biogeochemical cycling.
Bohu, Tsing; Santelli, Cara M; Akob, Denise M; Neu, Thomas R; Ciobota, Valerian; Rösch, Petra; Popp, Jürgen; Nietzsche, Sándor; Küsel, Kirsten
2015-01-01
Despite the ubiquity of Mn oxides in natural environments, there are only a few observations of biological Mn(II) oxidation at pH < 6. The lack of low pH Mn-oxidizing bacteria (MOB) isolates limits our understanding of how pH influences biological Mn(II) oxidation in extreme environments. Here, we report that a novel MOB isolate, Mesorhizobium australicum strain T-G1, isolated from an acidic and metalliferous uranium mining area, can oxidize Mn(II) at both acidic and neutral pH using different enzymatic pathways. X-ray diffraction, Raman spectroscopy, and scanning electron microscopy with energy dispersive X-ray spectroscopy revealed that T-G1 initiated bixbyite-like Mn oxide formation at pH 5.5 which coincided with multi-copper oxidase expression from early exponential phase to late stationary phase. In contrast, reactive oxygen species (ROS), particularly superoxide, appeared to be more important for T-G1 mediated Mn(II) oxidation at neutral pH. ROS was produced in parallel with the occurrence of Mn(II) oxidation at pH 7.2 from early stationary phase. Solid phase Mn oxides did not precipitate, which is consistent with the presence of a high amount of H2O2 and lower activity of catalase in the liquid culture at pH 7.2. Our results show that M. australicum T-G1, an acid tolerant MOB, can initiate Mn(II) oxidation by varying its oxidation mechanisms depending on the pH and may play an important role in low pH manganese biogeochemical cycling.
Relativistic quantum private database queries
NASA Astrophysics Data System (ADS)
Sun, Si-Jia; Yang, Yu-Guang; Zhang, Ming-Ou
2015-04-01
Recently, Jakobi et al. (Phys Rev A 83, 022301, 2011) suggested the first practical private database query protocol (J-protocol) based on the Scarani et al. (Phys Rev Lett 92, 057901, 2004) quantum key distribution protocol. Unfortunately, the J-protocol is just a cheat-sensitive private database query protocol. In this paper, we present an idealized relativistic quantum private database query protocol based on Minkowski causality and the properties of quantum information. Also, we prove that the protocol is secure in terms of the user security and the database security.
Walter User’s Manual (Version 1.0).
1987-09-01
queries and/or commands. 1.2 - How Walter Uses the Screen As shown in Figure 1-1, Walter divides the screen of your terminal into five separate areas...our attention to queries and how to submit them to the database. 1.3.1 - Submitting Queries A query is an expression consisting of words, parentheses...dates, but also with ranges of dates, such as "oct 15 : nov 15". Waiter recognizes three kinds of dates: * Specific dates of the form [date <month> <day
Flexible Decision Support in Device-Saturated Environments
2003-10-01
also output tuples to a remote MySQL or Postgres database. 3.3 GUI The GUI allows the user to pose queries using SQL and to display query...DatabaseConnection.java – handles connections to an external database (such as MySQL or Postgres ). • Debug.java – contains the code for printing out Debug messages...also provided. It is possible to output the results of queries to a MySQL or Postgres database for archival and the GUI can query those results
Hintermair, Corinna; Voß, Kirsten; Forné, Ignasi; Heidemann, Martin; Flatley, Andrew; Kremmer, Elisabeth; Imhof, Axel; Eick, Dirk
2016-01-01
Dynamic phosphorylation of Tyr1-Ser2-Pro3-Thr4-Ser5-Pro6-Ser7 heptad-repeats in the C-terminal domain (CTD) of the large subunit coordinates progression of RNA polymerase (Pol) II through the transcription cycle. Here, we describe an M phase-specific form of Pol II phosphorylated at Thr4, but not at Tyr1, Ser2, Ser5, and Ser7 residues. Thr4 phosphorylated Pol II binds to centrosomes and midbody and interacts with the Thr4-specific Polo-like kinase 1. Binding of Pol II to centrosomes does not require the CTD but may involve subunits of the non-canonical R2TP-Prefoldin-like complex, which bind to and co-localize with Pol II at centrosomes. CTD Thr4 mutants, but not Ser2 and Ser5 mutants, display severe mitosis and cytokinesis defects characterized by multipolar spindles and polyploid cells. We conclude that proper M phase progression of cells requires binding of Pol II to centrosomes to facilitate regulation of mitosis and cytokinesis in a CTD Thr4-P dependent manner. PMID:27264542
Hintermair, Corinna; Voß, Kirsten; Forné, Ignasi; Heidemann, Martin; Flatley, Andrew; Kremmer, Elisabeth; Imhof, Axel; Eick, Dirk
2016-06-06
Dynamic phosphorylation of Tyr1-Ser2-Pro3-Thr4-Ser5-Pro6-Ser7 heptad-repeats in the C-terminal domain (CTD) of the large subunit coordinates progression of RNA polymerase (Pol) II through the transcription cycle. Here, we describe an M phase-specific form of Pol II phosphorylated at Thr4, but not at Tyr1, Ser2, Ser5, and Ser7 residues. Thr4 phosphorylated Pol II binds to centrosomes and midbody and interacts with the Thr4-specific Polo-like kinase 1. Binding of Pol II to centrosomes does not require the CTD but may involve subunits of the non-canonical R2TP-Prefoldin-like complex, which bind to and co-localize with Pol II at centrosomes. CTD Thr4 mutants, but not Ser2 and Ser5 mutants, display severe mitosis and cytokinesis defects characterized by multipolar spindles and polyploid cells. We conclude that proper M phase progression of cells requires binding of Pol II to centrosomes to facilitate regulation of mitosis and cytokinesis in a CTD Thr4-P dependent manner.
Implementation of a Proficiency-Based Diploma System in Maine: Phase II--District Level Analysis
ERIC Educational Resources Information Center
Silvernail, David L.; Stump, Erika K.; McCafferty, Anita Stewart; Hawes, Kathryn M.
2014-01-01
This report describes the findings from Phase II of a study of Maine's implementation of a proficiency-based diploma system. At the request of the Joint Standing Committee on Education and Cultural Affairs of the Maine Legislature, the Maine Policy Research Institute (MEPRI) has conducted a two-phased study of the implementation of Maine law…
40 CFR 76.8 - Early election for Group 1, Phase II boilers.
Code of Federal Regulations, 2010 CFR
2010-07-01
... PROGRAMS (CONTINUED) ACID RAIN NITROGEN OXIDES EMISSION REDUCTION PROGRAM § 76.8 Early election for Group 1... plan and: (i) If a Phase I Acid Rain permit governing the source at which the unit is located has been... chapter to include the early election plan; or (ii) If a Phase I Acid Rain permit governing the source at...
40 CFR 76.8 - Early election for Group 1, Phase II boilers.
Code of Federal Regulations, 2014 CFR
2014-07-01
... PROGRAMS (CONTINUED) ACID RAIN NITROGEN OXIDES EMISSION REDUCTION PROGRAM § 76.8 Early election for Group 1... plan and: (i) If a Phase I Acid Rain permit governing the source at which the unit is located has been... chapter to include the early election plan; or (ii) If a Phase I Acid Rain permit governing the source at...
40 CFR 76.8 - Early election for Group 1, Phase II boilers.
Code of Federal Regulations, 2011 CFR
2011-07-01
... PROGRAMS (CONTINUED) ACID RAIN NITROGEN OXIDES EMISSION REDUCTION PROGRAM § 76.8 Early election for Group 1... plan and: (i) If a Phase I Acid Rain permit governing the source at which the unit is located has been... chapter to include the early election plan; or (ii) If a Phase I Acid Rain permit governing the source at...
40 CFR 76.8 - Early election for Group 1, Phase II boilers.
Code of Federal Regulations, 2012 CFR
2012-07-01
... PROGRAMS (CONTINUED) ACID RAIN NITROGEN OXIDES EMISSION REDUCTION PROGRAM § 76.8 Early election for Group 1... plan and: (i) If a Phase I Acid Rain permit governing the source at which the unit is located has been... chapter to include the early election plan; or (ii) If a Phase I Acid Rain permit governing the source at...
40 CFR 76.8 - Early election for Group 1, Phase II boilers.
Code of Federal Regulations, 2013 CFR
2013-07-01
... PROGRAMS (CONTINUED) ACID RAIN NITROGEN OXIDES EMISSION REDUCTION PROGRAM § 76.8 Early election for Group 1... plan and: (i) If a Phase I Acid Rain permit governing the source at which the unit is located has been... chapter to include the early election plan; or (ii) If a Phase I Acid Rain permit governing the source at...
Extension and Public Service in the University of Illinois. Phase II Report.
ERIC Educational Resources Information Center
Illinois Univ., Urbana.
Phase II of the report on the problem outlined in Phase I deals with specific recommendations for expanding and improving the extension and public service functions of the University of Illinois. To be effective, the university needs a master plan in which the four essential ingredients must be (1) broad, strong and explicit policy commitments by…
Abdulla, Ahmed AbdoAziz Ahmed; Lin, Hongfei; Xu, Bo; Banbhrani, Santosh Kumar
2016-07-25
Biomedical literature retrieval is becoming increasingly complex, and there is a fundamental need for advanced information retrieval systems. Information Retrieval (IR) programs scour unstructured materials such as text documents in large reserves of data that are usually stored on computers. IR is related to the representation, storage, and organization of information items, as well as to access. In IR one of the main problems is to determine which documents are relevant and which are not to the user's needs. Under the current regime, users cannot precisely construct queries in an accurate way to retrieve particular pieces of data from large reserves of data. Basic information retrieval systems are producing low-quality search results. In our proposed system for this paper we present a new technique to refine Information Retrieval searches to better represent the user's information need in order to enhance the performance of information retrieval by using different query expansion techniques and apply a linear combinations between them, where the combinations was linearly between two expansion results at one time. Query expansions expand the search query, for example, by finding synonyms and reweighting original terms. They provide significantly more focused, particularized search results than do basic search queries. The retrieval performance is measured by some variants of MAP (Mean Average Precision) and according to our experimental results, the combination of best results of query expansion is enhanced the retrieved documents and outperforms our baseline by 21.06 %, even it outperforms a previous study by 7.12 %. We propose several query expansion techniques and their combinations (linearly) to make user queries more cognizable to search engines and to produce higher-quality search results.
A distributed query execution engine of big attributed graphs.
Batarfi, Omar; Elshawi, Radwa; Fayoumi, Ayman; Barnawi, Ahmed; Sakr, Sherif
2016-01-01
A graph is a popular data model that has become pervasively used for modeling structural relationships between objects. In practice, in many real-world graphs, the graph vertices and edges need to be associated with descriptive attributes. Such type of graphs are referred to as attributed graphs. G-SPARQL has been proposed as an expressive language, with a centralized execution engine, for querying attributed graphs. G-SPARQL supports various types of graph querying operations including reachability, pattern matching and shortest path where any G-SPARQL query may include value-based predicates on the descriptive information (attributes) of the graph edges/vertices in addition to the structural predicates. In general, a main limitation of centralized systems is that their vertical scalability is always restricted by the physical limits of computer systems. This article describes the design, implementation in addition to the performance evaluation of DG-SPARQL, a distributed, hybrid and adaptive parallel execution engine of G-SPARQL queries. In this engine, the topology of the graph is distributed over the main memory of the underlying nodes while the graph data are maintained in a relational store which is replicated on the disk of each of the underlying nodes. DG-SPARQL evaluates parts of the query plan via SQL queries which are pushed to the underlying relational stores while other parts of the query plan, as necessary, are evaluated via indexless memory-based graph traversal algorithms. Our experimental evaluation shows the efficiency and the scalability of DG-SPARQL on querying massive attributed graph datasets in addition to its ability to outperform the performance of Apache Giraph, a popular distributed graph processing system, by orders of magnitudes.
Folksonomical P2P File Sharing Networks Using Vectorized KANSEI Information as Search Tags
NASA Astrophysics Data System (ADS)
Ohnishi, Kei; Yoshida, Kaori; Oie, Yuji
We present the concept of folksonomical peer-to-peer (P2P) file sharing networks that allow participants (peers) to freely assign structured search tags to files. These networks are similar to folksonomies in the present Web from the point of view that users assign search tags to information distributed over a network. As a concrete example, we consider an unstructured P2P network using vectorized Kansei (human sensitivity) information as structured search tags for file search. Vectorized Kansei information as search tags indicates what participants feel about their files and is assigned by the participant to each of their files. A search query also has the same form of search tags and indicates what participants want to feel about files that they will eventually obtain. A method that enables file search using vectorized Kansei information is the Kansei query-forwarding method, which probabilistically propagates a search query to peers that are likely to hold more files having search tags that are similar to the query. The similarity between the search query and the search tags is measured in terms of their dot product. The simulation experiments examine if the Kansei query-forwarding method can provide equal search performance for all peers in a network in which only the Kansei information and the tendency with respect to file collection are different among all of the peers. The simulation results show that the Kansei query forwarding method and a random-walk-based query forwarding method, for comparison, work effectively in different situations and are complementary. Furthermore, the Kansei query forwarding method is shown, through simulations, to be superior to or equal to the random-walk based one in terms of search speed.