DocCube: Multi-Dimensional Visualization and Exploration of Large Document Sets.
ERIC Educational Resources Information Center
Mothe, Josiane; Chrisment, Claude; Dousset, Bernard; Alaux, Joel
2003-01-01
Describes a user interface that provides global visualizations of large document sets to help users formulate the query that corresponds to their information needs. Highlights include concept hierarchies that users can browse to specify and refine information needs; knowledge discovery in databases and texts; and multidimensional modeling.…
New generic indexing technology
NASA Technical Reports Server (NTRS)
Freeston, Michael
1996-01-01
There has been no fundamental change in the dynamic indexing methods supporting database systems since the invention of the B-tree twenty-five years ago. And yet the whole classical approach to dynamic database indexing has long since become inappropriate and increasingly inadequate. We are moving rapidly from the conventional one-dimensional world of fixed-structure text and numbers to a multi-dimensional world of variable structures, objects and images, in space and time. But, even before leaving the confines of conventional database indexing, the situation is highly unsatisfactory. In fact, our research has led us to question the basic assumptions of conventional database indexing. We have spent the past ten years studying the properties of multi-dimensional indexing methods, and in this paper we draw the strands of a number of developments together - some quite old, some very new, to show how we now have the basis for a new generic indexing technology for the next generation of database systems.
The Efficacy of Multidimensional Constraint Keys in Database Query Performance
ERIC Educational Resources Information Center
Cardwell, Leslie K.
2012-01-01
This work is intended to introduce a database design method to resolve the two-dimensional complexities inherent in the relational data model and its resulting performance challenges through abstract multidimensional constructs. A multidimensional constraint is derived and utilized to implement an indexed Multidimensional Key (MK) to abstract a…
Multidimensional Measurement of Poverty among Women in Sub-Saharan Africa
ERIC Educational Resources Information Center
Batana, Yele Maweki
2013-01-01
Since the seminal work of Sen, poverty has been recognized as a multidimensional phenomenon. The recent availability of relevant databases renewed the interest in this approach. This paper estimates multidimensional poverty among women in fourteen Sub-Saharan African countries using the Alkire and Foster multidimensional poverty measures, whose…
Integrating Scientific Array Processing into Standard SQL
NASA Astrophysics Data System (ADS)
Misev, Dimitar; Bachhuber, Johannes; Baumann, Peter
2014-05-01
We live in a time that is dominated by data. Data storage is cheap and more applications than ever accrue vast amounts of data. Storing the emerging multidimensional data sets efficiently, however, and allowing them to be queried by their inherent structure, is a challenge many databases have to face today. Despite the fact that multidimensional array data is almost always linked to additional, non-array information, array databases have mostly developed separately from relational systems, resulting in a disparity between the two database categories. The current SQL standard and SQL DBMS supports arrays - and in an extension also multidimensional arrays - but does so in a very rudimentary and inefficient way. This poster demonstrates the practicality of an SQL extension for array processing, implemented in a proof-of-concept multi-faceted system that manages a federation of array and relational database systems, providing transparent, efficient and scalable access to the heterogeneous data in them.
Risch, John S [Kennewick, WA; Dowson, Scott T [West Richland, WA; Hart, Michelle L [Richland, WA; Hatley, Wes L [Kennewick, WA
2008-05-13
A method of displaying correlations among information objects comprises receiving a query against a database; obtaining a query result set; and generating a visualization representing the components of the result set, the visualization including one of a plane and line to represent a data field, nodes representing data values, and links showing correlations among fields and values. Other visualization methods and apparatus are disclosed.
Risch, John S [Kennewick, WA; Dowson, Scott T [West Richland, WA
2012-03-06
A method of displaying correlations among information objects includes receiving a query against a database; obtaining a query result set; and generating a visualization representing the components of the result set, the visualization including one of a plane and line to represent a data field, nodes representing data values, and links showing correlations among fields and values. Other visualization methods and apparatus are disclosed.
Multidimensional Learner Model In Intelligent Learning System
NASA Astrophysics Data System (ADS)
Deliyska, B.; Rozeva, A.
2009-11-01
The learner model in an intelligent learning system (ILS) has to ensure the personalization (individualization) and the adaptability of e-learning in an online learner-centered environment. ILS is a distributed e-learning system whose modules can be independent and located in different nodes (servers) on the Web. This kind of e-learning is achieved through the resources of the Semantic Web and is designed and developed around a course, group of courses or specialty. An essential part of ILS is learner model database which contains structured data about learner profile and temporal status in the learning process of one or more courses. In the paper a learner model position in ILS is considered and a relational database is designed from learner's domain ontology. Multidimensional modeling agent for the source database is designed and resultant learner data cube is presented. Agent's modules are proposed with corresponding algorithms and procedures. Multidimensional (OLAP) analysis guidelines on the resultant learner module for designing dynamic learning strategy have been highlighted.
Computer systems and methods for the query and visualization multidimensional databases
Stolte, Chris; Tang, Diane L.; Hanrahan, Patrick
2017-04-25
A method of generating a data visualization is performed at a computer having a display, one or more processors, and memory. The memory stores one or more programs for execution by the one or more processors. The process receives user specification of a plurality of characteristics of a data visualization. The data visualization is based on data from a multidimensional database. The characteristics specify at least x-position and y-position of data marks corresponding to tuples of data retrieved from the database. The process generates a data visualization according to the specified plurality of characteristics. The data visualization has an x-axis defined based on data for one or more first fields from the database that specify x-position of the data marks and the data visualization has a y-axis defined based on data for one or more second fields from the database that specify y-position of the data marks.
Shatkay, Hagit; Pan, Fengxia; Rzhetsky, Andrey; Wilbur, W. John
2008-01-01
Motivation: Much current research in biomedical text mining is concerned with serving biologists by extracting certain information from scientific text. We note that there is no ‘average biologist’ client; different users have distinct needs. For instance, as noted in past evaluation efforts (BioCreative, TREC, KDD) database curators are often interested in sentences showing experimental evidence and methods. Conversely, lab scientists searching for known information about a protein may seek facts, typically stated with high confidence. Text-mining systems can target specific end-users and become more effective, if the system can first identify text regions rich in the type of scientific content that is of interest to the user, retrieve documents that have many such regions, and focus on fact extraction from these regions. Here, we study the ability to characterize and classify such text automatically. We have recently introduced a multi-dimensional categorization and annotation scheme, developed to be applicable to a wide variety of biomedical documents and scientific statements, while intended to support specific biomedical retrieval and extraction tasks. Results: The annotation scheme was applied to a large corpus in a controlled effort by eight independent annotators, where three individual annotators independently tagged each sentence. We then trained and tested machine learning classifiers to automatically categorize sentence fragments based on the annotation. We discuss here the issues involved in this task, and present an overview of the results. The latter strongly suggest that automatic annotation along most of the dimensions is highly feasible, and that this new framework for scientific sentence categorization is applicable in practice. Contact: shatkay@cs.queensu.ca PMID:18718948
A Graphics Design Framework to Visualize Multi-Dimensional Economic Datasets
ERIC Educational Resources Information Center
Chandramouli, Magesh; Narayanan, Badri; Bertoline, Gary R.
2013-01-01
This study implements a prototype graphics visualization framework to visualize multidimensional data. This graphics design framework serves as a "visual analytical database" for visualization and simulation of economic models. One of the primary goals of any kind of visualization is to extract useful information from colossal volumes of…
ComVisMD - compact visualization of multidimensional data: experimenting with cricket players data
NASA Astrophysics Data System (ADS)
Dandin, Shridhar B.; Ducassé, Mireille
2018-03-01
Database information is multidimensional and often displayed in tabular format (row/column display). Presented in aggregated form, multidimensional data can be used to analyze the records or objects. Online Analytical database Processing (OLAP) proposes mechanisms to display multidimensional data in aggregated forms. A choropleth map is a thematic map in which areas are colored in proportion to the measurement of a statistical variable being displayed, such as population density. They are used mostly for compact graphical representation of geographical information. We propose a system, ComVisMD inspired by choropleth map and the OLAP cube to visualize multidimensional data in a compact way. ComVisMD displays multidimensional data like OLAP Cube, where we are mapping an attribute a (first dimension, e.g. year started playing cricket) in vertical direction, object coloring based on b (second dimension, e.g. batting average), mapping varying-size circles based on attribute c (third dimension, e.g. highest score), mapping numbers based on attribute d (fourth dimension, e.g. matches played). We illustrate our approach on cricket players data, namely on two tables Country and Player. They have a large number of rows and columns: 246 rows and 17 columns for players of one country. ComVisMD’s visualization reduces the size of the tabular display by a factor of about 4, allowing users to grasp more information at a time than the bare table display.
Carazo, J M; Stelzer, E H
1999-01-01
The BioImage Database Project collects and structures multidimensional data sets recorded by various microscopic techniques relevant to modern life sciences. It provides, as precisely as possible, the circumstances in which the sample was prepared and the data were recorded. It grants access to the actual data and maintains links between related data sets. In order to promote the interdisciplinary approach of modern science, it offers a large set of key words, which covers essentially all aspects of microscopy. Nonspecialists can, therefore, access and retrieve significant information recorded and submitted by specialists in other areas. A key issue of the undertaking is to exploit the available technology and to provide a well-defined yet flexible structure for dealing with data. Its pivotal element is, therefore, a modern object relational database that structures the metadata and ameliorates the provision of a complete service. The BioImage database can be accessed through the Internet. Copyright 1999 Academic Press.
Primor, Liron; Pierce, Margaret E; Katzir, Tami
2011-12-01
The aim of this study was to investigate which cognitive and reading-related linguistic skills contribute to reading comprehension of narrative and expository texts. The study examined an Israeli national database of Hebrew-speaking readers in fourth grade, from which a subsample of 190 readers with a reading disability (RD) and 190 readers with no reading disability (NRD) was selected. IQ, text reading, reading comprehension, and various linguistic and cognitive skills were assessed. Structural equation modeling results suggested that both groups rely on lower level processes such as text reading accuracy and orthographic knowledge for reading comprehension of both genres. However, RD readers depend more heavily upon these lower level processes compared with NRD for whom higher level processes contribute more to reading comprehension. The various variables accounted for only 25-34% of reading comprehension variance, and possible explanations are discussed. Taken together, these findings highlight the variety of factors influencing reading comprehension and its multidimensional nature.
Peter, Samuel C; Whelan, James P; Pfund, Rory A; Meyers, Andrew W
2018-06-14
Although readability has been traditionally operationalized and even become synonymous with the concept of word and sentence length, modern text analysis theory and technology have shifted toward multidimensional comprehension-based analytic techniques. In an effort to make use of these advancements and demonstrate their general utility, 6 commonly used measures of gambling disorder were submitted to readability analyses using 2 of these advanced approaches, Coh-Metrix and Question Understanding Aid (QUAID), and one traditional approach, the Flesch-Kincaid Grade Level. As hypothesized, significant variation was found across measures, with some questionnaires emerging as more appropriate than others for use in samples that may include individuals with low literacy. Recommendations are made for the use of these modern approaches to readability to inform decisions on measure selection and development. (PsycINFO Database Record (c) 2018 APA, all rights reserved).
Hout, Michael C; Goldinger, Stephen D; Brady, Kyle J
2014-01-01
Cognitive theories in visual attention and perception, categorization, and memory often critically rely on concepts of similarity among objects, and empirically require measures of "sameness" among their stimuli. For instance, a researcher may require similarity estimates among multiple exemplars of a target category in visual search, or targets and lures in recognition memory. Quantifying similarity, however, is challenging when everyday items are the desired stimulus set, particularly when researchers require several different pictures from the same category. In this article, we document a new multidimensional scaling database with similarity ratings for 240 categories, each containing color photographs of 16-17 exemplar objects. We collected similarity ratings using the spatial arrangement method. Reports include: the multidimensional scaling solutions for each category, up to five dimensions, stress and fit measures, coordinate locations for each stimulus, and two new classifications. For each picture, we categorized the item's prototypicality, indexed by its proximity to other items in the space. We also classified pairs of images along a continuum of similarity, by assessing the overall arrangement of each MDS space. These similarity ratings will be useful to any researcher that wishes to control the similarity of experimental stimuli according to an objective quantification of "sameness."
Mining a Web Citation Database for Author Co-Citation Analysis.
ERIC Educational Resources Information Center
He, Yulan; Hui, Siu Cheung
2002-01-01
Proposes a mining process to automate author co-citation analysis based on the Web Citation Database, a data warehouse for storing citation indices of Web publications. Describes the use of agglomerative hierarchical clustering for author clustering and multidimensional scaling for displaying author cluster maps, and explains PubSearch, a…
Computer systems and methods for the query and visualization of multidimensional databases
Stolte, Chris; Tang, Diane L.; Hanrahan, Patrick
2006-08-08
A method and system for producing graphics. A hierarchical structure of a database is determined. A visual table, comprising a plurality of panes, is constructed by providing a specification that is in a language based on the hierarchical structure of the database. In some cases, this language can include fields that are in the database schema. The database is queried to retrieve a set of tuples in accordance with the specification. A subset of the set of tuples is associated with a pane in the plurality of panes.
Computer systems and methods for the query and visualization of multidimensional database
Stolte, Chris; Tang, Diane L.; Hanrahan, Patrick
2010-05-11
A method and system for producing graphics. A hierarchical structure of a database is determined. A visual table, comprising a plurality of panes, is constructed by providing a specification that is in a language based on the hierarchical structure of the database. In some cases, this language can include fields that are in the database schema. The database is queried to retrieve a set of tuples in accordance with the specification. A subset of the set of tuples is associated with a pane in the plurality of panes.
Al-Nasheri, Ahmed; Muhammad, Ghulam; Alsulaiman, Mansour; Ali, Zulfiqar; Mesallam, Tamer A; Farahat, Mohamed; Malki, Khalid H; Bencherif, Mohamed A
2017-01-01
Automatic voice-pathology detection and classification systems may help clinicians to detect the existence of any voice pathologies and the type of pathology from which patients suffer in the early stages. The main aim of this paper is to investigate Multidimensional Voice Program (MDVP) parameters to automatically detect and classify the voice pathologies in multiple databases, and then to find out which parameters performed well in these two processes. Samples of the sustained vowel /a/ of normal and pathological voices were extracted from three different databases, which have three voice pathologies in common. The selected databases in this study represent three distinct languages: (1) the Arabic voice pathology database; (2) the Massachusetts Eye and Ear Infirmary database (English database); and (3) the Saarbruecken Voice Database (German database). A computerized speech lab program was used to extract MDVP parameters as features, and an acoustical analysis was performed. The Fisher discrimination ratio was applied to rank the parameters. A t test was performed to highlight any significant differences in the means of the normal and pathological samples. The experimental results demonstrate a clear difference in the performance of the MDVP parameters using these databases. The highly ranked parameters also differed from one database to another. The best accuracies were obtained by using the three highest ranked MDVP parameters arranged according to the Fisher discrimination ratio: these accuracies were 99.68%, 88.21%, and 72.53% for the Saarbruecken Voice Database, the Massachusetts Eye and Ear Infirmary database, and the Arabic voice pathology database, respectively. Copyright © 2017 The Voice Foundation. Published by Elsevier Inc. All rights reserved.
Prologue: Reading Comprehension Is Not a Single Ability.
Catts, Hugh W; Kamhi, Alan G
2017-04-20
In this initial article of the clinical forum on reading comprehension, we argue that reading comprehension is not a single ability that can be assessed by one or more general reading measures or taught by a small set of strategies or approaches. We present evidence for a multidimensional view of reading comprehension that demonstrates how it varies as a function of reader ability, text, and task. The implications of this view for instruction of reading comprehension are considered. Reading comprehension is best conceptualized with a multidimensional model. The multidimensionality of reading comprehension means that instruction will be more effective when tailored to student performance with specific texts and tasks.
NALDB: nucleic acid ligand database for small molecules targeting nucleic acid
Kumar Mishra, Subodh; Kumar, Amit
2016-01-01
Nucleic acid ligand database (NALDB) is a unique database that provides detailed information about the experimental data of small molecules that were reported to target several types of nucleic acid structures. NALDB is the first ligand database that contains ligand information for all type of nucleic acid. NALDB contains more than 3500 ligand entries with detailed pharmacokinetic and pharmacodynamic information such as target name, target sequence, ligand 2D/3D structure, SMILES, molecular formula, molecular weight, net-formal charge, AlogP, number of rings, number of hydrogen bond donor and acceptor, potential energy along with their Ki, Kd, IC50 values. All these details at single platform would be helpful for the development and betterment of novel ligands targeting nucleic acids that could serve as a potential target in different diseases including cancers and neurological disorders. With maximum 255 conformers for each ligand entry, our database is a multi-conformer database and can facilitate the virtual screening process. NALDB provides powerful web-based search tools that make database searching efficient and simplified using option for text as well as for structure query. NALDB also provides multi-dimensional advanced search tool which can screen the database molecules on the basis of molecular properties of ligand provided by database users. A 3D structure visualization tool has also been included for 3D structure representation of ligands. NALDB offers an inclusive pharmacological information and the structurally flexible set of small molecules with their three-dimensional conformers that can accelerate the virtual screening and other modeling processes and eventually complement the nucleic acid-based drug discovery research. NALDB can be routinely updated and freely available on bsbe.iiti.ac.in/bsbe/naldb/HOME.php. Database URL: http://bsbe.iiti.ac.in/bsbe/naldb/HOME.php PMID:26896846
SMMRNA: a database of small molecule modulators of RNA
Mehta, Ankita; Sonam, Surabhi; Gouri, Isha; Loharch, Saurabh; Sharma, Deepak K.; Parkesh, Raman
2014-01-01
We have developed SMMRNA, an interactive database, available at http://www.smmrna.org, with special focus on small molecule ligands targeting RNA. Currently, SMMRNA consists of ∼770 unique ligands along with structural images of RNA molecules. Each ligand in the SMMRNA contains information such as Kd, Ki, IC50, ΔTm, molecular weight (MW), hydrogen donor and acceptor count, XlogP, number of rotatable bonds, number of aromatic rings and 2D and 3D structures. These parameters can be explored using text search, advanced search, substructure and similarity-based analysis tools that are embedded in SMMRNA. A structure editor is provided for 3D visualization of ligands. Advance analysis can be performed using substructure and OpenBabel-based chemical similarity fingerprints. Upload facility for both RNA and ligands is also provided. The physicochemical properties of the ligands were further examined using OpenBabel descriptors, hierarchical clustering, binning partition and multidimensional scaling. We have also generated a 3D conformation database of ligands to support the structure and ligand-based screening. SMMRNA provides comprehensive resource for further design, development and refinement of small molecule modulators for selective targeting of RNA molecules. PMID:24163098
NASA Technical Reports Server (NTRS)
Brodsky, Alexander; Segal, Victor E.
1999-01-01
The EOSCUBE constraint database system is designed to be a software productivity tool for high-level specification and efficient generation of EOSDIS and other scientific products. These products are typically derived from large volumes of multidimensional data which are collected via a range of scientific instruments.
Multidimensional human dynamics in mobile phone communications.
Quadri, Christian; Zignani, Matteo; Capra, Lorenzo; Gaito, Sabrina; Rossi, Gian Paolo
2014-01-01
In today's technology-assisted society, social interactions may be expressed through a variety of techno-communication channels, including online social networks, email and mobile phones (calls, text messages). Consequently, a clear grasp of human behavior through the diverse communication media is considered a key factor in understanding the formation of the today's information society. So far, all previous research on user communication behavior has focused on a sole communication activity. In this paper we move forward another step on this research path by performing a multidimensional study of human sociality as an expression of the use of mobile phones. The paper focuses on user temporal communication behavior in the interplay between the two complementary communication media, text messages and phone calls, that represent the bi-dimensional scenario of analysis. Our study provides a theoretical framework for analyzing multidimensional bursts as the most general burst category, that includes one-dimensional bursts as the simplest case, and offers empirical evidence of their nature by following the combined phone call/text message communication patterns of approximately one million people over three-month period. This quantitative approach enables the design of a generative model rooted in the three most significant features of the multidimensional burst - the number of dimensions, prevalence and interleaving degree - able to reproduce the main media usage attitude. The other findings of the paper include a novel multidimensional burst detection algorithm and an insight analysis of the human media selection process.
Solutions for medical databases optimal exploitation.
Branescu, I; Purcarea, V L; Dobrescu, R
2014-03-15
The paper discusses the methods to apply OLAP techniques for multidimensional databases that leverage the existing, performance-enhancing technique, known as practical pre-aggregation, by making this technique relevant to a much wider range of medical applications, as a logistic support to the data warehousing techniques. The transformations have practically low computational complexity and they may be implemented using standard relational database technology. The paper also describes how to integrate the transformed hierarchies in current OLAP systems, transparently to the user and proposes a flexible, "multimodel" federated system for extending OLAP querying to external object databases.
Chalkley, Robert J; Baker, Peter R; Hansen, Kirk C; Medzihradszky, Katalin F; Allen, Nadia P; Rexach, Michael; Burlingame, Alma L
2005-08-01
An in-depth analysis of a multidimensional chromatography-mass spectrometry dataset acquired on a quadrupole selecting, quadrupole collision cell, time-of-flight (QqTOF) geometry instrument was carried out. A total of 3269 CID spectra were acquired. Through manual verification of database search results and de novo interpretation of spectra 2368 spectra could be confidently determined as predicted tryptic peptides. A detailed analysis of the non-matching spectra was also carried out, highlighting what the non-matching spectra in a database search typically are composed of. The results of this comprehensive dataset study demonstrate that QqTOF instruments produce information-rich data of which a high percentage of the data is readily interpretable.
Computer systems and methods for the query and visualization of multidimensional databases
Stolte, Chris; Tang, Diane L; Hanrahan, Patrick
2014-04-29
In response to a user request, a computer generates a graphical user interface on a computer display. A schema information region of the graphical user interface includes multiple operand names, each operand name associated with one or more fields of a multi-dimensional database. A data visualization region of the graphical user interface includes multiple shelves. Upon detecting a user selection of the operand names and a user request to associate each user-selected operand name with a respective shelf in the data visualization region, the computer generates a visual table in the data visualization region in accordance with the associations between the operand names and the corresponding shelves. The visual table includes a plurality of panes, each pane having at least one axis defined based on data for the fields associated with a respective operand name.
Computer systems and methods for the query and visualization of multidimensional databases
Stolte, Chris [Palo Alto, CA; Tang, Diane L [Palo Alto, CA; Hanrahan, Patrick [Portola Valley, CA
2011-02-01
In response to a user request, a computer generates a graphical user interface on a computer display. A schema information region of the graphical user interface includes multiple operand names, each operand name associated with one or more fields of a multi-dimensional database. A data visualization region of the graphical user interface includes multiple shelves. Upon detecting a user selection of the operand names and a user request to associate each user-selected operand name with a respective shelf in the data visualization region, the computer generates a visual table in the data visualization region in accordance with the associations between the operand names and the corresponding shelves. The visual table includes a plurality of panes, each pane having at least one axis defined based on data for the fields associated with a respective operand name.
Computer systems and methods for the query and visualization of multidimensional databases
Stolte, Chris [Palo Alto, CA; Tang, Diane L [Palo Alto, CA; Hanrahan, Patrick [Portola Valley, CA
2012-03-20
In response to a user request, a computer generates a graphical user interface on a computer display. A schema information region of the graphical user interface includes multiple operand names, each operand name associated with one or more fields of a multi-dimensional database. A data visualization region of the graphical user interface includes multiple shelves. Upon detecting a user selection of the operand names and a user request to associate each user-selected operand name with a respective shelf in the data visualization region, the computer generates a visual table in the data visualization region in accordance with the associations between the operand names and the corresponding shelves. The visual table includes a plurality of panes, each pane having at least one axis defined based on data for the fields associated with a respective operand name.
Jin, Rui; Lin, Zhi-jian; Xue, Chun-miao; Zhang, Bing
2013-09-01
Knowledge Discovery in Databases is gaining attention and raising new hopes for traditional Chinese medicine (TCM) researchers. It is a useful tool in understanding and deciphering TCM theories. Aiming for a better understanding of Chinese herbal property theory (CHPT), this paper performed an improved association rule learning to analyze semistructured text in the book entitled Shennong's Classic of Materia Medica. The text was firstly annotated and transformed to well-structured multidimensional data. Subsequently, an Apriori algorithm was employed for producing association rules after the sensitivity analysis of parameters. From the confirmed 120 resulting rules that described the intrinsic relationships between herbal property (qi, flavor and their combinations) and herbal efficacy, two novel fundamental principles underlying CHPT were acquired and further elucidated: (1) the many-to-one mapping of herbal efficacy to herbal property; (2) the nonrandom overlap between the related efficacy of qi and flavor. This work provided an innovative knowledge about CHPT, which would be helpful for its modern research.
Nuclear Forensic Inferences Using Iterative Multidimensional Statistics
DOE Office of Scientific and Technical Information (OSTI.GOV)
Robel, M; Kristo, M J; Heller, M A
2009-06-09
Nuclear forensics involves the analysis of interdicted nuclear material for specific material characteristics (referred to as 'signatures') that imply specific geographical locations, production processes, culprit intentions, etc. Predictive signatures rely on expert knowledge of physics, chemistry, and engineering to develop inferences from these material characteristics. Comparative signatures, on the other hand, rely on comparison of the material characteristics of the interdicted sample (the 'questioned sample' in FBI parlance) with those of a set of known samples. In the ideal case, the set of known samples would be a comprehensive nuclear forensics database, a database which does not currently exist. Inmore » fact, our ability to analyze interdicted samples and produce an extensive list of precise materials characteristics far exceeds our ability to interpret the results. Therefore, as we seek to develop the extensive databases necessary for nuclear forensics, we must also develop the methods necessary to produce the necessary inferences from comparison of our analytical results with these large, multidimensional sets of data. In the work reported here, we used a large, multidimensional dataset of results from quality control analyses of uranium ore concentrate (UOC, sometimes called 'yellowcake'). We have found that traditional multidimensional techniques, such as principal components analysis (PCA), are especially useful for understanding such datasets and drawing relevant conclusions. In particular, we have developed an iterative partial least squares-discriminant analysis (PLS-DA) procedure that has proven especially adept at identifying the production location of unknown UOC samples. By removing classes which fell far outside the initial decision boundary, and then rebuilding the PLS-DA model, we have consistently produced better and more definitive attributions than with a single pass classification approach. Performance of the iterative PLS-DA method compared favorably to that of classification and regression tree (CART) and k nearest neighbor (KNN) algorithms, with the best combination of accuracy and robustness, as tested by classifying samples measured independently in our laboratories against the vendor QC based reference set.« less
The LSST Data Mining Research Agenda
NASA Astrophysics Data System (ADS)
Borne, K.; Becla, J.; Davidson, I.; Szalay, A.; Tyson, J. A.
2008-12-01
We describe features of the LSST science database that are amenable to scientific data mining, object classification, outlier identification, anomaly detection, image quality assurance, and survey science validation. The data mining research agenda includes: scalability (at petabytes scales) of existing machine learning and data mining algorithms; development of grid-enabled parallel data mining algorithms; designing a robust system for brokering classifications from the LSST event pipeline (which may produce 10,000 or more event alerts per night) multi-resolution methods for exploration of petascale databases; indexing of multi-attribute multi-dimensional astronomical databases (beyond spatial indexing) for rapid querying of petabyte databases; and more.
Solutions for medical databases optimal exploitation
Branescu, I; Purcarea, VL; Dobrescu, R
2014-01-01
The paper discusses the methods to apply OLAP techniques for multidimensional databases that leverage the existing, performance-enhancing technique, known as practical pre-aggregation, by making this technique relevant to a much wider range of medical applications, as a logistic support to the data warehousing techniques. The transformations have practically low computational complexity and they may be implemented using standard relational database technology. The paper also describes how to integrate the transformed hierarchies in current OLAP systems, transparently to the user and proposes a flexible, “multimodel" federated system for extending OLAP querying to external object databases. PMID:24653769
Biermann, Martin
2014-04-01
Clinical trials aiming for regulatory approval of a therapeutic agent must be conducted according to Good Clinical Practice (GCP). Clinical Data Management Systems (CDMS) are specialized software solutions geared toward GCP-trials. They are however less suited for data management in small non-GCP research projects. For use in researcher-initiated non-GCP studies, we developed a client-server database application based on the public domain CakePHP framework. The underlying MySQL database uses a simple data model based on only five data tables. The graphical user interface can be run in any web browser inside the hospital network. Data are validated upon entry. Data contained in external database systems can be imported interactively. Data are automatically anonymized on import, and the key lists identifying the subjects being logged to a restricted part of the database. Data analysis is performed by separate statistics and analysis software connecting to the database via a generic Open Database Connectivity (ODBC) interface. Since its first pilot implementation in 2011, the solution has been applied to seven different clinical research projects covering different clinical problems in different organ systems such as cancer of the thyroid and the prostate glands. This paper shows how the adoption of a generic web application framework is a feasible, flexible, low-cost, and user-friendly way of managing multidimensional research data in researcher-initiated non-GCP clinical projects. Copyright © 2014 The Authors. Published by Elsevier Ireland Ltd.. All rights reserved.
A Systematic Review of Studies Using the Multidimensional Assessment of Fatigue Scale.
Belza, Basia; Miyawaki, Christina E; Liu, Minhui; Aree-Ue, Suparb; Fessel, Melissa; Minott, Kenya R; Zhang, Xi
2018-04-01
To review how the Multidimensional Assessment of Fatigue (MAF) has been used and evaluate its psychometric properties. We conducted a database search using "multidimensional assessment of fatigue" or "MAF" as key terms from 1993 to 2015, and located 102 studies. Eighty-three were empirical studies and 19 were reviews/evaluations. Research was conducted in 17 countries; 32 diseases were represented. Nine language versions of the MAF were used. The mean of the Global Fatigue Index ranged from 10.9 to 49.4. The MAF was reported to be easy-to-use, had strong reliability and validity, and was used in populations who spoke languages other than English. The MAF is an acceptable assessment tool to measure fatigue and intervention effectiveness in various languages, diseases, and settings across the world.
Computer systems and methods for the query and visualization of multidimensional databases
Stolte, Chris; Tang, Diane L; Hanrahan, Patrick
2015-03-03
A computer displays a graphical user interface on its display. The graphical user interface includes a schema information region and a data visualization region. The schema information region includes multiple operand names, each operand corresponding to one or more fields of a multi-dimensional database that includes at least one data hierarchy. The data visualization region includes a columns shelf and a rows shelf. The computer detects user actions to associate one or more first operands with the columns shelf and to associate one or more second operands with the rows shelf. The computer generates a visual table in the data visualization region in accordance with the user actions. The visual table includes one or more panes. Each pane has an x-axis defined based on data for the one or more first operands, and each pane has a y-axis defined based on data for the one or more second operands.
Integrated Array/Metadata Analytics
NASA Astrophysics Data System (ADS)
Misev, Dimitar; Baumann, Peter
2015-04-01
Data comes in various forms and types, and integration usually presents a problem that is often simply ignored and solved with ad-hoc solutions. Multidimensional arrays are an ubiquitous data type, that we find at the core of virtually all science and engineering domains, as sensor, model, image, statistics data. Naturally, arrays are richly described by and intertwined with additional metadata (alphanumeric relational data, XML, JSON, etc). Database systems, however, a fundamental building block of what we call "Big Data", lack adequate support for modelling and expressing these array data/metadata relationships. Array analytics is hence quite primitive or non-existent at all in modern relational DBMS. Recognizing this, we extended SQL with a new SQL/MDA part seamlessly integrating multidimensional array analytics into the standard database query language. We demonstrate the benefits of SQL/MDA with real-world examples executed in ASQLDB, an open-source mediator system based on HSQLDB and rasdaman, that already implements SQL/MDA.
Computer systems and methods for the query and visualization of multidimensional databases
Stolte, Chris; Tang, Diane L.; Hanrahan, Patrick
2015-11-10
A computer displays a graphical user interface on its display. The graphical user interface includes a schema information region and a data visualization region. The schema information region includes a plurality of fields of a multi-dimensional database that includes at least one data hierarchy. The data visualization region includes a columns shelf and a rows shelf. The computer detects user actions to associate one or more first fields with the columns shelf and to associate one or more second fields with the rows shelf. The computer generates a visual table in the data visualization region in accordance with the user actions. The visual table includes one or more panes. Each pane has an x-axis defined based on data for the one or more first fields, and each pane has a y-axis defined based on data for the one or more second fields.
NALDB: nucleic acid ligand database for small molecules targeting nucleic acid.
Kumar Mishra, Subodh; Kumar, Amit
2016-01-01
Nucleic acid ligand database (NALDB) is a unique database that provides detailed information about the experimental data of small molecules that were reported to target several types of nucleic acid structures. NALDB is the first ligand database that contains ligand information for all type of nucleic acid. NALDB contains more than 3500 ligand entries with detailed pharmacokinetic and pharmacodynamic information such as target name, target sequence, ligand 2D/3D structure, SMILES, molecular formula, molecular weight, net-formal charge, AlogP, number of rings, number of hydrogen bond donor and acceptor, potential energy along with their Ki, Kd, IC50 values. All these details at single platform would be helpful for the development and betterment of novel ligands targeting nucleic acids that could serve as a potential target in different diseases including cancers and neurological disorders. With maximum 255 conformers for each ligand entry, our database is a multi-conformer database and can facilitate the virtual screening process. NALDB provides powerful web-based search tools that make database searching efficient and simplified using option for text as well as for structure query. NALDB also provides multi-dimensional advanced search tool which can screen the database molecules on the basis of molecular properties of ligand provided by database users. A 3D structure visualization tool has also been included for 3D structure representation of ligands. NALDB offers an inclusive pharmacological information and the structurally flexible set of small molecules with their three-dimensional conformers that can accelerate the virtual screening and other modeling processes and eventually complement the nucleic acid-based drug discovery research. NALDB can be routinely updated and freely available on bsbe.iiti.ac.in/bsbe/naldb/HOME.php. Database URL: http://bsbe.iiti.ac.in/bsbe/naldb/HOME.php. © The Author(s) 2016. Published by Oxford University Press.
On continuous user authentication via typing behavior.
Roth, Joseph; Liu, Xiaoming; Metaxas, Dimitris
2014-10-01
We hypothesize that an individual computer user has a unique and consistent habitual pattern of hand movements, independent of the text, while typing on a keyboard. As a result, this paper proposes a novel biometric modality named typing behavior (TB) for continuous user authentication. Given a webcam pointing toward a keyboard, we develop real-time computer vision algorithms to automatically extract hand movement patterns from the video stream. Unlike the typical continuous biometrics, such as keystroke dynamics (KD), TB provides a reliable authentication with a short delay, while avoiding explicit key-logging. We collect a video database where 63 unique subjects type static text and free text for multiple sessions. For one typing video, the hands are segmented in each frame and a unique descriptor is extracted based on the shape and position of hands, as well as their temporal dynamics in the video sequence. We propose a novel approach, named bag of multi-dimensional phrases, to match the cross-feature and cross-temporal pattern between a gallery sequence and probe sequence. The experimental results demonstrate a superior performance of TB when compared with KD, which, together with our ultrareal-time demo system, warrant further investigation of this novel vision application and biometric modality.
Naz, Saeeda; Umar, Arif Iqbal; Ahmed, Riaz; Razzak, Muhammad Imran; Rashid, Sheikh Faisal; Shafait, Faisal
2016-01-01
The recognition of Arabic script and its derivatives such as Urdu, Persian, Pashto etc. is a difficult task due to complexity of this script. Particularly, Urdu text recognition is more difficult due to its Nasta'liq writing style. Nasta'liq writing style inherits complex calligraphic nature, which presents major issues to recognition of Urdu text owing to diagonality in writing, high cursiveness, context sensitivity and overlapping of characters. Therefore, the work done for recognition of Arabic script cannot be directly applied to Urdu recognition. We present Multi-dimensional Long Short Term Memory (MDLSTM) Recurrent Neural Networks with an output layer designed for sequence labeling for recognition of printed Urdu text-lines written in the Nasta'liq writing style. Experiments show that MDLSTM attained a recognition accuracy of 98% for the unconstrained Urdu Nasta'liq printed text, which significantly outperforms the state-of-the-art techniques.
Härtig, Claus
2008-01-04
A multidimensional approach for the identification of fatty acid methyl esters (FAME) based on GC/MS analysis is described. Mass spectra and retention data of more than 130 FAME from various sources (chain lengths in the range from 4 to 24 carbon atoms) were collected in a database. Hints for the interpretation of FAME mass spectra are given and relevant diagnostic marker ions are deduced indicating specific groups of fatty acids. To verify the identity of single species and to ensure an optimized chromatographic resolution, the database was compiled with retention data libraries acquired on columns of different polarity (HP-5, DB-23, and HP-88). For a combined use of mass spectra and retention data standardized methods of measurement for each of these columns are required. Such master methods were developed and always applied under the conditions of retention time locking (RTL) which allowed an excellent reproducibility and comparability of absolute retention times. Moreover, as a relative retention index system, equivalent chain lengths (ECL) of FAME were determined by linear interpolation. To compare and to predict ECL values by means of structural features, fractional chain lengths (FCL) were calculated and fitted as well. As shown in an example, the use of retention data and mass spectral information together in a database search leads to an improved and reliable identification of FAME (including positional and geometrical isomers) without further derivatizations.
Rasdaman for Big Spatial Raster Data
NASA Astrophysics Data System (ADS)
Hu, F.; Huang, Q.; Scheele, C. J.; Yang, C. P.; Yu, M.; Liu, K.
2015-12-01
Spatial raster data have grown exponentially over the past decade. Recent advancements on data acquisition technology, such as remote sensing, have allowed us to collect massive observation data of various spatial resolution and domain coverage. The volume, velocity, and variety of such spatial data, along with the computational intensive nature of spatial queries, pose grand challenge to the storage technologies for effective big data management. While high performance computing platforms (e.g., cloud computing) can be used to solve the computing-intensive issues in big data analysis, data has to be managed in a way that is suitable for distributed parallel processing. Recently, rasdaman (raster data manager) has emerged as a scalable and cost-effective database solution to store and retrieve massive multi-dimensional arrays, such as sensor, image, and statistics data. Within this paper, the pros and cons of using rasdaman to manage and query spatial raster data will be examined and compared with other common approaches, including file-based systems, relational databases (e.g., PostgreSQL/PostGIS), and NoSQL databases (e.g., MongoDB and Hive). Earth Observing System (EOS) data collected from NASA's Atmospheric Scientific Data Center (ASDC) will be used and stored in these selected database systems, and a set of spatial and non-spatial queries will be designed to benchmark their performance on retrieving large-scale, multi-dimensional arrays of EOS data. Lessons learnt from using rasdaman will be discussed as well.
An empirical study of multidimensional fidelity of COMPASS consultation.
Wong, Venus; Ruble, Lisa A; McGrew, John H; Yu, Yue
2018-06-01
Consultation is essential to the daily practice of school psychologists (National Association of School Psychologist, 2010). Successful consultation requires fidelity at both the consultant (implementation) and consultee (intervention) levels. We applied a multidimensional, multilevel conception of fidelity (Dunst, Trivette, & Raab, 2013) to a consultative intervention called the Collaborative Model for Promoting Competence and Success (COMPASS) for students with autism. The study provided 3 main findings. First, multidimensional, multilevel fidelity is a stable construct and increases over time with consultation support. Second, mediation analyses revealed that implementation-level fidelity components had distant, indirect effects on student Individualized Education Program (IEP) outcomes. Third, 3 fidelity components correlated with IEP outcomes: teacher coaching responsiveness at the implementation level, and teacher quality of delivery and student responsiveness at the intervention levels. Implications and future directions are discussed. (PsycINFO Database Record (c) 2018 APA, all rights reserved).
Wallach, Geraldine P; Ocampo, Alaine
2017-04-20
In this discussion as part of a response to Catts and Kamhi's "Prologue: Reading Comprehension Is Not a Single Activity" (2017), the authors provide selected examples from 4th-, 5th-, and 6th-grade texts to demonstrate, in agreement with Catts and Kamhi, that reading comprehension is a multifaceted and complex ability. The authors were asked to provide readers with evidence-based practices that lend support to applications of a multidimensional model of comprehension. We present examples from the reading comprehension literature that support the notion that reading is a complex set of abilities that include a reader's ability, especially background knowledge; the type of text the reader is being asked to comprehend; and the task or technique used in assessment or intervention paradigms. An intervention session from 6th grade serves to demonstrate how background knowledge, a text's demands, and tasks may come together in the real world as clinicians and educators aim to help students comprehend complex material. The authors agree with the conceptual framework proposed by Catts and Kamhi that clinicians and educators should consider the multidimensional nature of reading comprehension (an interaction of reader, text, and task) when creating assessment and intervention programs. The authors might depart slightly by considering, more closely, those reading comprehension strategies that might facilitate comprehension across texts and tasks with an understanding of students' individual needs at different points in time.
Loop-Extended Symbolic Execution on Binary Programs
2009-03-02
1434. Based on its speci- fication [35], one valid message format contains 2 fields: a header byte of value 4, followed by a string giving a database ...potentially become expensive. For instance the polyhedron technique [16] requires costly conversion operations on a multi-dimensional abstract representation
RDBMS Based Lexical Resource for Indian Heritage: The Case of Mahābhārata
NASA Astrophysics Data System (ADS)
Mani, Diwakar
The paper describes a lexical resource in the form of a relational database based indexing system for Sanskrit documents - Mahābhārata (MBh) as an example. The system is available online on http://sanskrit.jnu.ac.in/mb with input and output in Devanāgarī Unicode, using technologies such as RDBMS and Java Servlet. The system works as an interactive and multi-dimensional indexing system with search facility for MBh and has potentials for use as a generic system for all Sanskrit texts of similar structure. Currently, the system allows three types of searching facilities- 'Direct Search', 'Alphabetical Search' and 'Search by Classes'. The input triggers an indexing process by which a temporary index is created for the search string, and then clicking on any indexed word displays the details for that word and also a facility to search that word in some other online lexical resources.
Empirical Determination of Competence Areas to Computer Science Education
ERIC Educational Resources Information Center
Zendler, Andreas; Klaudt, Dieter; Seitz, Cornelia
2014-01-01
The authors discuss empirically determined competence areas to K-12 computer science education, emphasizing the cognitive level of competence. The results of a questionnaire with 120 professors of computer science serve as a database. By using multi-dimensional scaling and cluster analysis, four competence areas to computer science education…
SorghumFDB: sorghum functional genomics database with multidimensional network analysis.
Tian, Tian; You, Qi; Zhang, Liwei; Yi, Xin; Yan, Hengyu; Xu, Wenying; Su, Zhen
2016-01-01
Sorghum (Sorghum bicolor [L.] Moench) has excellent agronomic traits and biological properties, such as heat and drought-tolerance. It is a C4 grass and potential bioenergy-producing plant, which makes it an important crop worldwide. With the sorghum genome sequence released, it is essential to establish a sorghum functional genomics data mining platform. We collected genomic data and some functional annotations to construct a sorghum functional genomics database (SorghumFDB). SorghumFDB integrated knowledge of sorghum gene family classifications (transcription regulators/factors, carbohydrate-active enzymes, protein kinases, ubiquitins, cytochrome P450, monolignol biosynthesis related enzymes, R-genes and organelle-genes), detailed gene annotations, miRNA and target gene information, orthologous pairs in the model plants Arabidopsis, rice and maize, gene loci conversions and a genome browser. We further constructed a dynamic network of multidimensional biological relationships, comprised of the co-expression data, protein-protein interactions and miRNA-target pairs. We took effective measures to combine the network, gene set enrichment and motif analyses to determine the key regulators that participate in related metabolic pathways, such as the lignin pathway, which is a major biological process in bioenergy-producing plants.Database URL: http://structuralbiology.cau.edu.cn/sorghum/index.html. © The Author(s) 2016. Published by Oxford University Press.
Multidimensional indexing structure for use with linear optimization queries
NASA Technical Reports Server (NTRS)
Bergman, Lawrence David (Inventor); Castelli, Vittorio (Inventor); Chang, Yuan-Chi (Inventor); Li, Chung-Sheng (Inventor); Smith, John Richard (Inventor)
2002-01-01
Linear optimization queries, which usually arise in various decision support and resource planning applications, are queries that retrieve top N data records (where N is an integer greater than zero) which satisfy a specific optimization criterion. The optimization criterion is to either maximize or minimize a linear equation. The coefficients of the linear equation are given at query time. Methods and apparatus are disclosed for constructing, maintaining and utilizing a multidimensional indexing structure of database records to improve the execution speed of linear optimization queries. Database records with numerical attributes are organized into a number of layers and each layer represents a geometric structure called convex hull. Such linear optimization queries are processed by searching from the outer-most layer of this multi-layer indexing structure inwards. At least one record per layer will satisfy the query criterion and the number of layers needed to be searched depends on the spatial distribution of records, the query-issued linear coefficients, and N, the number of records to be returned. When N is small compared to the total size of the database, answering the query typically requires searching only a small fraction of all relevant records, resulting in a tremendous speedup as compared to linearly scanning the entire dataset.
NASA Astrophysics Data System (ADS)
Verma, Surendra P.; Rivera-Gómez, M. Abdelaly; Díaz-González, Lorena; Pandarinath, Kailasa; Amezcua-Valdez, Alejandra; Rosales-Rivera, Mauricio; Verma, Sanjeet K.; Quiroz-Ruiz, Alfredo; Armstrong-Altrin, John S.
2017-05-01
A new multidimensional scheme consistent with the International Union of Geological Sciences (IUGS) is proposed for the classification of igneous rocks in terms of four magma types: ultrabasic, basic, intermediate, and acid. Our procedure is based on an extensive database of major element composition of a total of 33,868 relatively fresh rock samples having a multinormal distribution (initial database with 37,215 samples). Multinormally distributed database in terms of log-ratios of samples was ascertained by a new computer program DOMuDaF, in which the discordancy test was applied at the 99.9% confidence level. Isometric log-ratio (ilr) transformation was used to provide overall percent correct classification of 88.7%, 75.8%, 88.0%, and 80.9% for ultrabasic, basic, intermediate, and acid rocks, respectively. Given the known mathematical and uncertainty propagation properties, this transformation could be adopted for routine applications. The incorrect classification was mainly for the "neighbour" magma types, e.g., basic for ultrabasic and vice versa. Some of these misclassifications do not have any effect on multidimensional tectonic discrimination. For an efficient application of this multidimensional scheme, a new computer program MagClaMSys_ilr (MagClaMSys-Magma Classification Major-element based System) was written, which is available for on-line processing on http://tlaloc.ier.unam.mx/index.html. This classification scheme was tested from newly compiled data for relatively fresh Neogene igneous rocks and was found to be consistent with the conventional IUGS procedure. The new scheme was successfully applied to inter-laboratory data for three geochemical reference materials (basalts JB-1 and JB-1a, and andesite JA-3) from Japan and showed that the inferred magma types are consistent with the rock name (basic for basalts JB-1 and JB-1a and intermediate for andesite JA-3). The scheme was also successfully applied to five case studies of older Archaean to Mesozoic igneous rocks. Similar or more reliable results were obtained from existing tectonomagmatic discrimination diagrams when used in conjunction with the new computer program as compared to the IUGS scheme. The application to three case studies of igneous provenance of sedimentary rocks was demonstrated as a novel approach. Finally, we show that the new scheme is more robust for post-emplacement compositional changes than the conventional IUGS procedure.
Developing a Multi-Dimensional Hydrodynamics Code with Astrochemical Reactions
NASA Astrophysics Data System (ADS)
Kwak, Kyujin; Yang, Seungwon
2015-08-01
The Atacama Large Millimeter/submillimeter Array (ALMA) revealed high resolution molecular lines some of which are still unidentified yet. Because formation of these astrochemical molecules has been seldom studied in traditional chemistry, observations of new molecular lines drew a lot of attention from not only astronomers but also chemists both experimental and theoretical. Theoretical calculations for the formation of these astrochemical molecules have been carried out providing reaction rates for some important molecules, and some of theoretical predictions have been measured in laboratories. The reaction rates for the astronomically important molecules are now collected to form databases some of which are publically available. By utilizing these databases, we develop a multi-dimensional hydrodynamics code that includes the reaction rates of astrochemical molecules. Because this type of hydrodynamics code is able to trace the molecular formation in a non-equilibrium fashion, it is useful to study the formation history of these molecules that affects the spatial distribution of some specific molecules. We present the development procedure of this code and some test problems in order to verify and validate the developed code.
Statistical Downscaling in Multi-dimensional Wave Climate Forecast
NASA Astrophysics Data System (ADS)
Camus, P.; Méndez, F. J.; Medina, R.; Losada, I. J.; Cofiño, A. S.; Gutiérrez, J. M.
2009-04-01
Wave climate at a particular site is defined by the statistical distribution of sea state parameters, such as significant wave height, mean wave period, mean wave direction, wind velocity, wind direction and storm surge. Nowadays, long-term time series of these parameters are available from reanalysis databases obtained by numerical models. The Self-Organizing Map (SOM) technique is applied to characterize multi-dimensional wave climate, obtaining the relevant "wave types" spanning the historical variability. This technique summarizes multi-dimension of wave climate in terms of a set of clusters projected in low-dimensional lattice with a spatial organization, providing Probability Density Functions (PDFs) on the lattice. On the other hand, wind and storm surge depend on instantaneous local large-scale sea level pressure (SLP) fields while waves depend on the recent history of these fields (say, 1 to 5 days). Thus, these variables are associated with large-scale atmospheric circulation patterns. In this work, a nearest-neighbors analog method is used to predict monthly multi-dimensional wave climate. This method establishes relationships between the large-scale atmospheric circulation patterns from numerical models (SLP fields as predictors) with local wave databases of observations (monthly wave climate SOM PDFs as predictand) to set up statistical models. A wave reanalysis database, developed by Puertos del Estado (Ministerio de Fomento), is considered as historical time series of local variables. The simultaneous SLP fields calculated by NCEP atmospheric reanalysis are used as predictors. Several applications with different size of sea level pressure grid and with different temporal domain resolution are compared to obtain the optimal statistical model that better represents the monthly wave climate at a particular site. In this work we examine the potential skill of this downscaling approach considering perfect-model conditions, but we will also analyze the suitability of this methodology to be used for seasonal forecast and for long-term climate change scenario projection of wave climate.
Oceans 2.0: Interactive tools for the Visualization of Multi-dimensional Ocean Sensor Data
NASA Astrophysics Data System (ADS)
Biffard, B.; Valenzuela, M.; Conley, P.; MacArthur, M.; Tredger, S.; Guillemot, E.; Pirenne, B.
2016-12-01
Ocean Networks Canada (ONC) operates ocean observatories on all three of Canada's coasts. The instruments produce 280 gigabytes of data per day with 1/2 petabyte archived so far. In 2015, 13 terabytes were downloaded by over 500 users from across the world. ONC's data management system is referred to as "Oceans 2.0" owing to its interactive, participative features. A key element of Oceans 2.0 is real time data acquisition and processing: custom device drivers implement the input-output protocol of each instrument. Automatic parsing and calibration takes place on the fly, followed by event detection and quality control. All raw data are stored in a file archive, while the processed data are copied to fast databases. Interactive access to processed data is provided through data download and visualization/quick look features that are adapted to diverse data types (scalar, acoustic, video, multi-dimensional, etc). Data may be post or re-processed to add features, analysis or correct errors, update calibrations, etc. A robust storage structure has been developed consisting of an extensive file system and a no-SQL database (Cassandra). Cassandra is a node-based open source distributed database management system. It is scalable and offers improved performance for big data. A key feature is data summarization. The system has also been integrated with web services and an ERDDAP OPeNDAP server, capable of serving scalar and multidimensional data from Cassandra for fixed or mobile devices.A complex data viewer has been developed making use of the big data capability to interactively display live or historic echo sounder and acoustic Doppler current profiler data, where users can scroll, apply processing filters and zoom through gigabytes of data with simple interactions. This new technology brings scientists one step closer to a comprehensive, web-based data analysis environment in which visual assessment, filtering, event detection and annotation can be integrated.
ERIC Educational Resources Information Center
Castillo, Cristina; Tolchinsky, Liliana
2018-01-01
Building a text is a multidimensional endeavor. Writers must work simultaneously on the content of the text, its discursive organization, the structure of the sentences, and the individual words themselves. Knowledge of vocabulary is central to this endeavor. This study intends (1) to trace the development of writer's vocabulary depth, their…
Protein Simulation Data in the Relational Model.
Simms, Andrew M; Daggett, Valerie
2012-10-01
High performance computing is leading to unprecedented volumes of data. Relational databases offer a robust and scalable model for storing and analyzing scientific data. However, these features do not come without a cost-significant design effort is required to build a functional and efficient repository. Modeling protein simulation data in a relational database presents several challenges: the data captured from individual simulations are large, multi-dimensional, and must integrate with both simulation software and external data sites. Here we present the dimensional design and relational implementation of a comprehensive data warehouse for storing and analyzing molecular dynamics simulations using SQL Server.
Protein Simulation Data in the Relational Model
Simms, Andrew M.; Daggett, Valerie
2011-01-01
High performance computing is leading to unprecedented volumes of data. Relational databases offer a robust and scalable model for storing and analyzing scientific data. However, these features do not come without a cost—significant design effort is required to build a functional and efficient repository. Modeling protein simulation data in a relational database presents several challenges: the data captured from individual simulations are large, multi-dimensional, and must integrate with both simulation software and external data sites. Here we present the dimensional design and relational implementation of a comprehensive data warehouse for storing and analyzing molecular dynamics simulations using SQL Server. PMID:23204646
1993-07-09
Calculate Oil and solve iteratively equation (18) for q and (l)-(S) forex . 4, Solve the velocity problemn through equation (19) to calculate q and (6)-(10) to...object.oriented models for the database to store the system information f1l. Using OOP on the formalism level is more difficult and a current field of...Multidimensional Physical Systems: Graph-theoretic Modeling, Systems and Cybernetics, vol 21 (1992), 5 .9-71 JV A RELATIONAL DATABASE FOR GENERAL
A review of materials for spectral design coatings in signature management applications
NASA Astrophysics Data System (ADS)
Andersson, Kent E.; Škerlind, Christina
2014-10-01
The current focus in Swedish policy towards national security and high-end technical systems, together with a rapid development in multispectral sensor technology, adds to the utility of developing advanced materials for spectral design in signature management applications. A literature study was performed probing research databases for advancements. Qualitative text analysis was performed using a six-indicator instrument: spectrally selective reflectance; low gloss; low degree of polarization; low infrared emissivity; non-destructive properties in radar and in general controllability of optical properties. Trends are identified and the most interesting materials and coating designs are presented with relevant performance metrics. They are sorted into categories in the order of increasing complexity: pigments and paints, one-dimensional structures, multidimensional structures (including photonic crystals), and lastly biomimic and metamaterials. The military utility of the coatings is assessed qualitatively. The need for developing a framework for assessing the military utility of incrementally increasing the performance of spectrally selective coatings is identified.
Epistemological and Reading Beliefs Profiles and Their Role in Multiple Text Comprehension
ERIC Educational Resources Information Center
Mateos, Mar; Solé, Isabel; Martín, Elena; Castells, Nuria; Cuevas, Isabel; González-Lamas, Jara
2016-01-01
Introduction: The aim of this study was to analyse the role of epistemological beliefs and reading beliefs in the comprehension of multiple texts which presented conflicting positions about a controversial topic (nuclear energy). More specifically, we investigated the influence of the multidimensional configuration of epistemological and reading…
Comparing NetCDF and SciDB on managing and querying 5D hydrologic dataset
NASA Astrophysics Data System (ADS)
Liu, Haicheng; Xiao, Xiao
2016-11-01
Efficiently extracting information from high dimensional hydro-meteorological modelling datasets requires smart solutions. Traditional methods are mostly based on files, which can be edited and accessed handily. But they have problems of efficiency due to contiguous storage structure. Others propose databases as an alternative for advantages such as native functionalities for manipulating multidimensional (MD) arrays, smart caching strategy and scalability. In this research, NetCDF file based solutions and the multidimensional array database management system (DBMS) SciDB applying chunked storage structure are benchmarked to determine the best solution for storing and querying 5D large hydrologic modelling dataset. The effect of data storage configurations including chunk size, dimension order and compression on query performance is explored. Results indicate that dimension order to organize storage of 5D data has significant influence on query performance if chunk size is very large. But the effect becomes insignificant when chunk size is properly set. Compression of SciDB mostly has negative influence on query performance. Caching is an advantage but may be influenced by execution of different query processes. On the whole, NetCDF solution without compression is in general more efficient than the SciDB DBMS.
A Multidimensional Investigation of Deep-Level and Surface-Level Processing
ERIC Educational Resources Information Center
Dinsmore, Daniel L.; Alexander, Patricia A.
2016-01-01
This study examines the moderating effects of a situational factor (i.e., text type) and an individual factor (i.e., subject-matter knowledge) on the relation between depth of processing and performance. One-hundred and fifty-one undergraduates completed measures of subject-matter knowledge, read either an expository or persuasive text about the…
SciDB versus Spark: A Preliminary Comparison Based on an Earth Science Use Case
NASA Astrophysics Data System (ADS)
Clune, T.; Kuo, K. S.; Doan, K.; Oloso, A.
2015-12-01
We compare two Big Data technologies, SciDB and Spark, for performance, usability, and extensibility, when applied to a representative Earth science use case. SciDB is a new-generation parallel distributed database management system (DBMS) based on the array data model that is capable of handling multidimensional arrays efficiently but requires lengthy data ingest prior to analysis, whereas Spark is a fast and general engine for large scale data processing that can immediately process raw data files and thereby avoid the ingest process. Once data have been ingested, SciDB is very efficient in database operations such as subsetting. Spark, on the other hand, provides greater flexibility by supporting a wide variety of high-level tools including DBMS's. For the performance aspect of this preliminary comparison, we configure Spark to operate directly on text or binary data files and thereby limit the need for additional tools. Arguably, a more appropriate comparison would involve exploring other configurations of Spark which exploit supported high-level tools, but that is beyond our current resources. To make the comparison as "fair" as possible, we export the arrays produced by SciDB into text files (or converting them to binary files) for the intake by Spark and thereby avoid any additional file processing penalties. The Earth science use case selected for this comparison is the identification and tracking of snowstorms in the NASA Modern Era Retrospective-analysis for Research and Applications (MERRA) reanalysis data. The identification portion of the use case is to flag all grid cells of the MERRA high-resolution hourly data that satisfies our criteria for snowstorm, whereas the tracking portion connects flagged cells adjacent in time and space to form a snowstorm episode. We will report the results of our comparisons at this presentation.
Associative memory model for searching an image database by image snippet
NASA Astrophysics Data System (ADS)
Khan, Javed I.; Yun, David Y.
1994-09-01
This paper presents an associative memory called an multidimensional holographic associative computing (MHAC), which can be potentially used to perform feature based image database query using image snippet. MHAC has the unique capability to selectively focus on specific segments of a query frame during associative retrieval. As a result, this model can perform search on the basis of featural significance described by a subset of the snippet pixels. This capability is critical for visual query in image database because quite often the cognitive index features in the snippet are statistically weak. Unlike, the conventional artificial associative memories, MHAC uses a two level representation and incorporates additional meta-knowledge about the reliability status of segments of information it receives and forwards. In this paper we present the analysis of focus characteristics of MHAC.
NASA Astrophysics Data System (ADS)
Verma, Surendra P.; Rivera-Gómez, M. Abdelaly; Díaz-González, Lorena; Quiroz-Ruiz, Alfredo
2016-12-01
A new multidimensional classification scheme consistent with the chemical classification of the International Union of Geological Sciences (IUGS) is proposed for the nomenclature of High-Mg altered rocks. Our procedure is based on an extensive database of major element (SiO2, TiO2, Al2O3, Fe2O3t, MnO, MgO, CaO, Na2O, K2O, and P2O5) compositions of a total of 33,868 (920 High-Mg and 32,948 "Common") relatively fresh igneous rock samples. The database consisting of these multinormally distributed samples in terms of their isometric log-ratios was used to propose a set of 11 discriminant functions and 6 diagrams to facilitate High-Mg rock classification. The multinormality required by linear discriminant and canonical analysis was ascertained by a new computer program DOMuDaF. One multidimensional function can distinguish the High-Mg and Common igneous rocks with high percent success values of about 86.4% and 98.9%, respectively. Similarly, from 10 discriminant functions the High-Mg rocks can also be classified as one of the four rock types (komatiite, meimechite, picrite, and boninite), with high success values of about 88%-100%. Satisfactory functioning of this new classification scheme was confirmed by seven independent tests. Five further case studies involving application to highly altered rocks illustrate the usefulness of our proposal. A computer program HMgClaMSys was written to efficiently apply the proposed classification scheme, which will be available for online processing of igneous rock compositional data. Monte Carlo simulation modeling and mass-balance computations confirmed the robustness of our classification with respect to analytical errors and postemplacement compositional changes.
2009-01-01
Background Electronic guideline-based decision support systems have been suggested to successfully deliver the knowledge embedded in clinical practice guidelines. A number of studies have already shown positive findings for decision support systems such as drug-dosing systems and computer-generated reminder systems for preventive care services. Methods A systematic literature search (1990 to December 2008) of the English literature indexed in the Medline database, Embase, the Cochrane Central Register of Controlled Trials, and CRD (DARE, HTA and NHS EED databases) was conducted to identify evaluation studies of electronic multi-step guideline implementation systems in ambulatory care settings. Important inclusion criterions were the multidimensionality of the guideline (the guideline needed to consist of several aspects or steps) and real-time interaction with the system during consultation. Clinical decision support systems such as one-time reminders for preventive care for which positive findings were shown in earlier reviews were excluded. Two comparisons were considered: electronic multidimensional guidelines versus usual care (comparison one) and electronic multidimensional guidelines versus other guideline implementation methods (comparison two). Results Twenty-seven publications were selected for analysis in this systematic review. Most designs were cluster randomized controlled trials investigating process outcomes more than patient outcomes. With success defined as at least 50% of the outcome variables being significant, none of the studies were successful in improving patient outcomes. Only seven of seventeen studies that investigated process outcomes showed improvements in process of care variables compared with the usual care group (comparison one). No incremental effect of the electronic implementation over the distribution of paper versions of the guideline was found, neither for the patient outcomes nor for the process outcomes (comparison two). Conclusions There is little evidence at the moment for the effectiveness of an increasingly used and commercialised instrument such as electronic multidimensional guidelines. After more than a decade of development of numerous electronic systems, research on the most effective implementation strategy for this kind of guideline-based decision support systems is still lacking. This conclusion implies a considerable risk towards inappropriate investments in ineffective implementation interventions and in suboptimal care. PMID:20042070
Full-Text Databases in Medicine.
ERIC Educational Resources Information Center
Sievert, MaryEllen C.; And Others
1995-01-01
Describes types of full-text databases in medicine; discusses features for searching full-text journal databases available through online vendors; reviews research on full-text databases in medicine; and describes the MEDLINE/Full-Text Research Project at the University of Missouri (Columbia) which investigated precision, recall, and relevancy.…
Pain assessment scales in newborns: integrative review
de Melo, Gleicia Martins; Lélis, Ana Luíza Paula de Aguiar; de Moura, Alline Falconieri; Cardoso, Maria Vera Lúcia Moreira Leitão; da Silva, Viviane Martins
2014-01-01
OBJECTIVE: To analyze studies on methods used to assess pain in newborns. DATA SOURCES: Integrative review study of articles published from 2001 to 2012, carried out in the following databases: Scopus, PubMed, CINAHL, LILACS and Cochrane. The sample consisted of 13 articles with level of evidence 5. DATA SYNTHESIS: 29 pain assessment scales in newborns, including 13 one-dimensional and 16 multidimensional, that assess acute and prolonged pain in preterm and full-term infants were available in scientific publications. CONCLUSION: Based on the characteristics of scales, one cannot choose a single one as the most appropriate scale, as this choice will depend on gestational age, type of painful stimulus and the environment in which the infant is inserted. It is suggested the use of multidimensional or one-dimensional scales; however, they must be reliable and validated. PMID:25511005
Rosenthal, Victor D; Pawar, Mandakini; Leblebicioglu, Hakan; Navoa-Ng, Josephine Anne; Villamil-Gómez, Wilmer; Armas-Ruiz, Alberto; Cuéllar, Luis E; Medeiros, Eduardo A; Mitrev, Zan; Gikas, Achilleas; Yang, Yun; Ahmed, Altaf; Kanj, Souha S; Dueñas, Lourdes; Gurskis, Vaidotas; Mapp, Trudell; Guanche-Garcell, Humberto; Fernández-Hidalgo, Rosalía; Kübler, Andrzej
2013-04-01
To assess the feasibility and effectiveness of the International Nosocomial Infection Control Consortium (INICC) multidimensional hand hygiene approach in 19 limited-resource countries and to analyze predictors of poor hand hygiene compliance. An observational, prospective, cohort, interventional, before-and-after study from April 1999 through December 2011. The study was divided into 2 periods: a 3-month baseline period and a 7-year follow-up period. Ninety-nine intensive care unit (ICU) members of the INICC in Argentina, Brazil, China, Colombia, Costa Rica, Cuba, El Salvador, Greece, India, Lebanon, Lithuania, Macedonia, Mexico, Pakistan, Panama, Peru, Philippines, Poland, and Turkey. Healthcare workers at 99 ICU members of the INICC. A multidimensional hand hygiene approach was used, including (1) administrative support, (2) supplies availability, (3) education and training, (4) reminders in the workplace, (5) process surveillance, and (6) performance feedback. Observations were made for hand hygiene compliance in each ICU, during randomly selected 30-minute periods. A total of 149,727 opportunities for hand hygiene were observed. Overall hand hygiene compliance increased from 48.3% to 71.4% ([Formula: see text]). Univariate analysis indicated that several variables were significantly associated with poor hand hygiene compliance, including males versus females (63% vs 70%; [Formula: see text]), physicians versus nurses (62% vs 72%; [Formula: see text]), and adult versus neonatal ICUs (67% vs 81%; [Formula: see text]), among others. Adherence to hand hygiene increased by 48% with the INICC approach. Specific programs directed to improve hand hygiene for variables found to be predictors of poor hand hygiene compliance should be implemented.
Lequerica, Anthony; Bushnik, Tamara; Wright, Jerry; Kolakowsky-Hayner, Stephanie A; Hammond, Flora M; Dijkers, Marcel P; Cantor, Joshua
2012-01-01
To investigate the psychometric properties of the Multidimensional Assessment of Fatigue (MAF) scale in a traumatic brain injury (TBI) sample. Prospective survey study. Community. One hundred sixty-seven individuals with TBI admitted for inpatient rehabilitation, enrolled into the TBI Model Systems national database, and followed up at either the first or second year postinjury. Not applicable. Multidimensional Assessment of Fatigue. The initial analysis, using items 1 to 14, which are based on a 10-point rating scale, found that only 1 item ("walking") misfit the overall construct of fatigue in this TBI population. However, this 10-point rating scale was found to have disordered thresholds. When ratings were collapsed into 4 response categories, all MAF items used to calculate the Global Fatigue Index formed a unidimensional scale. Findings generally support the unidimensionality of the MAF when used in a TBI population but call into question the use of a 10-point rating scale for items 1 to 14. Further study is needed to investigate the use of a 4-category rating scale across all items and the fit of the "walking" item for a measure of fatigue among individuals with TBI.
Annual Progress Report for July 1, 1981 through June 30, 1982,
1982-08-01
Online Search Service .....................93 14.5 Database Analyses ......................................... 0000093 14.6 Automatic Detection of...D. Dow, "Deformatio potentials of "uperlattices and Interfaces, L. at Vsaunm Sienc Ma Tchnolly. vol. 19, pp. $64-566, 1981. 4.17 3. D. Oberstar, No...cince, vol. 15, no. 3, pp. 311-320, Sept. 1981. 12.11 M. C. Loi, "Simulations among multidimensional Turing machines," Theoretilna Comanaz Sience (to
Agile Datacube Analytics (not just) for the Earth Sciences
NASA Astrophysics Data System (ADS)
Misev, Dimitar; Merticariu, Vlad; Baumann, Peter
2017-04-01
Metadata are considered small, smart, and queryable; data, on the other hand, are known as big, clumsy, hard to analyze. Consequently, gridded data - such as images, image timeseries, and climate datacubes - are managed separately from the metadata, and with different, restricted retrieval capabilities. One reason for this silo approach is that databases, while good at tables, XML hierarchies, RDF graphs, etc., traditionally do not support multi-dimensional arrays well. This gap is being closed by Array Databases which extend the SQL paradigm of "any query, anytime" to NoSQL arrays. They introduce semantically rich modelling combined with declarative, high-level query languages on n-D arrays. On Server side, such queries can be optimized, parallelized, and distributed based on partitioned array storage. This way, they offer new vistas in flexibility, scalability, performance, and data integration. In this respect, the forthcoming ISO SQL extension MDA ("Multi-dimensional Arrays") will be a game changer in Big Data Analytics. We introduce concepts and opportunities through the example of rasdaman ("raster data manager") which in fact has pioneered the field of Array Databases and forms the blueprint for ISO SQL/MDA and further Big Data standards, such as OGC WCPS for querying spatio-temporal Earth datacubes. With operational installations exceeding 140 TB queries have been split across more than one thousand cloud nodes, using CPUs as well as GPUs. Installations can easily be mashed up securely, enabling large-scale location-transparent query processing in federations. Federation queries have been demonstrated live at EGU 2016 spanning Europe and Australia in the context of the intercontinental EarthServer initiative, visualized through NASA WorldWind.
Agile Datacube Analytics (not just) for the Earth Sciences
NASA Astrophysics Data System (ADS)
Baumann, P.
2016-12-01
Metadata are considered small, smart, and queryable; data, on the other hand, are known as big, clumsy, hard to analyze. Consequently, gridded data - such as images, image timeseries, and climate datacubes - are managed separately from the metadata, and with different, restricted retrieval capabilities. One reason for this silo approach is that databases, while good at tables, XML hierarchies, RDF graphs, etc., traditionally do not support multi-dimensional arrays well.This gap is being closed by Array Databases which extend the SQL paradigm of "any query, anytime" to NoSQL arrays. They introduce semantically rich modelling combined with declarative, high-level query languages on n-D arrays. On Server side, such queries can be optimized, parallelized, and distributed based on partitioned array storage. This way, they offer new vistas in flexibility, scalability, performance, and data integration. In this respect, the forthcoming ISO SQL extension MDA ("Multi-dimensional Arrays") will be a game changer in Big Data Analytics.We introduce concepts and opportunities through the example of rasdaman ("raster data manager") which in fact has pioneered the field of Array Databases and forms the blueprint for ISO SQL/MDA and further Big Data standards, such as OGC WCPS for querying spatio-temporal Earth datacubes. With operational installations exceeding 140 TB queries have been split across more than one thousand cloud nodes, using CPUs as well as GPUs. Installations can easily be mashed up securely, enabling large-scale location-transparent query processing in federations. Federation queries have been demonstrated live at EGU 2016 spanning Europe and Australia in the context of the intercontinental EarthServer initiative, visualized through NASA WorldWind.
Austvoll-Dahlgren, Astrid; Guttersrud, Øystein; Nsangi, Allen; Semakula, Daniel; Oxman, Andrew D
2017-05-25
The Claim Evaluation Tools database contains multiple-choice items for measuring people's ability to apply the key concepts they need to know to be able to assess treatment claims. We assessed items from the database using Rasch analysis to develop an outcome measure to be used in two randomised trials in Uganda. Rasch analysis is a form of psychometric testing relying on Item Response Theory. It is a dynamic way of developing outcome measures that are valid and reliable. To assess the validity, reliability and responsiveness of 88 items addressing 22 key concepts using Rasch analysis. We administrated four sets of multiple-choice items in English to 1114 people in Uganda and Norway, of which 685 were children and 429 were adults (including 171 health professionals). We scored all items dichotomously. We explored summary and individual fit statistics using the RUMM2030 analysis package. We used SPSS to perform distractor analysis. Most items conformed well to the Rasch model, but some items needed revision. Overall, the four item sets had satisfactory reliability. We did not identify significant response dependence between any pairs of items and, overall, the magnitude of multidimensionality in the data was acceptable. The items had a high level of difficulty. Most of the items conformed well to the Rasch model's expectations. Following revision of some items, we concluded that most of the items were suitable for use in an outcome measure for evaluating the ability of children or adults to assess treatment claims. © Article author(s) (or their employer(s) unless otherwise stated in the text of the article) 2017. All rights reserved. No commercial use is permitted unless otherwise expressly granted.
High fold computer disk storage DATABASE for fast extended analysis of γ-rays events
NASA Astrophysics Data System (ADS)
Stézowski, O.; Finck, Ch.; Prévost, D.
1999-03-01
Recently spectacular technical developments have been achieved to increase the resolving power of large γ-ray spectrometers. With these new eyes, physicists are able to study the intricate nature of atomic nuclei. Concurrently more and more complex multidimensional analyses are needed to investigate very weak phenomena. In this article, we first present a software (DATABASE) allowing high fold coincidences γ-rays events to be stored on hard disk. Then, a non-conventional method of analysis, anti-gating procedure, is described. Two physical examples are given to explain how it can be used and Monte Carlo simulations have been performed to test the validity of this method.
NASA Astrophysics Data System (ADS)
Yin, Lucy; Andrews, Jennifer; Heaton, Thomas
2018-05-01
Earthquake parameter estimations using nearest neighbor searching among a large database of observations can lead to reliable prediction results. However, in the real-time application of Earthquake Early Warning (EEW) systems, the accurate prediction using a large database is penalized by a significant delay in the processing time. We propose to use a multidimensional binary search tree (KD tree) data structure to organize large seismic databases to reduce the processing time in nearest neighbor search for predictions. We evaluated the performance of KD tree on the Gutenberg Algorithm, a database-searching algorithm for EEW. We constructed an offline test to predict peak ground motions using a database with feature sets of waveform filter-bank characteristics, and compare the results with the observed seismic parameters. We concluded that large database provides more accurate predictions of the ground motion information, such as peak ground acceleration, velocity, and displacement (PGA, PGV, PGD), than source parameters, such as hypocenter distance. Application of the KD tree search to organize the database reduced the average searching process by 85% time cost of the exhaustive method, allowing the method to be feasible for real-time implementation. The algorithm is straightforward and the results will reduce the overall time of warning delivery for EEW.
Acoustic analysis of normal Saudi adult voices.
Malki, Khalid H; Al-Habib, Salman F; Hagr, Abulrahman A; Farahat, Mohamed M
2009-08-01
To determine the acoustic differences between Saudi adult male and female voices, and to compare the acoustic variables of the Multidimensional Voice Program (MDVP) obtained from North American adults to a group of Saudi males and females. A cross-sectional survey of normal adult male and female voices was conducted at King Abdulaziz University Hospital, Riyadh, Kingdom of Saudi Arabia between March 2007 and December 2008. Ninety-five Saudi subjects sustained the vowel /a/ 6 times, and the steady state portion of 3 samples was analyzed and compared with the samples of the KayPentax normative voice database. Significant differences were found between Saudi and North American KayPentax database groups. In the male subjects, 15 of 33 MDVP variables, and 10 of 33 variables in the female subjects were found to be significantly different from the KayPentax database. We conclude that the acoustical differences may reflect laryngeal anatomical or tissue differences between the Saudi and the KayPentax database.
Chemical Space: Big Data Challenge for Molecular Diversity.
Awale, Mahendra; Visini, Ricardo; Probst, Daniel; Arús-Pous, Josep; Reymond, Jean-Louis
2017-10-25
Chemical space describes all possible molecules as well as multi-dimensional conceptual spaces representing the structural diversity of these molecules. Part of this chemical space is available in public databases ranging from thousands to billions of compounds. Exploiting these databases for drug discovery represents a typical big data problem limited by computational power, data storage and data access capacity. Here we review recent developments of our laboratory, including progress in the chemical universe databases (GDB) and the fragment subset FDB-17, tools for ligand-based virtual screening by nearest neighbor searches, such as our multi-fingerprint browser for the ZINC database to select purchasable screening compounds, and their application to discover potent and selective inhibitors for calcium channel TRPV6 and Aurora A kinase, the polypharmacology browser (PPB) for predicting off-target effects, and finally interactive 3D-chemical space visualization using our online tools WebDrugCS and WebMolCS. All resources described in this paper are available for public use at www.gdb.unibe.ch.
Emotion Recognition from EEG Signals Using Multidimensional Information in EMD Domain.
Zhuang, Ning; Zeng, Ying; Tong, Li; Zhang, Chi; Zhang, Hanming; Yan, Bin
2017-01-01
This paper introduces a method for feature extraction and emotion recognition based on empirical mode decomposition (EMD). By using EMD, EEG signals are decomposed into Intrinsic Mode Functions (IMFs) automatically. Multidimensional information of IMF is utilized as features, the first difference of time series, the first difference of phase, and the normalized energy. The performance of the proposed method is verified on a publicly available emotional database. The results show that the three features are effective for emotion recognition. The role of each IMF is inquired and we find that high frequency component IMF1 has significant effect on different emotional states detection. The informative electrodes based on EMD strategy are analyzed. In addition, the classification accuracy of the proposed method is compared with several classical techniques, including fractal dimension (FD), sample entropy, differential entropy, and discrete wavelet transform (DWT). Experiment results on DEAP datasets demonstrate that our method can improve emotion recognition performance.
Calibration of the Test of Relational Reasoning.
Dumas, Denis; Alexander, Patricia A
2016-10-01
Relational reasoning, or the ability to discern meaningful patterns within a stream of information, is a critical cognitive ability associated with academic and professional success. Importantly, relational reasoning has been described as taking multiple forms, depending on the type of higher order relations being drawn between and among concepts. However, the reliable and valid measurement of such a multidimensional construct of relational reasoning has been elusive. The Test of Relational Reasoning (TORR) was designed to tap 4 forms of relational reasoning (i.e., analogy, anomaly, antinomy, and antithesis). In this investigation, the TORR was calibrated and scored using multidimensional item response theory in a large, representative undergraduate sample. The bifactor model was identified as the best-fitting model, and used to estimate item parameters and construct reliability. To improve the usefulness of the TORR to educators, scaled scores were also calculated and presented. (PsycINFO Database Record (c) 2016 APA, all rights reserved).
Borchani, Hanen; Bielza, Concha; Toro, Carlos; Larrañaga, Pedro
2013-03-01
Our aim is to use multi-dimensional Bayesian network classifiers in order to predict the human immunodeficiency virus type 1 (HIV-1) reverse transcriptase and protease inhibitors given an input set of respective resistance mutations that an HIV patient carries. Multi-dimensional Bayesian network classifiers (MBCs) are probabilistic graphical models especially designed to solve multi-dimensional classification problems, where each input instance in the data set has to be assigned simultaneously to multiple output class variables that are not necessarily binary. In this paper, we introduce a new method, named MB-MBC, for learning MBCs from data by determining the Markov blanket around each class variable using the HITON algorithm. Our method is applied to both reverse transcriptase and protease data sets obtained from the Stanford HIV-1 database. Regarding the prediction of antiretroviral combination therapies, the experimental study shows promising results in terms of classification accuracy compared with state-of-the-art MBC learning algorithms. For reverse transcriptase inhibitors, we get 71% and 11% in mean and global accuracy, respectively; while for protease inhibitors, we get more than 84% and 31% in mean and global accuracy, respectively. In addition, the analysis of MBC graphical structures lets us gain insight into both known and novel interactions between reverse transcriptase and protease inhibitors and their respective resistance mutations. MB-MBC algorithm is a valuable tool to analyze the HIV-1 reverse transcriptase and protease inhibitors prediction problem and to discover interactions within and between these two classes of inhibitors. Copyright © 2012 Elsevier B.V. All rights reserved.
An Evaluation of Database Solutions to Spatial Object Association
DOE Office of Scientific and Technical Information (OSTI.GOV)
Kumar, V S; Kurc, T; Saltz, J
2008-06-24
Object association is a common problem encountered in many applications. Spatial object association, also referred to as crossmatch of spatial datasets, is the problem of identifying and comparing objects in two datasets based on their positions in a common spatial coordinate system--one of the datasets may correspond to a catalog of objects observed over time in a multi-dimensional domain; the other dataset may consist of objects observed in a snapshot of the domain at a time point. The use of database management systems to the solve the object association problem provides portability across different platforms and also greater flexibility. Increasingmore » dataset sizes in today's applications, however, have made object association a data/compute-intensive problem that requires targeted optimizations for efficient execution. In this work, we investigate how database-based crossmatch algorithms can be deployed on different database system architectures and evaluate the deployments to understand the impact of architectural choices on crossmatch performance and associated trade-offs. We investigate the execution of two crossmatch algorithms on (1) a parallel database system with active disk style processing capabilities, (2) a high-throughput network database (MySQL Cluster), and (3) shared-nothing databases with replication. We have conducted our study in the context of a large-scale astronomy application with real use-case scenarios.« less
Coyne, Sarah M; Padilla-Walker, Laura M; Holmgren, Hailey G; Davis, Emilie J; Collier, Kevin M; Memmott-Elison, Madison K; Hawkins, Alan J
2018-02-01
Studies examining the effects of exposure to prosocial media on positive outcomes are increasing in number and strength. However, existing meta-analyses use a broad definition of prosocial media that does not recognize the multidimensionality of prosocial behavior. The aim of the current study is to conduct a meta-analysis on the effects of exposure to prosocial media on prosocial behavior, aggression, and empathic concern while examining multiple moderators that the prosocial behavior literature suggests are important to our understanding of why individuals voluntarily help others (e.g., target, type, cost). Results from 72 studies involving 243 effect sizes revealed that exposure to prosocial media was related to higher levels of prosocial behavior and empathic concern and lower levels of aggressive behavior. Moderation analyses suggest that several moderators accounted for heterogeneity in the model, including age of participant, region, media type (active vs. passive), and study design. In terms of multidimensional moderators, prosocial media had stronger effects on prosocial behavior toward strangers than did any other target and on helping and prosocial thinking but not donating or volunteering. Comparisons with other meta-analyses on media effects are made and implications for parents, media producers, and researchers are discussed. (PsycINFO Database Record (c) 2018 APA, all rights reserved).
Horn’s Curve Estimation Through Multi-Dimensional Interpolation
2013-03-01
complex nature of human behavior has not yet been broached. This is not to say analysts play favorites in reaching conclusions, only that varied...Chapter III, Section 3.7. For now, it is sufficient to say underdetermined data presents technical challenges and all such datasets will be excluded from...database lookup table and then use the method of linear interpolation to instantaneously estimate the unknown points on an as-needed basis ( say from a user
Academic Journal Embargoes and Full Text Databases.
ERIC Educational Resources Information Center
Brooks, Sam
2003-01-01
Documents the reasons for embargoes of academic journals in full text databases (i.e., publisher-imposed delays on the availability of full text content) and provides insight regarding common misconceptions. Tables present data on selected journals covering a cross-section of subjects and publishers and comparing two full text business databases.…
Full Text Psychology Journals Available from Popular Library Databases
ERIC Educational Resources Information Center
Joswick, Kathleen E.
2006-01-01
The author identified 433 core journals in psychology and investigated their full text availability in popular databases. While 62 percent of the studied journals were available in at least one database, access from individual databases ranged from 1.4 percent to 38.1 percent of the titles. The full text of influential psychology journals is not…
Visualizing the semantic content of large text databases using text maps
NASA Technical Reports Server (NTRS)
Combs, Nathan
1993-01-01
A methodology for generating text map representations of the semantic content of text databases is presented. Text maps provide a graphical metaphor for conceptualizing and visualizing the contents and data interrelationships of large text databases. Described are a set of experiments conducted against the TIPSTER corpora of Wall Street Journal articles. These experiments provide an introduction to current work in the representation and visualization of documents by way of their semantic content.
Data Visualization in Information Retrieval and Data Mining (SIG VIS).
ERIC Educational Resources Information Center
Efthimiadis, Efthimis
2000-01-01
Presents abstracts that discuss using data visualization for information retrieval and data mining, including immersive information space and spatial metaphors; spatial data using multi-dimensional matrices with maps; TREC (Text Retrieval Conference) experiments; users' information needs in cartographic information retrieval; and users' relevance…
The collation of forensic DNA case data into a multi-dimensional intelligence database.
Walsh, S J; Moss, D S; Kliem, C; Vintiner, G M
2002-01-01
The primary aim of any DNA Database is to link individuals to unsolved offenses and unsolved offenses to each other via DNA profiling. This aim has been successfully realised during the operation of the New Zealand (NZ) DNA Databank over the past five years. The DNA Intelligence Project (DIP), a collaborative project involving NZ forensic and law enforcement agencies, interrogated the forensic case data held on the NZ DNA databank and collated it into a functional intelligence database. This database has been used to identify significant trends which direct Police and forensic personnel towards the most appropriate use of DNA technology. Intelligence is being provided in areas such as the level of usage of DNA techniques in criminal investigation, the relative success of crime scene samples and the geographical distribution of crimes. The DIP has broadened the dimensions of the information offered through the NZ DNA Databank and has furthered the understanding and investigative capability of both Police and forensic scientists. The outcomes of this research fit soundly with the current policies of 'intelligence led policing', which are being adopted by Police jurisdictions locally and overseas.
Modeling Powered Aerodynamics for the Orion Launch Abort Vehicle Aerodynamic Database
NASA Technical Reports Server (NTRS)
Chan, David T.; Walker, Eric L.; Robinson, Philip E.; Wilson, Thomas M.
2011-01-01
Modeling the aerodynamics of the Orion Launch Abort Vehicle (LAV) has presented many technical challenges to the developers of the Orion aerodynamic database. During a launch abort event, the aerodynamic environment around the LAV is very complex as multiple solid rocket plumes interact with each other and the vehicle. It is further complicated by vehicle separation events such as between the LAV and the launch vehicle stack or between the launch abort tower and the crew module. The aerodynamic database for the LAV was developed mainly from wind tunnel tests involving powered jet simulations of the rocket exhaust plumes, supported by computational fluid dynamic simulations. However, limitations in both methods have made it difficult to properly capture the aerodynamics of the LAV in experimental and numerical simulations. These limitations have also influenced decisions regarding the modeling and structure of the aerodynamic database for the LAV and led to compromises and creative solutions. Two database modeling approaches are presented in this paper (incremental aerodynamics and total aerodynamics), with examples showing strengths and weaknesses of each approach. In addition, the unique problems presented to the database developers by the large data space required for modeling a launch abort event illustrate the complexities of working with multi-dimensional data.
Perspectives in astrophysical databases
NASA Astrophysics Data System (ADS)
Frailis, Marco; de Angelis, Alessandro; Roberto, Vito
2004-07-01
Astrophysics has become a domain extremely rich of scientific data. Data mining tools are needed for information extraction from such large data sets. This asks for an approach to data management emphasizing the efficiency and simplicity of data access; efficiency is obtained using multidimensional access methods and simplicity is achieved by properly handling metadata. Moreover, clustering and classification techniques on large data sets pose additional requirements in terms of computation and memory scalability and interpretability of results. In this study we review some possible solutions.
Empirical Validation of Reading Proficiency Guidelines
ERIC Educational Resources Information Center
Clifford, Ray; Cox, Troy L.
2013-01-01
The validation of ability scales describing multidimensional skills is always challenging, but not impossible. This study applies a multistage, criterion-referenced approach that uses a framework of aligned texts and reading tasks to explore the validity of the ACTFL and related reading proficiency guidelines. Rasch measurement and statistical…
Midtgaard, Julie; Stelter, Reinhard; Rørth, Mikael; Adamsen, Lis
2007-04-01
Evidence is emerging that exercise can reduce psychological distress in cancer patients undergoing treatment. The present study aimed to (qualitatively) explore the experiences of advanced disease cancer patients participating in a 6-week, 9-hours weekly, structured, group-based multidimensional exercise intervention while undergoing chemotherapy. Unstructured diaries from a purposive sample of three females and two males (28-52 years old) who participated in the program served as the database. Data were analyzed using a phenomenological, narrative method. The analysis yielded three themes: shifting position, self-surveillance, and negotiated strength. The intervention highlighted situations making it possible for the participants to negate psychological and physical constraints. The concept of structured exercise contains viable psychotherapeutic potentials by allowing the development of alternative bodily and mental realities complying with cancer patients' demands and abilities to regain autonomy and commitment to discover and adopt a sense of agency and shared self-reliance.
ERIC Educational Resources Information Center
Cotton, P. L.
1987-01-01
Defines two types of online databases: source, referring to those intended to be complete in themselves, whether full-text or abstracts; and bibliographic, meaning those that are not complete. Predictions are made about the future growth rate of these two types of databases, as well as full-text versus abstract databases. (EM)
Profiling Oman education data using data visualization technique
NASA Astrophysics Data System (ADS)
Alalawi, Sultan Juma Sultan; Shaharanee, Izwan Nizal Mohd; Jamil, Jastini Mohd
2016-10-01
This research works presents an innovative data visualization technique to understand and visualize the information of Oman's education data generated from the Ministry of Education Oman "Educational Portal". The Ministry of Education in Sultanate of Oman have huge databases contains massive information. The volume of data in the database increase yearly as many students, teachers and employees enter into the database. The task for discovering and analyzing these vast volumes of data becomes increasingly difficult. Information visualization and data mining offer a better ways in dealing with large volume of information. In this paper, an innovative information visualization technique is developed to visualize the complex multidimensional educational data. Microsoft Excel Dashboard, Visual Basic Application (VBA) and Pivot Table are utilized to visualize the data. Findings from the summarization of the data are presented, and it is argued that information visualization can help related stakeholders to become aware of hidden and interesting information from large amount of data drowning in their educational portal.
Datacube Services in Action, Using Open Source and Open Standards
NASA Astrophysics Data System (ADS)
Baumann, P.; Misev, D.
2016-12-01
Array Databases comprise novel, promising technology for massive spatio-temporal datacubes, extending the SQL paradigm of "any query, anytime" to n-D arrays. On server side, such queries can be optimized, parallelized, and distributed based on partitioned array storage. The rasdaman ("raster data manager") system, which has pioneered Array Databases, is available in open source on www.rasdaman.org. Its declarative query language extends SQL with array operators which are optimized and parallelized on server side. The rasdaman engine, which is part of OSGeo Live, is mature and in operational use databases individually holding dozens of Terabytes. Further, the rasdaman concepts have strongly impacted international Big Data standards in the field, including the forthcoming MDA ("Multi-Dimensional Array") extension to ISO SQL, the OGC Web Coverage Service (WCS) and Web Coverage Processing Service (WCPS) standards, and the forthcoming INSPIRE WCS/WCPS; in both OGC and INSPIRE, OGC is WCS Core Reference Implementation. In our talk we present concepts, architecture, operational services, and standardization impact of open-source rasdaman, as well as experiences made.
Horst, Jessica S; Hout, Michael C
2016-12-01
Many experimental research designs require images of novel objects. Here we introduce the Novel Object and Unusual Name (NOUN) Database. This database contains 64 primary novel object images and additional novel exemplars for ten basic- and nine global-level object categories. The objects' novelty was confirmed by both self-report and a lack of consensus on questions that required participants to name and identify the objects. We also found that object novelty correlated with qualifying naming responses pertaining to the objects' colors. The results from a similarity sorting task (and a subsequent multidimensional scaling analysis on the similarity ratings) demonstrated that the objects are complex and distinct entities that vary along several featural dimensions beyond simply shape and color. A final experiment confirmed that additional item exemplars comprised both sub- and superordinate categories. These images may be useful in a variety of settings, particularly for developmental psychology and other research in the language, categorization, perception, visual memory, and related domains.
Ge, Lixia; Mordiffi, Siti Zubaidah
Caring for elderly cancer patients may cause multidimensional burden on family caregivers. Recognition of factors associated with caregiver burden is important for providing proactive support to caregivers at risk. The aim of this study was to identify factors associated with high caregiver burden among family caregivers of elderly cancer patients. A systematic search of 7 electronic databases was conducted from database inception to October 2014. The identified studies were screened, and full text was further assessed. The quality of included studies was assessed using a checklist, and relevant data were extracted using a predeveloped data extraction form. Best-evidence synthesis model was used for data synthesis. The search yielded a total of 3339 studies, and 7 studies involving 1233 family caregivers were included after screening and full assessment of 116 studies. Moderate evidence supported that younger caregivers, solid tumors, and assistance with patient's activities of daily living were significantly associated with high caregiver burden. Eighteen factors were supported by limited evidence, and 1 was a conflicting factor. The scientific literature to date proved that caregiver burden was commonly experienced by family caregivers of elderly cancer patients. The evidence indicated that family caregivers who were at younger age, caring for solid tumor patients, and providing assistance with patient's activities of daily living reported high caregiver burden. The data provide evidence in identifying family caregivers at high risk of high caregiver burden. More high-quality studies are needed to clarify and determine the estimates of the effects of individual factors.
Collaborative development for setup, execution, sharing and analytics of complex NMR experiments.
Irvine, Alistair G; Slynko, Vadim; Nikolaev, Yaroslav; Senthamarai, Russell R P; Pervushin, Konstantin
2014-02-01
Factory settings of NMR pulse sequences are rarely ideal for every scenario in which they are utilised. The optimisation of NMR experiments has for many years been performed locally, with implementations often specific to an individual spectrometer. Furthermore, these optimised experiments are normally retained solely for the use of an individual laboratory, spectrometer or even single user. Here we introduce a web-based service that provides a database for the deposition, annotation and optimisation of NMR experiments. The application uses a Wiki environment to enable the collaborative development of pulse sequences. It also provides a flexible mechanism to automatically generate NMR experiments from deposited sequences. Multidimensional NMR experiments of proteins and other macromolecules consume significant resources, in terms of both spectrometer time and effort required to analyse the results. Systematic analysis of simulated experiments can enable optimal allocation of NMR resources for structural analysis of proteins. Our web-based application (http://nmrplus.org) provides all the necessary information, includes the auxiliaries (waveforms, decoupling sequences etc.), for analysis of experiments by accurate numerical simulation of multidimensional NMR experiments. The online database of the NMR experiments, together with a systematic evaluation of their sensitivity, provides a framework for selection of the most efficient pulse sequences. The development of such a framework provides a basis for the collaborative optimisation of pulse sequences by the NMR community, with the benefits of this collective effort being available to the whole community. Copyright © 2013 Elsevier Inc. All rights reserved.
rasdaman Array Database: current status
NASA Astrophysics Data System (ADS)
Merticariu, George; Toader, Alexandru
2015-04-01
rasdaman (Raster Data Manager) is a Free Open Source Array Database Management System which provides functionality for storing and processing massive amounts of raster data in the form of multidimensional arrays. The user can access, process and delete the data using SQL. The key features of rasdaman are: flexibility (datasets of any dimensionality can be processed with the help of SQL queries), scalability (rasdaman's distributed architecture enables it to seamlessly run on cloud infrastructures while offering an increase in performance with the increase of computation resources), performance (real-time access, processing, mixing and filtering of arrays of any dimensionality) and reliability (legacy communication protocol replaced with a new one based on cutting edge technology - Google Protocol Buffers and ZeroMQ). Among the data with which the system works, we can count 1D time series, 2D remote sensing imagery, 3D image time series, 3D geophysical data, and 4D atmospheric and climate data. Most of these representations cannot be stored only in the form of raw arrays, as the location information of the contents is also important for having a correct geoposition on Earth. This is defined by ISO 19123 as coverage data. rasdaman provides coverage data support through the Petascope service. Extensions were added on top of rasdaman in order to provide support for the Geoscience community. The following OGC standards are currently supported: Web Map Service (WMS), Web Coverage Service (WCS), and Web Coverage Processing Service (WCPS). The Web Map Service is an extension which provides zoom and pan navigation over images provided by a map server. Starting with version 9.1, rasdaman supports WMS version 1.3. The Web Coverage Service provides capabilities for downloading multi-dimensional coverage data. Support is also provided for several extensions of this service: Subsetting Extension, Scaling Extension, and, starting with version 9.1, Transaction Extension, which defines request types for inserting, updating and deleting coverages. A web client, designed for both novice and experienced users, is also available for the service and its extensions. The client offers an intuitive interface that allows users to work with multi-dimensional coverages by abstracting the specifics of the standard definitions of the requests. The Web Coverage Processing Service defines a language for on-the-fly processing and filtering multi-dimensional raster coverages. rasdaman exposes this service through the WCS processing extension. Demonstrations are provided online via the Earthlook website (earthlook.org) which presents use-cases from a wide variety of application domains, using the rasdaman system as processing engine.
Database citation in full text biomedical articles.
Kafkas, Şenay; Kim, Jee-Hyub; McEntyre, Johanna R
2013-01-01
Molecular biology and literature databases represent essential infrastructure for life science research. Effective integration of these data resources requires that there are structured cross-references at the level of individual articles and biological records. Here, we describe the current patterns of how database entries are cited in research articles, based on analysis of the full text Open Access articles available from Europe PMC. Focusing on citation of entries in the European Nucleotide Archive (ENA), UniProt and Protein Data Bank, Europe (PDBe), we demonstrate that text mining doubles the number of structured annotations of database record citations supplied in journal articles by publishers. Many thousands of new literature-database relationships are found by text mining, since these relationships are also not present in the set of articles cited by database records. We recommend that structured annotation of database records in articles is extended to other databases, such as ArrayExpress and Pfam, entries from which are also cited widely in the literature. The very high precision and high-throughput of this text-mining pipeline makes this activity possible both accurately and at low cost, which will allow the development of new integrated data services.
Database Citation in Full Text Biomedical Articles
Kafkas, Şenay; Kim, Jee-Hyub; McEntyre, Johanna R.
2013-01-01
Molecular biology and literature databases represent essential infrastructure for life science research. Effective integration of these data resources requires that there are structured cross-references at the level of individual articles and biological records. Here, we describe the current patterns of how database entries are cited in research articles, based on analysis of the full text Open Access articles available from Europe PMC. Focusing on citation of entries in the European Nucleotide Archive (ENA), UniProt and Protein Data Bank, Europe (PDBe), we demonstrate that text mining doubles the number of structured annotations of database record citations supplied in journal articles by publishers. Many thousands of new literature-database relationships are found by text mining, since these relationships are also not present in the set of articles cited by database records. We recommend that structured annotation of database records in articles is extended to other databases, such as ArrayExpress and Pfam, entries from which are also cited widely in the literature. The very high precision and high-throughput of this text-mining pipeline makes this activity possible both accurately and at low cost, which will allow the development of new integrated data services. PMID:23734176
Keyless Entry: Building a Text Database Using OCR Technology.
ERIC Educational Resources Information Center
Grotophorst, Clyde W.
1989-01-01
Discusses the use of optical character recognition (OCR) technology to produce an ASCII text database. A tutorial on digital scanning and OCR is provided, and a systems integration project which used the Calera CDP-3000XF scanner and text retrieval software to construct a database of dissertations at George Mason University is described. (four…
A Tutorial in Creating Web-Enabled Databases with Inmagic DB/TextWorks through ODBC.
ERIC Educational Resources Information Center
Breeding, Marshall
2000-01-01
Explains how to create Web-enabled databases. Highlights include Inmagic's DB/Text WebPublisher product called DB/TextWorks; ODBC (Open Database Connectivity) drivers; Perl programming language; HTML coding; Structured Query Language (SQL); Common Gateway Interface (CGI) programming; and examples of HTML pages and Perl scripts. (LRW)
Santosh, Paramala; Lievesley, Kate; Fiori, Federico; Singh, Jatinder
2017-06-21
Rett syndrome (RTT) is a pervasive neurodevelopmental disorder that presents with deficits in brain functioning leading to language and learning regression, characteristic hand stereotypies and developmental delay. Different mutations in the gene implicated in RTT- methyl-CpG-binding protein 2 ( MECP2 ) establishes RTT as a disorder with divergent symptomatology ranging from individuals with severe to milder phenotypes. A reliable and single multidimensional questionnaire is needed that can embrace all symptoms, and the relationships between them, and can map clinically meaningful data to symptomatology across the lifespan in patients with RTT. As part of the HealthTracker-based Tailored Rett Intervention and Assessment Longitudinal (TRIAL) database, the Rett Evaluation of Symptoms and Treatments (REST) Questionnaire will be able to marry with the physiological aspects of the disease obtained using wearable sensor technology, along with genetic and psychosocial data to stratify patients. Taken together, the web-based TRIAL database will empower clinicians and researchers with the confidence to delineate between different aspects of disorder symptomatology to streamline care pathways for individuals or for those patients entering clinical trials. This protocol describes the anticipated development of the REST questionnaire and the TRIAL database which links with the outcomes of the wearable sensor technology, and will serve as a barometer for longitudinal patient monitoring in patients with RTT. The US Food and Drug Administration Guidance for Patient-Reported Outcome Measures will be used as a template to inform the methodology of the study. It will follow an iterative framework that will include item/concept identification, item/concept elicitation in parent/carer-mediated focus groups, expert clinician feedback, web-based presentation of questionnaires, initial scale development, instrument refinement and instrument validation. The study has received favourable opinion from the National Health Service (NHS) Research Ethics Committee (REC): NHS Research Ethics Committee (REC)-London, Bromley Research Ethics Committee (reference: 15/LO/1772). © Article author(s) (or their employer(s) unless otherwise stated in the text of the article) 2017. All rights reserved. No commercial use is permitted unless otherwise expressly granted.
Testlet-Based Multidimensional Adaptive Testing.
Frey, Andreas; Seitz, Nicki-Nils; Brandt, Steffen
2016-01-01
Multidimensional adaptive testing (MAT) is a highly efficient method for the simultaneous measurement of several latent traits. Currently, no psychometrically sound approach is available for the use of MAT in testlet-based tests. Testlets are sets of items sharing a common stimulus such as a graph or a text. They are frequently used in large operational testing programs like TOEFL, PISA, PIRLS, or NAEP. To make MAT accessible for such testing programs, we present a novel combination of MAT with a multidimensional generalization of the random effects testlet model (MAT-MTIRT). MAT-MTIRT compared to non-adaptive testing is examined for several combinations of testlet effect variances (0.0, 0.5, 1.0, and 1.5) and testlet sizes (3, 6, and 9 items) with a simulation study considering three ability dimensions with simple loading structure. MAT-MTIRT outperformed non-adaptive testing regarding the measurement precision of the ability estimates. Further, the measurement precision decreased when testlet effect variances and testlet sizes increased. The suggested combination of the MTIRT model therefore provides a solution to the substantial problems of testlet-based tests while keeping the length of the test within an acceptable range.
Database Management Systems: New Homes for Migrating Bibliographic Records.
ERIC Educational Resources Information Center
Brooks, Terrence A.; Bierbaum, Esther G.
1987-01-01
Assesses bibliographic databases as part of visionary text systems such as hypertext and scholars' workstations. Downloading is discussed in terms of the capability to search records and to maintain unique bibliographic descriptions, and relational database management systems, file managers, and text databases are reviewed as possible hosts for…
Translation and Validation of the Multidimensional Dyspnea-12 Questionnaire.
Amado Diago, Carlos Antonio; Puente Maestu, Luis; Abascal Bolado, Beatriz; Agüero Calvo, Juan; Hernando Hernando, Mercedes; Puente Bats, Irene; Agüero Balbín, Ramón
2018-02-01
Dyspnea is a multidimensional symptom, but this multidimensionality is not considered in most dyspnea questionnaires. The Dyspnea-12 takes a multidimensional approach to the assessment of dyspnea, specifically the sensory and the affective response. The objective of this study was to translate into Spanish and validate the Dyspnea-12 questionnaire. The original English version of the Dyspnea-12 questionnaire was translated into Spanish and backtranslated to analyze its equivalence. Comprehension of the text was verified by analyzing the responses of 10 patients. Reliability and validation of the questionnaire were studied in an independent group of COPD patients attending the pulmonology clinics of Hospital Universitario Marqués de Valdecilla, diagnosed and categorized according to GOLD guidelines. The mean age of the group (n=51) was 65 years and mean FEV1 was 50%. All patients understood all questions of the translated version of Dyspnea-12. Internal consistency of the questionnaire was α=0.937 and intraclass correlation coefficient was=.969; P<.001. Statistically significant correlations were found with HADS (anxiety r=.608 and depression r=.615), mMRC dyspnea (r=.592), 6MWT (r=-0.445), FEV1 (r=-0.312), all dimensions of CRQ-SAS (dyspnea r=-0.626; fatigue r=-0.718; emotional function r=-0.663; mastery r=-0.740), CAT (r=0.669), and baseline dyspnea index (r=-0.615). Dyspnea-12 scores were 10.32 points higher in symptomatic GOLD groups (B and D) (P<.001). The Spanish version of Dyspnea-12 is a valid and reliable instrument to study the multidimensional nature of dyspnea. Copyright © 2017 SEPAR. Publicado por Elsevier España, S.L.U. All rights reserved.
The Most Frequent Metacognitive Strategies Used in Reading Comprehension among ESP Learners
ERIC Educational Resources Information Center
Khoshsima, Hooshang; Samani, Elham Amiri
2015-01-01
Reading strategies are plans for solving problems encountered during reading while learners are deeply engage with the text. So, comprehension is not a simple decoding of symbols, but a complex multidimensional process in which the leaner draws on previous schemata applying strategies consciously. In fact, metacognitive strategies are accessible…
NASA Astrophysics Data System (ADS)
Carkin, Susan
The broad goal of this study is to represent the linguistic variation of textbooks and lectures, the primary input for student learning---and sometimes the sole input in the large introductory classes which characterize General Education at many state universities. Computer techniques are used to analyze a corpus of textbooks and lectures from first-year university classes in macroeconomics and biology. These spoken and written variants are compared to each other as well as to benchmark texts from other multi-dimensional studies in order to examine their patterns, relations, and functions. A corpus consisting of 147,000 words was created from macroeconomics and biology lectures at a medium-large state university and from a set of nationally "best-selling" textbooks used in these same introductory survey courses. The corpus was analyzed using multi-dimensional methodology (Biber, 1988). The analysis consists of both empirical and qualitative phases. Quantitative analyses are undertaken on the linguistic features, their patterns of co-occurrence, and on the contextual elements of classrooms and textbooks. The contextual analysis is used to functionally interpret the statistical patterns of co-occurrence along five dimensions of textual variation, demonstrating patterns of difference and similarity with reference to text excerpts. Results of the analysis suggest that academic discourse is far from monolithic. Pedagogic discourse in introductory classes varies by modality and discipline, but not always in the directions expected. In the present study the most abstract texts were biology lectures---more abstract than written genres of academic prose and more abstract than introductory textbooks. Academic lectures in both disciplines, monologues which carry a heavy informational load, were extremely interactive, more like conversation than academic prose. A third finding suggests that introductory survey textbooks differ from those used in upper division classes by being relatively less marked for information density, abstraction, and non-overt argumentation. In addition to the findings mentioned here, numerous other relationships among the texts exhibit complex patterns of variation related to a number of situational variables. Pedagogical implications are discussed in relation to General Education courses, differing student populations, and the reading and listening demands which students encounter in large introductory classes in the university.
Understanding International GNC Hardware Trends
NASA Technical Reports Server (NTRS)
Greenbaum, Adam; Brady, Tye; Dennehy, Cornelius; Airey, Stephen P.; Roelke, Evan; Judd, Samuel Brady
2015-01-01
An industry-wide survey of guidance, navigation and control (GNC) sensors, namely star trackers, gyros, and sun sensors was undertaken in 2014, in which size, mass, power, and various performance metrics were recorded for each category. A multidimensional analysis was performed, looking at the spectrum of available sensors, with the intent of identifying gaps in the available capability range. Mission types that are not currently well served by the available components were discussed, as well as some missions that would be enabled by filling gaps in the component space. This paper continues that study, with a focus on reaction wheels and magnetometers, as well as with updates to the listings of star trackers, gyros, and sun sensors. Also discussed are a framework for making the database available to the community at large, and the continued maintenance of this database and the analysis of its contents.
NASA Astrophysics Data System (ADS)
Smith, David C.
2005-08-01
The "RAMANITA ©" method, for semi-quantitative chemical analysis of mineral solid-solutions by multidimensional calibration of Raman wavenumber shifts and mathematical calculation by simultaneous equations, is published here in detail in English for the first time. It was conceived by the present writer 20 years ago for binary and ternary pyroxene and garnet systems. The mathematical description was set out in 1989, but in an abstract in an obscure French special publication. Detailed "step-by-step" calibration of two garnet ternaries, followed by their linking, by M. Pinet and D.C. Smith in the early 1990s provided a hexary garnet database. Much later, using this garnet database, which forms part of his personal database called RAMANITA ©, the present writer began to develop the method by improving the terminology, automating the calculations, discussing problems and experimenting with different real chemical problems in archaeometry. Although this RAMANITA © method has been very briefly mentioned in two recent books, the necessary full mathematical explanation is given only here. The method will find application in any study which requires obtaining a non-destructive semi-quantitative chemical analysis from mineral solid solutions that cannot be analysed by any destructive analytical method, in particular for archaeological, geological or extraterrestrial research projects, e.g. from gemstones or other crystalline artworks of the cultural heritage (especially by Mobile Raman Microscopy (MRM)) in situ in museums or at archaeological sites, including under water for subaquatic archaeometry; from scientifically precious mineral microinclusions (such as garnet or pyroxene within diamond); from minerals in rocks analysed in situ on planetary bodies by a rover (especially "at distance" by telescopy). Recently some other workers have begun deducing chemical compositions from Raman wavenumber shifts in multivariate chemical space, but the philosophical approach is quite different.
Preparing College Students To Search Full-Text Databases: Is Instruction Necessary?
ERIC Educational Resources Information Center
Riley, Cheryl; Wales, Barbara
Full-text databases allow Central Missouri State University's clients to access some of the serials that libraries have had to cancel due to escalating subscription costs; EbscoHost, the subject of this study, is one such database. The database is available free to all Missouri residents. A survey was designed consisting of 21 questions intended…
ERIC Educational Resources Information Center
Bell, Steven J.
2003-01-01
Discusses full-text databases and whether existing aggregator databases are meeting user needs. Topics include the need for better search interfaces; concepts of quality research and information retrieval; information overload; full text in electronic journal collections versus aggregator databases; underrepresentation of certain disciplines; and…
NASA Technical Reports Server (NTRS)
Marshall, Jospeh R.; Morris, Allan T.
2007-01-01
Since 2003, AIAA's Computer Systems and Software Systems Technical Committees (TCs) have developed a database that aids technical committee management to map technical topics to their members. This Topics/Interest (T/I) database grew out of a collection of charts and spreadsheets maintained by the TCs. Since its inception, the tool has evolved into a multi-dimensional database whose dimensions include the importance, interest and expertise of TC members and whether or not a member and/or a TC is actively involved with the topic. In 2005, the database was expanded to include the TCs in AIAA s Information Systems Group and then expanded further to include all AIAA TCs. It was field tested at an AIAA Technical Activities Committee (TAC) Workshop in early 2006 through live access by over 80 users. Through the use of the topics database, TC and program committee (PC) members can accomplish relevant tasks such as: to identify topic experts (for Aerospace America articles or external contacts), to determine the interest of its members, to identify overlapping topics between diverse TCs and PCs, to guide new member drives and to reveal emerging topics. This paper will describe the origins, inception, initial development, field test and current version of the tool as well as elucidate the benefits and insights gained by using the database to aid the management of various TC functions. Suggestions will be provided to guide future development of the database for the purpose of providing dynamics and system level benefits to AIAA that currently do not exist in any technical organization.
NASA Astrophysics Data System (ADS)
Afonso, J. C.; Zlotnik, S.; Diez, P.
2015-12-01
We present a flexible, general and efficient approach for implementing thermodynamic phase equilibria information (in the form of sets of physical parameters) into geophysical and geodynamic studies. The approach is based on multi-dimensional decomposition methods, which transform the original multi-dimensional discrete information into a dimensional-separated representation. This representation has the property of increasing the number of coefficients to be stored linearly with the number of dimensions (opposite to a full multi-dimensional cube requiring exponential storage depending on the number of dimensions). Thus, the amount of information to be stored in memory during a numerical simulation or geophysical inversion is drastically reduced. Accordingly, the amount and resolution of the thermodynamic information that can be used in a simulation or inversion increases substantially. In addition, the method is independent of the actual software used to obtain the primary thermodynamic information, and therefore it can be used in conjunction with any thermodynamic modeling program and/or database. Also, the errors associated with the decomposition procedure are readily controlled by the user, depending on her/his actual needs (e.g. preliminary runs vs full resolution runs). We illustrate the benefits, generality and applicability of our approach with several examples of practical interest for both geodynamic modeling and geophysical inversion/modeling. Our results demonstrate that the proposed method is a competitive and attractive candidate for implementing thermodynamic constraints into a broad range of geophysical and geodynamic studies.
Word-level recognition of multifont Arabic text using a feature vector matching approach
NASA Astrophysics Data System (ADS)
Erlandson, Erik J.; Trenkle, John M.; Vogt, Robert C., III
1996-03-01
Many text recognition systems recognize text imagery at the character level and assemble words from the recognized characters. An alternative approach is to recognize text imagery at the word level, without analyzing individual characters. This approach avoids the problem of individual character segmentation, and can overcome local errors in character recognition. A word-level recognition system for machine-printed Arabic text has been implemented. Arabic is a script language, and is therefore difficult to segment at the character level. Character segmentation has been avoided by recognizing text imagery of complete words. The Arabic recognition system computes a vector of image-morphological features on a query word image. This vector is matched against a precomputed database of vectors from a lexicon of Arabic words. Vectors from the database with the highest match score are returned as hypotheses for the unknown image. Several feature vectors may be stored for each word in the database. Database feature vectors generated using multiple fonts and noise models allow the system to be tuned to its input stream. Used in conjunction with database pruning techniques, this Arabic recognition system has obtained promising word recognition rates on low-quality multifont text imagery.
Curran, Thomas; Hill, Andrew P
2017-12-28
From the 1980s onward, neoliberal governance in the United States, Canada, and the United Kingdom has emphasized competitive individualism and people have seemingly responded, in kind, by agitating to perfect themselves and their lifestyles. In this study, the authors examine whether cultural changes have coincided with an increase in multidimensional perfectionism in college students over the last 27 years. Their analyses are based on 164 samples and 41,641 American, Canadian, and British college students, who completed the Multidimensional Perfectionism Scale (Hewitt & Flett, 1991) between 1989 and 2016 (70.92% female, Mage = 20.66). Cross-temporal meta-analysis revealed that levels of self-oriented perfectionism, socially prescribed perfectionism, and other-oriented perfectionism have linearly increased. These trends remained when controlling for gender and between-country differences in perfectionism scores. Overall, in order of magnitude of the observed increase, the findings indicate that recent generations of young people perceive that others are more demanding of them, are more demanding of others, and are more demanding of themselves. (PsycINFO Database Record (c) 2017 APA, all rights reserved).
ERIC Educational Resources Information Center
Muijselaar, Marloes M. L.; Swart, Nicole M.; Steenbeek-Planting, Esther G.; Droop, Mienke; Verhoeven, Ludo; de Jong, Peter F.
2017-01-01
Many recent studies have aimed to demonstrate that specific types of reading comprehension depend on different underlying cognitive abilities. In these studies, it is often implicitly assumed that reading comprehension is a multidimensional construct. The general aim of this study was to examine the dimensionality of a large pool of reading…
ERIC Educational Resources Information Center
Smith, Karen
2018-01-01
Policy texts are representations of practice that both reflect and shape the world around them. There is, however, little higher education research that critically analyses the impact of higher education policy on educational developers and educational development practice. Extending methods from critical discourse analysis by combining textual…
ERIC Educational Resources Information Center
Jacob, Laura Beth
2012-01-01
Virtual world environments have evolved from object-oriented, text-based online games to complex three-dimensional immersive social spaces where the lines between reality and computer-generated begin to blur. Educators use virtual worlds to create engaging three-dimensional learning spaces for students, but the impact of virtual worlds in…
Biocuration workflows and text mining: overview of the BioCreative 2012 Workshop Track II.
Lu, Zhiyong; Hirschman, Lynette
2012-01-01
Manual curation of data from the biomedical literature is a rate-limiting factor for many expert curated databases. Despite the continuing advances in biomedical text mining and the pressing needs of biocurators for better tools, few existing text-mining tools have been successfully integrated into production literature curation systems such as those used by the expert curated databases. To close this gap and better understand all aspects of literature curation, we invited submissions of written descriptions of curation workflows from expert curated databases for the BioCreative 2012 Workshop Track II. We received seven qualified contributions, primarily from model organism databases. Based on these descriptions, we identified commonalities and differences across the workflows, the common ontologies and controlled vocabularies used and the current and desired uses of text mining for biocuration. Compared to a survey done in 2009, our 2012 results show that many more databases are now using text mining in parts of their curation workflows. In addition, the workshop participants identified text-mining aids for finding gene names and symbols (gene indexing), prioritization of documents for curation (document triage) and ontology concept assignment as those most desired by the biocurators. DATABASE URL: http://www.biocreative.org/tasks/bc-workshop-2012/workflow/.
Fei, Lin; Zhao, Jing; Leng, Jiahao; Zhang, Shujian
2017-10-12
The ALIPORC full-text database is targeted at a specific full-text database of acupuncture literature in the Republic of China. Starting in 2015, till now, the database has been getting completed, focusing on books relevant with acupuncture, articles and advertising documents, accomplished or published in the Republic of China. The construction of this database aims to achieve the source sharing of acupuncture medical literature in the Republic of China through the retrieval approaches to diversity and accurate content presentation, contributes to the exchange of scholars, reduces the paper damage caused by paging and simplify the retrieval of the rare literature. The writers have made the explanation of the database in light of sources, characteristics and current situation of construction; and have discussed on improving the efficiency and integrity of the database and deepening the development of acupuncture literature in the Republic of China.
Dworkin, Robert H; Bruehl, Stephen; Fillingim, Roger B; Loeser, John D; Terman, Gregory W; Turk, Dennis C
2016-09-01
A variety of approaches have been used to develop diagnostic criteria for chronic pain. The published evidence of the reliability and validity of existing diagnostic criteria is limited, and these criteria have typically not been used in clinical practice. The availability of a widely accepted, consistently applied, and evidence-based taxonomy of diagnostic criteria would improve the quality of clinical research on chronic pain and would be of great value in clinical practice. To address the need for evidence-based diagnostic criteria for the major chronic pain conditions, the Analgesic, Anesthetic, and Addiction Clinical Trial Translations, Innovations, Opportunities, and Networks (ACTTION) public-private partnership with the US Food and Drug Administration and the American Pain Society (APS) have collaborated on the development of the ACTTION-APS Pain Taxonomy (AAPT). AAPT provides a multidimensional framework that is applied systematically in the development of diagnostic criteria. This article (1) describes the background and rationale for AAPT; (2) presents the AAPT taxonomy and the specific conditions for which diagnostic criteria have been developed (to be published separately); (3) briefly reviews the 5 dimensions that constitute the AAPT multidimensional framework and describes the 7 accompanying articles that discuss these dimensions and other important issues involving AAPT; and (4) provides an overview of next steps, specifically, the general processes by which the initial set of diagnostic criteria (for which the evidence base has been drawn from the literature, systematic reviews, and secondary analyses of existing databases) will undergo additional assessments of reliability and validity. To address the need for evidence-based diagnostic criteria for the major chronic pain conditions, the AAPT provides a multidimensional framework that is applied systematically in the development of diagnostic criteria. The long-term objective of AAPT is to advance the scientific understanding of chronic pain and its treatment. Copyright © 2016 American Pain Society. Published by Elsevier Inc. All rights reserved.
Learning from physics text: A synthesis of recent research
NASA Astrophysics Data System (ADS)
Alexander, Patricia A.; Kulikowich, Jonna M.
Learning from physics text is described as a complex interaction of learner, text, and context variables. As a multidimensional procedure, text processing in the domain of physics relies on readers' knowledge and interest, and on readers' ability to monitor or regulate their processing. Certain textual features intended to assist readers in understanding and remembering physics content may actually work to the detriment of those very processes. Inclusion of seductive details and the incorporation of analogies may misdirect readers' attention or may increase processing demands, particularly in those cases when readers' physics knowledge is low. The questioning behaviors of teachers also impact on the task of comprehending physics texts. Finally, within the context of the classroom, the information that teachers dispense or the materials they employ can significantly influence the process of learning from physics text.
Testlet-Based Multidimensional Adaptive Testing
Frey, Andreas; Seitz, Nicki-Nils; Brandt, Steffen
2016-01-01
Multidimensional adaptive testing (MAT) is a highly efficient method for the simultaneous measurement of several latent traits. Currently, no psychometrically sound approach is available for the use of MAT in testlet-based tests. Testlets are sets of items sharing a common stimulus such as a graph or a text. They are frequently used in large operational testing programs like TOEFL, PISA, PIRLS, or NAEP. To make MAT accessible for such testing programs, we present a novel combination of MAT with a multidimensional generalization of the random effects testlet model (MAT-MTIRT). MAT-MTIRT compared to non-adaptive testing is examined for several combinations of testlet effect variances (0.0, 0.5, 1.0, and 1.5) and testlet sizes (3, 6, and 9 items) with a simulation study considering three ability dimensions with simple loading structure. MAT-MTIRT outperformed non-adaptive testing regarding the measurement precision of the ability estimates. Further, the measurement precision decreased when testlet effect variances and testlet sizes increased. The suggested combination of the MTIRT model therefore provides a solution to the substantial problems of testlet-based tests while keeping the length of the test within an acceptable range. PMID:27917132
CELL5M: A geospatial database of agricultural indicators for Africa South of the Sahara.
Koo, Jawoo; Cox, Cindy M; Bacou, Melanie; Azzarri, Carlo; Guo, Zhe; Wood-Sichra, Ulrike; Gong, Queenie; You, Liangzhi
2016-01-01
Recent progress in large-scale georeferenced data collection is widening opportunities for combining multi-disciplinary datasets from biophysical to socioeconomic domains, advancing our analytical and modeling capacity. Granular spatial datasets provide critical information necessary for decision makers to identify target areas, assess baseline conditions, prioritize investment options, set goals and targets and monitor impacts. However, key challenges in reconciling data across themes, scales and borders restrict our capacity to produce global and regional maps and time series. This paper provides overview, structure and coverage of CELL5M-an open-access database of geospatial indicators at 5 arc-minute grid resolution-and introduces a range of analytical applications and case-uses. CELL5M covers a wide set of agriculture-relevant domains for all countries in Africa South of the Sahara and supports our understanding of multi-dimensional spatial variability inherent in farming landscapes throughout the region.
NASA Astrophysics Data System (ADS)
Sagarminaga, Y.; Galparsoro, I.; Reig, R.; Sánchez, J. A.
2012-04-01
Since 2000, an intense effort was conducted in AZTI's Marine Research Division to set up a data management system which could gather all the marine datasets that were being produced by different in-house research projects. For that, a corporative GIS was designed that included a data and metadata repository, a database, a layer catalog & search application and an internet map viewer. Several layers, mostly dealing with physical, chemical and biological in-situ sampling, and basic and thematic cartography including bathymetry, geomorphology, different species habitat maps, and human pressure and activities maps, were successfully gathered in this system. Very soon, it was realised that new marine technologies yielding continuous multidimensional data, sometimes called FES (Fluid Earth System) data, were difficult to handle in this structure. The data affected, mainly included numerical oceanographic and meteorological models, remote sensing data, coastal RADAR data, and some in-situ observational systems such as CTD's casts, moored or lagrangian buoys, etc. A management system for gridded multidimensional data was developed using standardized formats (netcdf using CF conventions) and tools such as THREDDS catalog (UNIDATA/UCAR) providing web services such as OPENDAP, NCSS, and WCS, as well as ncWMS service developed by the Reading e-science Center. At present, a system (ITSASGIS-5D) is being developed, based on OGC standards and open-source tools to allow interoperability between all the data types mentioned before. This system includes, in the server side, postgresql/postgis databases and geoserver for GIS layers, and THREDDS/Opendap and ncWMS services for FES gridded data. Moreover, an on-line client is being developed to allow joint access, user configuration, data visualisation & query and data distribution. This client is using mapfish, ExtJS - GeoEXT, and openlayers libraries. Through this presentation the elements of the first released version of this system will be described and showed, together with the new topics to be developed in new versions that include among others, the integration of geoNetwork libraries and tools for both FES and GIS metadata management, and the use of new OGC Sensor Observation Services (SOS) to integrate non gridded multidimensional data such as time series, depth profiles or trajectories provided by different observational systems. The final aim of this approach is to contribute to the multidisciplinary access and use of marine data for management and research activities, and facilitate the implementation of integrated ecosystem based approaches in the fields of fisheries advice and management, marine spatial planning, or the implementation of the European policies such as the Water Framework Directive, the Marine Strategy Framework Directive or the Habitat Framework Directive.
Subject Retrieval from Full-Text Databases in the Humanities
ERIC Educational Resources Information Center
East, John W.
2007-01-01
This paper examines the problems involved in subject retrieval from full-text databases of secondary materials in the humanities. Ten such databases were studied and their search functionality evaluated, focusing on factors such as Boolean operators, document surrogates, limiting by subject area, proximity operators, phrase searching, wildcards,…
Intelligent Data Analysis in the EMERCOM Information System
NASA Astrophysics Data System (ADS)
Elena, Sharafutdinova; Tatiana, Avdeenko; Bakaev, Maxim
2017-01-01
The paper describes an information system development project for the Russian Ministry of Emergency Situations (MES, whose international operations body is known as EMERCOM), which was attended by the representatives of both the IT industry and the academia. Besides the general description of the system, we put forward OLAP and Data Mining-based approaches towards the intelligent analysis of the data accumulated in the database. In particular, some operational OLAP reports and an example of multi-dimensional information space based on OLAP Data Warehouse are presented. Finally, we outline Data Mining application to support decision-making regarding security inspections planning and results consideration.
Towards computational improvement of DNA database indexing and short DNA query searching.
Stojanov, Done; Koceski, Sašo; Mileva, Aleksandra; Koceska, Nataša; Bande, Cveta Martinovska
2014-09-03
In order to facilitate and speed up the search of massive DNA databases, the database is indexed at the beginning, employing a mapping function. By searching through the indexed data structure, exact query hits can be identified. If the database is searched against an annotated DNA query, such as a known promoter consensus sequence, then the starting locations and the number of potential genes can be determined. This is particularly relevant if unannotated DNA sequences have to be functionally annotated. However, indexing a massive DNA database and searching an indexed data structure with millions of entries is a time-demanding process. In this paper, we propose a fast DNA database indexing and searching approach, identifying all query hits in the database, without having to examine all entries in the indexed data structure, limiting the maximum length of a query that can be searched against the database. By applying the proposed indexing equation, the whole human genome could be indexed in 10 hours on a personal computer, under the assumption that there is enough RAM to store the indexed data structure. Analysing the methodology proposed by Reneker, we observed that hits at starting positions [Formula: see text] are not reported, if the database is searched against a query shorter than [Formula: see text] nucleotides, such that [Formula: see text] is the length of the DNA database words being mapped and [Formula: see text] is the length of the query. A solution of this drawback is also presented.
ERIC Educational Resources Information Center
Perkins, Kathleen M.
2016-01-01
Theatre is a multi-dimensional discipline encompassing aspects of several domains in the arts and humanities. Therefore, an array of scholarly practices, pedagogies, and methods might be available to a SoTL researcher from the close reading of texts in script analysis to portfolio critiques in set, costume, and lighting design--approaches shared…
Assessing semantic similarity of texts - Methods and algorithms
NASA Astrophysics Data System (ADS)
Rozeva, Anna; Zerkova, Silvia
2017-12-01
Assessing the semantic similarity of texts is an important part of different text-related applications like educational systems, information retrieval, text summarization, etc. This task is performed by sophisticated analysis, which implements text-mining techniques. Text mining involves several pre-processing steps, which provide for obtaining structured representative model of the documents in a corpus by means of extracting and selecting the features, characterizing their content. Generally the model is vector-based and enables further analysis with knowledge discovery approaches. Algorithms and measures are used for assessing texts at syntactical and semantic level. An important text-mining method and similarity measure is latent semantic analysis (LSA). It provides for reducing the dimensionality of the document vector space and better capturing the text semantics. The mathematical background of LSA for deriving the meaning of the words in a given text by exploring their co-occurrence is examined. The algorithm for obtaining the vector representation of words and their corresponding latent concepts in a reduced multidimensional space as well as similarity calculation are presented.
Ravikumar, Komandur Elayavilli; Wagholikar, Kavishwar B; Li, Dingcheng; Kocher, Jean-Pierre; Liu, Hongfang
2015-06-06
Advances in the next generation sequencing technology has accelerated the pace of individualized medicine (IM), which aims to incorporate genetic/genomic information into medicine. One immediate need in interpreting sequencing data is the assembly of information about genetic variants and their corresponding associations with other entities (e.g., diseases or medications). Even with dedicated effort to capture such information in biological databases, much of this information remains 'locked' in the unstructured text of biomedical publications. There is a substantial lag between the publication and the subsequent abstraction of such information into databases. Multiple text mining systems have been developed, but most of them focus on the sentence level association extraction with performance evaluation based on gold standard text annotations specifically prepared for text mining systems. We developed and evaluated a text mining system, MutD, which extracts protein mutation-disease associations from MEDLINE abstracts by incorporating discourse level analysis, using a benchmark data set extracted from curated database records. MutD achieves an F-measure of 64.3% for reconstructing protein mutation disease associations in curated database records. Discourse level analysis component of MutD contributed to a gain of more than 10% in F-measure when compared against the sentence level association extraction. Our error analysis indicates that 23 of the 64 precision errors are true associations that were not captured by database curators and 68 of the 113 recall errors are caused by the absence of associated disease entities in the abstract. After adjusting for the defects in the curated database, the revised F-measure of MutD in association detection reaches 81.5%. Our quantitative analysis reveals that MutD can effectively extract protein mutation disease associations when benchmarking based on curated database records. The analysis also demonstrates that incorporating discourse level analysis significantly improved the performance of extracting the protein-mutation-disease association. Future work includes the extension of MutD for full text articles.
Damming the genomic data flood using a comprehensive analysis and storage data structure
Bouffard, Marc; Phillips, Michael S.; Brown, Andrew M.K.; Marsh, Sharon; Tardif, Jean-Claude; van Rooij, Tibor
2010-01-01
Data generation, driven by rapid advances in genomic technologies, is fast outpacing our analysis capabilities. Faced with this flood of data, more hardware and software resources are added to accommodate data sets whose structure has not specifically been designed for analysis. This leads to unnecessarily lengthy processing times and excessive data handling and storage costs. Current efforts to address this have centered on developing new indexing schemas and analysis algorithms, whereas the root of the problem lies in the format of the data itself. We have developed a new data structure for storing and analyzing genotype and phenotype data. By leveraging data normalization techniques, database management system capabilities and the use of a novel multi-table, multidimensional database structure we have eliminated the following: (i) unnecessarily large data set size due to high levels of redundancy, (ii) sequential access to these data sets and (iii) common bottlenecks in analysis times. The resulting novel data structure horizontally divides the data to circumvent traditional problems associated with the use of databases for very large genomic data sets. The resulting data set required 86% less disk space and performed analytical calculations 6248 times faster compared to a standard approach without any loss of information. Database URL: http://castor.pharmacogenomics.ca PMID:21159730
An information model for managing multi-dimensional gridded data in a GIS
NASA Astrophysics Data System (ADS)
Xu, H.; Abdul-Kadar, F.; Gao, P.
2016-04-01
Earth observation agencies like NASA and NOAA produce huge volumes of historical, near real-time, and forecasting data representing terrestrial, atmospheric, and oceanic phenomena. The data drives climatological and meteorological studies, and underpins operations ranging from weather pattern prediction and forest fire monitoring to global vegetation analysis. These gridded data sets are distributed mostly as files in HDF, GRIB, or netCDF format and quantify variables like precipitation, soil moisture, or sea surface temperature, along one or more dimensions like time and depth. Although the data cube is a well-studied model for storing and analyzing multi-dimensional data, the GIS community remains in need of a solution that simplifies interactions with the data, and elegantly fits with existing database schemas and dissemination protocols. This paper presents an information model that enables Geographic Information Systems (GIS) to efficiently catalog very large heterogeneous collections of geospatially-referenced multi-dimensional rasters—towards providing unified access to the resulting multivariate hypercubes. We show how the implementation of the model encapsulates format-specific variations and provides unified access to data along any dimension. We discuss how this framework lends itself to familiar GIS concepts like image mosaics, vector field visualization, layer animation, distributed data access via web services, and scientific computing. Global data sources like MODIS from USGS and HYCOM from NOAA illustrate how one would employ this framework for cataloging, querying, and intuitively visualizing such hypercubes. ArcGIS—an established platform for processing, analyzing, and visualizing geospatial data—serves to demonstrate how this integration brings the full power of GIS to the scientific community.
NASA Astrophysics Data System (ADS)
Marchand, Pierre; Brisebois, Alexandre; Bédard, Yvan; Edwards, Geoffrey
This paper presents the results obtained with a new type of spatiotemporal topological dimension implemented within a hypercube, i.e., within a multidimensional database (MDDB) structure formed by the conjunction of several thematic, spatial and temporal dimensions. Our goal is to support efficient SpatioTemporal Exploration and Analysis (STEA) in the context of Automatic Position Reporting System (APRS), the worldwide amateur radio system for position report transmission. Mobile APRS stations are equipped with GPS navigation systems to provide real-time positioning reports. Previous research about the multidimensional approach has proved good potential for spatiotemporal exploration and analysis despite a lack of explicit topological operators (spatial, temporal and spatiotemporal). Our project implemented such operators through a hierarchy of operators that are applied to pairs of instances of objects. At the top of the hierarchy, users can use simple operators such as "same place", "same time" or "same time, same place". As they drill down into the hierarchy, more detailed topological operators are made available such as "adjacent immediately after", "touch during" or more detailed operators. This hierarchy is structured according to four levels of granularity based on cognitive models, generalized relationships and formal models of topological relationships. In this paper, we also describe the generic approach which allows efficient STEA within the multidimensional approach. Finally, we demonstrate that such an implementation offers query run times which permit to maintain a "train-of-thought" during exploration and analysis operations as they are compatible with Newell's cognitive band (query runtime<10 s) (Newell, A., 1990. Unified theories of cognition. Harvard University Press, Cambridge MA, 549 p.).
NASA Astrophysics Data System (ADS)
Merticariu, Vlad; Misev, Dimitar; Baumann, Peter
2017-04-01
While python has developed into the lingua franca in Data Science there is often a paradigm break when accessing specialized tools. In particular for one of the core data categories in science and engineering, massive multi-dimensional arrays, out-of-memory solutions typically employ their own, different models. We discuss this situation on the example of the scalable open-source array engine, rasdaman ("raster data manager") which offers access to and processing of Petascale multi-dimensional arrays through an SQL-style array query language, rasql. Such queries are executed in the server on a storage engine utilizing adaptive array partitioning and based on a processing engine implementing a "tile streaming" paradigm to allow processing of arrays massively larger than server RAM. The rasdaman QL has acted as blueprint for forthcoming ISO Array SQL and the Open Geospatial Consortium (OGC) geo analytics language, Web Coverage Processing Service, adopted in 2008. Not surprisingly, rasdaman is OGC and INSPIRE Reference Implementation for their "Big Earth Data" standards suite. Recently, rasdaman has been augmented with a python interface which allows to transparently interact with the database (credits go to Siddharth Shukla's Master Thesis at Jacobs University). Programmers do not need to know the rasdaman query language, as the operators are silently transformed, through lazy evaluation, into queries. Arrays delivered are likewise automatically transformed into their python representation. In the talk, the rasdaman concept will be illustrated with the help of large-scale real-life examples of operational satellite image and weather data services, and sample python code.
Multidimensional measures validated for home health needs of older persons: A systematic review.
de Rossi Figueiredo, Daniela; Paes, Lucilene Gama; Warmling, Alessandra Martins; Erdmann, Alacoque Lorenzini; de Mello, Ana Lúcia Schaefer Ferreira
2018-01-01
To conduct a systematic review of the literature on valid and reliable multidimensional instruments to assess home health needs of older persons. Systematic review. Electronic databases, PubMed/Medline, Web of Science, Scopus, Cumulative Index to Nursing and Allied Health Literature, Scientific Electronic Library Online and the Latin American and Caribbean Health Sciences Information. All English, Portuguese and Spanish literature which included studies of reliability and validity of instruments that assessed at least two dimensions: physical, psychological, social support and functional independence, self-rated health behaviors and contextual environment and if such instruments proposed interventions after evaluation and/or monitoring changes over a period of time. Older persons aged 60 years or older. Of the 2397 studies identified, 32 were considered eligible. Two-thirds of the instruments proposed the physical, psychological, social support and functional independence dimensions. Inter-observer and intra-observer reliability and internal consistency values were 0.7 or above. More than two-thirds of the studies included validity (n=26) and more than one validity was tested in 15% (n=4) of these. Only 7% (n=2) proposed interventions after evaluation and/or monitoring changes over a period of time. Although the multidimensional assessment was performed, and the reliability values of the reviewed studies were satisfactory, different validity tests were not present in several studies. A gap at the instrument conception was observed related to interventions after evaluation and/or monitoring changes over a period of time. Further studies with this purpose are necessary for home health needs of the older persons. Copyright © 2017 Elsevier Ltd. All rights reserved.
Searching Harvard Business Review Online. . . Lessons in Searching a Full Text Database.
ERIC Educational Resources Information Center
Tenopir, Carol
1985-01-01
This article examines the Harvard Business Review Online (HBRO) database (bibliographic description fields, abstracts, extracted information, full text, subject descriptors) and reports on 31 sample HBRO searches conducted in Bibliographic Retrieval Services to test differences between searching full text and searching bibliographic record. Sample…
AccuNet/AP (Associated Press) Multimedia Archive
ERIC Educational Resources Information Center
Young, Terrence E., Jr.
2004-01-01
The AccuNet/AP Multimedia Archive is an electronic library containing the AP's current photos and a selection of pictures from their enormous print and negative library, as well as text and graphic material. It is composed of two photo databases as well as graphics, text, and audio databases. The features of this database are briefly described in…
Geodata Modeling and Query in Geographic Information Systems
NASA Technical Reports Server (NTRS)
Adam, Nabil
1996-01-01
Geographic information systems (GIS) deal with collecting, modeling, man- aging, analyzing, and integrating spatial (locational) and non-spatial (attribute) data required for geographic applications. Examples of spatial data are digital maps, administrative boundaries, road networks, and those of non-spatial data are census counts, land elevations and soil characteristics. GIS shares common areas with a number of other disciplines such as computer- aided design, computer cartography, database management, and remote sensing. None of these disciplines however, can by themselves fully meet the requirements of a GIS application. Examples of such requirements include: the ability to use locational data to produce high quality plots, perform complex operations such as network analysis, enable spatial searching and overlay operations, support spatial analysis and modeling, and provide data management functions such as efficient storage, retrieval, and modification of large datasets; independence, integrity, and security of data; and concurrent access to multiple users. It is on the data management issues that we devote our discussions in this monograph. Traditionally, database management technology have been developed for business applications. Such applications require, among other things, capturing the data requirements of high-level business functions and developing machine- level implementations; supporting multiple views of data and yet providing integration that would minimize redundancy and maintain data integrity and security; providing a high-level language for data definition and manipulation; allowing concurrent access to multiple users; and processing user transactions in an efficient manner. The demands on database management systems have been for speed, reliability, efficiency, cost effectiveness, and user-friendliness. Significant progress have been made in all of these areas over the last two decades to the point that many generalized database platforms are now available for developing data intensive applications that run in real-time. While continuous improvement is still being made at a very fast-paced and competitive rate, new application areas such as computer aided design, image processing, VLSI design, and GIS have been identified by many as the next generation of database applications. These new application areas pose serious challenges to the currently available database technology. At the core of these challenges is the nature of data that is manipulated. In traditional database applications, the database objects do not have any spatial dimension, and as such, can be thought of as point data in a multi-dimensional space. For example, each instance of an entity EMPLOYEE will have a unique value corresponding to every attribute such as employee id, employee name, employee address and so on. Thus, every Employee instance can be thought of as a point in a multi-dimensional space where each dimension is represented by an attribute. Furthermore, all operations on such data are one-dimensional. Thus, users may retrieve all entities satisfying one or more constraints. Examples of such constraints include employees with addresses in a certain area code, or salaries within a certain range. Even though constraints can be specified on multiple attributes (dimensions), the search for such data is essentially orthogonal across these dimensions.
ERIC Educational Resources Information Center
Tauchert, Wolfgang; And Others
1991-01-01
Describes the PADOK-II project in Germany, which was designed to give information on the effects of linguistic algorithms on retrieval in a full-text database, the German Patent Information System (GPI). Relevance assessments are discussed, statistical evaluations are described, and searches are compared for the full-text section versus the…
ELNET--The Electronic Library Database System.
ERIC Educational Resources Information Center
King, Shirley V.
1991-01-01
ELNET (Electronic Library Network), a Japanese language database, allows searching of index terms and free text terms from articles and stores the full text of the articles on an optical disc system. Users can order fax copies of the text from the optical disc. This article also explains online searching and discusses machine translation. (LRW)
The Weaknesses of Full-Text Searching
ERIC Educational Resources Information Center
Beall, Jeffrey
2008-01-01
This paper provides a theoretical critique of the deficiencies of full-text searching in academic library databases. Because full-text searching relies on matching words in a search query with words in online resources, it is an inefficient method of finding information in a database. This matching fails to retrieve synonyms, and it also retrieves…
NASA Astrophysics Data System (ADS)
Hwang, Darryl H.; Ma, Kevin; Yepes, Fernando; Nadamuni, Mridula; Nayyar, Megha; Liu, Brent; Duddalwar, Vinay; Lepore, Natasha
2015-12-01
A conventional radiology report primarily consists of a large amount of unstructured text, and lacks clear, concise, consistent and content-rich information. Hence, an area of unmet clinical need consists of developing better ways to communicate radiology findings and information specific to each patient. Here, we design a new workflow and reporting system that combines and integrates advances in engineering technology with those from the medical sciences, the Multidimensional Interactive Radiology Report and Analysis (MIRRA). Until recently, clinical standards have primarily relied on 2D images for the purpose of measurement, but with the advent of 3D processing, many of the manually measured metrics can be automated, leading to better reproducibility and less subjective measurement placement. Hence, we make use this newly available 3D processing in our workflow. Our pipeline is used here to standardize the labeling, tracking, and quantifying of metrics for renal masses.
Eigenhuis, Annemarie; Kamphuis, Jan H; Noordhof, Arjen
2017-09-01
A growing body of research suggests that the same general dimensions can describe normal and pathological personality, but most of the supporting evidence is exploratory. We aim to determine in a confirmatory framework the extent to which responses on the Multidimensional Personality Questionnaire (MPQ) are identical across general and clinical samples. We tested the Dutch brief form of the MPQ (MPQ-BF-NL) for measurement invariance across a general population subsample (N = 365) and a clinical sample (N = 365), using Multiple Group Confirmatory Factor Analysis (MGCFA) and Multiple Group Exploratory Structural Equation Modeling (MGESEM). As an omnibus personality test, the MPQ-BF-NL revealed strict invariance, indicating absence of bias. Unidimensional per scale tests for measurement invariance revealed that 10% of items appeared to contain bias across samples. Item bias only affected the scale interpretation of Achievement, with individuals from the clinical sample more readily admitting to put high demands on themselves than individuals from the general sample, regardless of trait level. This formal test of equivalence provides strong evidence for the common structure of normal and pathological personality and lends further support to the clinical utility of the MPQ. (PsycINFO Database Record (c) 2017 APA, all rights reserved).
"Science SQL" as a Building Block for Flexible, Standards-based Data Infrastructures
NASA Astrophysics Data System (ADS)
Baumann, Peter
2016-04-01
We have learnt to live with the pain of separating data and metadata into non-interoperable silos. For metadata, we enjoy the flexibility of databases, be they relational, graph, or some other NoSQL. Contrasting this, users still "drown in files" as an unstructured, low-level archiving paradigm. It is time to bridge this chasm which once was technologically induced, but today can be overcome. One building block towards a common re-integrated information space is to support massive multi-dimensional spatio-temporal arrays. These "datacubes" appear as sensor, image, simulation, and statistics data in all science and engineering domains, and beyond. For example, 2-D satellilte imagery, 2-D x/y/t image timeseries and x/y/z geophysical voxel data, and 4-D x/y/z/t climate data contribute to today's data deluge in the Earth sciences. Virtual observatories in the Space sciences routinely generate Petabytes of such data. Life sciences deal with microarray data, confocal microscopy, human brain data, which all fall into the same category. The ISO SQL/MDA (Multi-Dimensional Arrays) candidate standard is extending SQL with modelling and query support for n-D arrays ("datacubes") in a flexible, domain-neutral way. This heralds a new generation of services with new quality parameters, such as flexibility, ease of access, embedding into well-known user tools, and scalability mechanisms that remain completely transparent to users. Technology like the EU rasdaman ("raster data manager") Array Database system can support all of the above examples simultaneously, with one technology. This is practically proven: As of today, rasdaman is in operational use on hundreds of Terabytes of satellite image timeseries datacubes, with transparent query distribution across more than 1,000 nodes. Therefore, Array Databases offering SQL/MDA constitute a natural common building block for next-generation data infrastructures. Being initiator and editor of the standard we present principles, implementation facets, and application examples as a basis for further discussion. Further, we highlight recent implementation progress in parallelization, data distribution, and query optimization showing their effects on real-life use cases.
Database tomography for commercial application
NASA Technical Reports Server (NTRS)
Kostoff, Ronald N.; Eberhart, Henry J.
1994-01-01
Database tomography is a method for extracting themes and their relationships from text. The algorithms, employed begin with word frequency and word proximity analysis and build upon these results. When the word 'database' is used, think of medical or police records, patents, journals, or papers, etc. (any text information that can be computer stored). Database tomography features a full text, user interactive technique enabling the user to identify areas of interest, establish relationships, and map trends for a deeper understanding of an area of interest. Database tomography concepts and applications have been reported in journals and presented at conferences. One important feature of the database tomography algorithm is that it can be used on a database of any size, and will facilitate the users ability to understand the volume of content therein. While employing the process to identify research opportunities it became obvious that this promising technology has potential applications for business, science, engineering, law, and academe. Examples include evaluating marketing trends, strategies, relationships and associations. Also, the database tomography process would be a powerful component in the area of competitive intelligence, national security intelligence and patent analysis. User interests and involvement cannot be overemphasized.
Earth science big data at users' fingertips: the EarthServer Science Gateway Mobile
NASA Astrophysics Data System (ADS)
Barbera, Roberto; Bruno, Riccardo; Calanducci, Antonio; Fargetta, Marco; Pappalardo, Marco; Rundo, Francesco
2014-05-01
The EarthServer project (www.earthserver.eu), funded by the European Commission under its Seventh Framework Program, aims at establishing open access and ad-hoc analytics on extreme-size Earth Science data, based on and extending leading-edge Array Database technology. The core idea is to use database query languages as client/server interface to achieve barrier-free "mix & match" access to multi-source, any-size, multi-dimensional space-time data -- in short: "Big Earth Data Analytics" - based on the open standards of the Open Geospatial Consortium Web Coverage Processing Service (OGC WCPS) and the W3C XQuery. EarthServer combines both, thereby achieving a tight data/metadata integration. Further, the rasdaman Array Database System (www.rasdaman.com) is extended with further space-time coverage data types. On server side, highly effective optimizations - such as parallel and distributed query processing - ensure scalability to Exabyte volumes. In this contribution we will report on the EarthServer Science Gateway Mobile, an app for both iOS and Android-based devices that allows users to seamlessly access some of the EarthServer applications using SAML-based federated authentication and fine-grained authorisation mechanisms.
Full-Text Linking: Affiliated versus Nonaffiliated Access in a Free Database.
ERIC Educational Resources Information Center
Grogg, Jill E.; Andreadis, Debra K.; Kirk, Rachel A.
2002-01-01
Presents a comparison of access to full-text articles from a free bibliographic database (PubSCIENCE) for affiliated and unaffiliated users. Found that affiliated users had access to more full-text articles than unaffiliated users had, and that both types of users could increase their level of access through additional searching and greater…
Crafting a Social Context for Medical Informatics Networks
NASA Astrophysics Data System (ADS)
Patel, Salil H.
Effective healthcare delivery is increasingly predicated upon the availability, accuracy, and integrity of personal health information. Tracking and analysis of medical information throughout its lifeeycle may be viewed through the lenses of both physical network architecture and the broader social context in which such information is gathered and applied. As information technology and evidence-based practice models evolve in tandem, the development of interlinked multimodal and multidimensional databases has shown great promise for improving public health. To this end. providers, regulators, payers, and individual patients each share rights and responsibilities in creating a milieu which both, fosters and protects the practice and promise of medical information.
Kafkas, Şenay; Kim, Jee-Hyub; Pi, Xingjun; McEntyre, Johanna R
2015-01-01
In this study, we present an analysis of data citation practices in full text research articles and their corresponding supplementary data files, made available in the Open Access set of articles from Europe PubMed Central. Our aim is to investigate whether supplementary data files should be considered as a source of information for integrating the literature with biomolecular databases. Using text-mining methods to identify and extract a variety of core biological database accession numbers, we found that the supplemental data files contain many more database citations than the body of the article, and that those citations often take the form of a relatively small number of articles citing large collections of accession numbers in text-based files. Moreover, citation of value-added databases derived from submission databases (such as Pfam, UniProt or Ensembl) is common, demonstrating the reuse of these resources as datasets in themselves. All the database accession numbers extracted from the supplementary data are publicly accessible from http://dx.doi.org/10.5281/zenodo.11771. Our study suggests that supplementary data should be considered when linking articles with data, in curation pipelines, and in information retrieval tasks in order to make full use of the entire research article. These observations highlight the need to improve the management of supplemental data in general, in order to make this information more discoverable and useful.
Potts, Maryellen; Cartmell, Kathleen B; Nemeth, Lynne; Bhattacharjee, Gautam; Qanungo, Suparna
2018-05-01
To meet the growing need for palliative care in low-resource countries, palliative care programs should be evidence based and contextually appropriate. This study was conducted to synthesize the current evidence to guide future programmatic and research efforts. This systematic review evaluated palliative care outcome measures, outcomes, and interventions in low-resource countries. After title searches, abstracts and full-text articles were screened for inclusion. Data were extracted to report on intervention models, outcome measures used, and intervention outcomes. Eighteen papers were reviewed, reporting on interventions conducted across nine low-resource countries. These interventions evaluated home-based palliative care models; a community-managed model; palliative care integrated with hospitals, hospices, or HIV clinics; and models focused on patients' self-management. Three studies were randomized controlled trials. Other studies used nonrandomized trials, cohort studies, mixed methods, pre-post test evaluation, cost-accounting evaluation, and cross-sectional surveys. Thirteen studies measured physical outcomes, 10 using multidimensional instruments. Nine studies measured psychological outcomes, eight using multidimensional instruments. Nine studies measured social outcomes, seven using multidimensional instruments. Nine studies measured outcomes across multiple domains. Across outcomes evaluated, results were reported in the direction of benefit associated with palliative care interventions. Many palliative care intervention models exist to serve patients in low-resource countries. Yet, limited high-quality evidence from low-resource countries is available to document intervention outcomes. Rigorous experimental studies and greater measurement of multidimensional aspects of palliative care are needed to advance the science of palliative care in low-resource settings. Copyright © 2017 American Academy of Hospice and Palliative Medicine. Published by Elsevier Inc. All rights reserved.
Full-text, Downloading, & Other Issues.
ERIC Educational Resources Information Center
Tenopir, Carol
1983-01-01
Issues having a possible impact on online search services in libraries are discussed including full text databases, front-end processors which translate user's input into the command language of an appropriate system, downloading to create personal files from commercial databases, and pricing. (EJS)
Modeling change from large-scale high-dimensional spatio-temporal array data
NASA Astrophysics Data System (ADS)
Lu, Meng; Pebesma, Edzer
2014-05-01
The massive data that come from Earth observation satellite and other sensors provide significant information for modeling global change. At the same time, the high dimensionality of the data has brought challenges in data acquisition, management, effective querying and processing. In addition, the output of earth system modeling tends to be data intensive and needs methodologies for storing, validation, analyzing and visualization, e.g. as maps. An important proportion of earth system observations and simulated data can be represented as multi-dimensional array data, which has received increasingly attention in big data management and spatial-temporal analysis. Study cases will be developed in natural science such as climate change, hydrological modeling, sediment dynamics, from which the addressing of big data problems is necessary. Multi-dimensional array-based database management and analytics system such as Rasdaman, SciDB, and R will be applied to these cases. From these studies will hope to learn the strengths and weaknesses of these systems, how they might work together or how semantics of array operations differ, through addressing the problems associated with big data. Research questions include: • How can we reduce dimensions spatially and temporally, or thematically? • How can we extend existing GIS functions to work on multidimensional arrays? • How can we combine data sets of different dimensionality or different resolutions? • Can map algebra be extended to an intelligible array algebra? • What are effective semantics for array programming of dynamic data driven applications? • In which sense are space and time special, as dimensions, compared to other properties? • How can we make the analysis of multi-spectral, multi-temporal and multi-sensor earth observation data easy?
Large-scale feature searches of collections of medical imagery
NASA Astrophysics Data System (ADS)
Hedgcock, Marcus W.; Karshat, Walter B.; Levitt, Tod S.; Vosky, D. N.
1993-09-01
Large scale feature searches of accumulated collections of medical imagery are required for multiple purposes, including clinical studies, administrative planning, epidemiology, teaching, quality improvement, and research. To perform a feature search of large collections of medical imagery, one can either search text descriptors of the imagery in the collection (usually the interpretation), or (if the imagery is in digital format) the imagery itself. At our institution, text interpretations of medical imagery are all available in our VA Hospital Information System. These are downloaded daily into an off-line computer. The text descriptors of most medical imagery are usually formatted as free text, and so require a user friendly database search tool to make searches quick and easy for any user to design and execute. We are tailoring such a database search tool (Liveview), developed by one of the authors (Karshat). To further facilitate search construction, we are constructing (from our accumulated interpretation data) a dictionary of medical and radiological terms and synonyms. If the imagery database is digital, the imagery which the search discovers is easily retrieved from the computer archive. We describe our database search user interface, with examples, and compare the efficacy of computer assisted imagery searches from a clinical text database with manual searches. Our initial work on direct feature searches of digital medical imagery is outlined.
Application of data mining approaches to drug delivery.
Ekins, Sean; Shimada, Jun; Chang, Cheng
2006-11-30
Computational approaches play a key role in all areas of the pharmaceutical industry from data mining, experimental and clinical data capture to pharmacoeconomics and adverse events monitoring. They will likely continue to be indispensable assets along with a growing library of software applications. This is primarily due to the increasingly massive amount of biology, chemistry and clinical data, which is now entering the public domain mainly as a result of NIH and commercially funded projects. We are therefore in need of new methods for mining this mountain of data in order to enable new hypothesis generation. The computational approaches include, but are not limited to, database compilation, quantitative structure activity relationships (QSAR), pharmacophores, network visualization models, decision trees, machine learning algorithms and multidimensional data visualization software that could be used to improve drug delivery after mining public and/or proprietary data. We will discuss some areas of unmet needs in the area of data mining for drug delivery that can be addressed with new software tools or databases of relevance to future pharmaceutical projects.
NASA Astrophysics Data System (ADS)
Eakins, John P.; Edwards, Jonathan D.; Riley, K. Jonathan; Rosin, Paul L.
2001-01-01
Many different kinds of features have been used as the basis for shape retrieval from image databases. This paper investigates the relative effectiveness of several types of global shape feature, both singly and in combination. The features compared include well-established descriptors such as Fourier coefficients and moment invariants, as well as recently-proposed measures of triangularity and ellipticity. Experiments were conducted within the framework of the ARTISAN shape retrieval system, and retrieval effectiveness assessed on a database of over 10,000 images, using 24 queries and associated ground truth supplied by the UK Patent Office . Our experiments revealed only minor differences in retrieval effectiveness between different measures, suggesting that a wide variety of shape feature combinations can provide adequate discriminating power for effective shape retrieval in multi-component image collections such as trademark registries. Marked differences between measures were observed for some individual queries, suggesting that there could be considerable scope for improving retrieval effectiveness by providing users with an improved framework for searching multi-dimensional feature space.
NASA Astrophysics Data System (ADS)
Eakins, John P.; Edwards, Jonathan D.; Riley, K. Jonathan; Rosin, Paul L.
2000-12-01
Many different kinds of features have been used as the basis for shape retrieval from image databases. This paper investigates the relative effectiveness of several types of global shape feature, both singly and in combination. The features compared include well-established descriptors such as Fourier coefficients and moment invariants, as well as recently-proposed measures of triangularity and ellipticity. Experiments were conducted within the framework of the ARTISAN shape retrieval system, and retrieval effectiveness assessed on a database of over 10,000 images, using 24 queries and associated ground truth supplied by the UK Patent Office . Our experiments revealed only minor differences in retrieval effectiveness between different measures, suggesting that a wide variety of shape feature combinations can provide adequate discriminating power for effective shape retrieval in multi-component image collections such as trademark registries. Marked differences between measures were observed for some individual queries, suggesting that there could be considerable scope for improving retrieval effectiveness by providing users with an improved framework for searching multi-dimensional feature space.
Method and system for data clustering for very large databases
NASA Technical Reports Server (NTRS)
Livny, Miron (Inventor); Zhang, Tian (Inventor); Ramakrishnan, Raghu (Inventor)
1998-01-01
Multi-dimensional data contained in very large databases is efficiently and accurately clustered to determine patterns therein and extract useful information from such patterns. Conventional computer processors may be used which have limited memory capacity and conventional operating speed, allowing massive data sets to be processed in a reasonable time and with reasonable computer resources. The clustering process is organized using a clustering feature tree structure wherein each clustering feature comprises the number of data points in the cluster, the linear sum of the data points in the cluster, and the square sum of the data points in the cluster. A dense region of data points is treated collectively as a single cluster, and points in sparsely occupied regions can be treated as outliers and removed from the clustering feature tree. The clustering can be carried out continuously with new data points being received and processed, and with the clustering feature tree being restructured as necessary to accommodate the information from the newly received data points.
Task-Driven Dynamic Text Summarization
ERIC Educational Resources Information Center
Workman, Terri Elizabeth
2011-01-01
The objective of this work is to examine the efficacy of natural language processing (NLP) in summarizing bibliographic text for multiple purposes. Researchers have noted the accelerating growth of bibliographic databases. Information seekers using traditional information retrieval techniques when searching large bibliographic databases are often…
Multidimensional poverty, household environment and short-term morbidity in India.
Dehury, Bidyadhar; Mohanty, Sanjay K
2017-01-01
Using the unit data from the second round of the Indian Human Development Survey (IHDS-II), 2011-2012, which covered 42,152 households, this paper examines the association between multidimensional poverty, household environmental deprivation and short-term morbidities (fever, cough and diarrhoea) in India. Poverty is measured in a multidimensional framework that includes the dimensions of education, health and income, while household environmental deprivation is defined as lack of access to improved sanitation, drinking water and cooking fuel. A composite index combining multidimensional poverty and household environmental deprivation has been computed, and households are classified as follows: multidimensional poor and living in a poor household environment, multidimensional non-poor and living in a poor household environment, multidimensional poor and living in a good household environment and multidimensional non-poor and living in a good household environment. Results suggest that about 23% of the population belonging to multidimensional poor households and living in a poor household environment had experienced short-term morbidities in a reference period of 30 days compared to 20% of the population belonging to multidimensional non-poor households and living in a poor household environment, 19% of the population belonging to multidimensional poor households and living in a good household environment and 15% of the population belonging to multidimensional non-poor households and living in a good household environment. Controlling for socioeconomic covariates, the odds of short-term morbidity was 1.47 [CI 1.40-1.53] among the multidimensional poor and living in a poor household environment, 1.28 [CI 1.21-1.37] among the multidimensional non-poor and living in a poor household environment and 1.21 [CI 1.64-1.28] among the multidimensional poor and living in a good household environment compared to the multidimensional non-poor and living in a good household environment. Results are robust across states and hold good for each of the three morbidities: fever, cough and diarrhoea. This establishes that along with poverty, household environmental conditions have a significant bearing on short-term morbidities in India. Public investment in sanitation, drinking water and cooking fuel can reduce the morbidity and improve the health of the population.
ERIC Educational Resources Information Center
Kollegger, James G.; And Others
1988-01-01
In the first of three articles, the producer of Energyline, Energynet, and Tele/Scope recalls the development of the databases and database business strategies. The second describes the development of biomedical online databases, and the third discusses future developments, including full text databases, database producers as online host, and…
A Query Expansion Framework in Image Retrieval Domain Based on Local and Global Analysis
Rahman, M. M.; Antani, S. K.; Thoma, G. R.
2011-01-01
We present an image retrieval framework based on automatic query expansion in a concept feature space by generalizing the vector space model of information retrieval. In this framework, images are represented by vectors of weighted concepts similar to the keyword-based representation used in text retrieval. To generate the concept vocabularies, a statistical model is built by utilizing Support Vector Machine (SVM)-based classification techniques. The images are represented as “bag of concepts” that comprise perceptually and/or semantically distinguishable color and texture patches from local image regions in a multi-dimensional feature space. To explore the correlation between the concepts and overcome the assumption of feature independence in this model, we propose query expansion techniques in the image domain from a new perspective based on both local and global analysis. For the local analysis, the correlations between the concepts based on the co-occurrence pattern, and the metrical constraints based on the neighborhood proximity between the concepts in encoded images, are analyzed by considering local feedback information. We also analyze the concept similarities in the collection as a whole in the form of a similarity thesaurus and propose an efficient query expansion based on the global analysis. The experimental results on a photographic collection of natural scenes and a biomedical database of different imaging modalities demonstrate the effectiveness of the proposed framework in terms of precision and recall. PMID:21822350
Guhlin, Joseph; Silverstein, Kevin A T; Zhou, Peng; Tiffin, Peter; Young, Nevin D
2017-08-10
Rapid generation of omics data in recent years have resulted in vast amounts of disconnected datasets without systemic integration and knowledge building, while individual groups have made customized, annotated datasets available on the web with few ways to link them to in-lab datasets. With so many research groups generating their own data, the ability to relate it to the larger genomic and comparative genomic context is becoming increasingly crucial to make full use of the data. The Omics Database Generator (ODG) allows users to create customized databases that utilize published genomics data integrated with experimental data which can be queried using a flexible graph database. When provided with omics and experimental data, ODG will create a comparative, multi-dimensional graph database. ODG can import definitions and annotations from other sources such as InterProScan, the Gene Ontology, ENZYME, UniPathway, and others. This annotation data can be especially useful for studying new or understudied species for which transcripts have only been predicted, and rapidly give additional layers of annotation to predicted genes. In better studied species, ODG can perform syntenic annotation translations or rapidly identify characteristics of a set of genes or nucleotide locations, such as hits from an association study. ODG provides a web-based user-interface for configuring the data import and for querying the database. Queries can also be run from the command-line and the database can be queried directly through programming language hooks available for most languages. ODG supports most common genomic formats as well as generic, easy to use tab-separated value format for user-provided annotations. ODG is a user-friendly database generation and query tool that adapts to the supplied data to produce a comparative genomic database or multi-layered annotation database. ODG provides rapid comparative genomic annotation and is therefore particularly useful for non-model or understudied species. For species for which more data are available, ODG can be used to conduct complex multi-omics, pattern-matching queries.
ERIC Educational Resources Information Center
Grooms, David W.
1988-01-01
Discusses the quality controls imposed on text and image data that is currently being converted from paper to digital images by the Patent and Trademark Office. The methods of inspection used on text and on images are described, and the quality of the data delivered thus far is discussed. (CLB)
Data Mining Research with the LSST
NASA Astrophysics Data System (ADS)
Borne, Kirk D.; Strauss, M. A.; Tyson, J. A.
2007-12-01
The LSST catalog database will exceed 10 petabytes, comprising several hundred attributes for 5 billion galaxies, 10 billion stars, and over 1 billion variable sources (optical variables, transients, or moving objects), extracted from over 20,000 square degrees of deep imaging in 5 passbands with thorough time domain coverage: 1000 visits over the 10-year LSST survey lifetime. The opportunities are enormous for novel scientific discoveries within this rich time-domain ultra-deep multi-band survey database. Data Mining, Machine Learning, and Knowledge Discovery research opportunities with the LSST are now under study, with a potential for new collaborations to develop to contribute to these investigations. We will describe features of the LSST science database that are amenable to scientific data mining, object classification, outlier identification, anomaly detection, image quality assurance, and survey science validation. We also give some illustrative examples of current scientific data mining research in astronomy, and point out where new research is needed. In particular, the data mining research community will need to address several issues in the coming years as we prepare for the LSST data deluge. The data mining research agenda includes: scalability (at petabytes scales) of existing machine learning and data mining algorithms; development of grid-enabled parallel data mining algorithms; designing a robust system for brokering classifications from the LSST event pipeline (which may produce 10,000 or more event alerts per night); multi-resolution methods for exploration of petascale databases; visual data mining algorithms for visual exploration of the data; indexing of multi-attribute multi-dimensional astronomical databases (beyond RA-Dec spatial indexing) for rapid querying of petabyte databases; and more. Finally, we will identify opportunities for synergistic collaboration between the data mining research group and the LSST Data Management and Science Collaboration teams.
Multidimensional chromatography in food analysis.
Herrero, Miguel; Ibáñez, Elena; Cifuentes, Alejandro; Bernal, Jose
2009-10-23
In this work, the main developments and applications of multidimensional chromatographic techniques in food analysis are reviewed. Different aspects related to the existing couplings involving chromatographic techniques are examined. These couplings include multidimensional GC, multidimensional LC, multidimensional SFC as well as all their possible combinations. Main advantages and drawbacks of each coupling are critically discussed and their key applications in food analysis described.
NASA Astrophysics Data System (ADS)
Balsara, Dinshaw S.; Nkonga, Boniface
2017-10-01
Just as the quality of a one-dimensional approximate Riemann solver is improved by the inclusion of internal sub-structure, the quality of a multidimensional Riemann solver is also similarly improved. Such multidimensional Riemann problems arise when multiple states come together at the vertex of a mesh. The interaction of the resulting one-dimensional Riemann problems gives rise to a strongly-interacting state. We wish to endow this strongly-interacting state with physically-motivated sub-structure. The fastest way of endowing such sub-structure consists of making a multidimensional extension of the HLLI Riemann solver for hyperbolic conservation laws. Presenting such a multidimensional analogue of the HLLI Riemann solver with linear sub-structure for use on structured meshes is the goal of this work. The multidimensional MuSIC Riemann solver documented here is universal in the sense that it can be applied to any hyperbolic conservation law. The multidimensional Riemann solver is made to be consistent with constraints that emerge naturally from the Galerkin projection of the self-similar states within the wave model. When the full eigenstructure in both directions is used in the present Riemann solver, it becomes a complete Riemann solver in a multidimensional sense. I.e., all the intermediate waves are represented in the multidimensional wave model. The work also presents, for the very first time, an important analysis of the dissipation characteristics of multidimensional Riemann solvers. The present Riemann solver results in the most efficient implementation of a multidimensional Riemann solver with sub-structure. Because it preserves stationary linearly degenerate waves, it might also help with well-balancing. Implementation-related details are presented in pointwise fashion for the one-dimensional HLLI Riemann solver as well as the multidimensional MuSIC Riemann solver.
Zanni, Martin Thomas; Damrauer, Niels H.
2010-07-20
A multidimensional spectrometer for the infrared, visible, and ultraviolet regions of the electromagnetic spectrum, and a method for making multidimensional spectroscopic measurements in the infrared, visible, and ultraviolet regions of the electromagnetic spectrum. The multidimensional spectrometer facilitates measurements of inter- and intra-molecular interactions.
Clark, A. S.; Shea, S.
1991-01-01
The use of Folio Views, a PC DOS based product for free text databases, is explored in three applications in an Integrated Academic Information System (IAIMS): (1) a telephone directory, (2) a grants and contracts newsletter, and (3) nursing care plans. PMID:1666967
The Impact of Online Bibliographic Databases on Teaching and Research in Political Science.
ERIC Educational Resources Information Center
Reichel, Mary
The availability of online bibliographic databases greatly facilitates literature searching in political science. The advantages to searching databases online include combination of concepts, comprehensiveness, multiple database searching, free-text searching, currency, current awareness services, document delivery service, and convenience.…
Knowns and unknowns in metabolomics identified by multidimensional NMR and hybrid MS/NMR methods
DOE Office of Scientific and Technical Information (OSTI.GOV)
Bingol, Kerem; Brüschweiler, Rafael
Metabolomics continues to make rapid progress through the development of new and better methods and their applications to gain insight into the metabolism of a wide range of different biological systems from a systems biology perspective. Customization of NMR databases and search tools allows the faster and more accurate identification of known metabolites, whereas the identification of unknowns, without a need for extensive purification, requires new strategies to integrate NMR with mass spectrometry, cheminformatics, and computational methods. For some applications, the use of covalent and non-covalent attachments in the form of labeled tags or nanoparticles can significantly reduce the complexitymore » of these tasks.« less
High Performance Analytics with the R3-Cache
NASA Astrophysics Data System (ADS)
Eavis, Todd; Sayeed, Ruhan
Contemporary data warehouses now represent some of the world’s largest databases. As these systems grow in size and complexity, however, it becomes increasingly difficult for brute force query processing approaches to meet the performance demands of end users. Certainly, improved indexing and more selective view materialization are helpful in this regard. Nevertheless, with warehouses moving into the multi-terabyte range, it is clear that the minimization of external memory accesses must be a primary performance objective. In this paper, we describe the R 3-cache, a natively multi-dimensional caching framework designed specifically to support sophisticated warehouse/OLAP environments. R 3-cache is based upon an in-memory version of the R-tree that has been extended to support buffer pages rather than disk blocks. A key strength of the R 3-cache is that it is able to utilize multi-dimensional fragments of previous query results so as to significantly minimize the frequency and scale of disk accesses. Moreover, the new caching model directly accommodates the standard relational storage model and provides mechanisms for pro-active updates that exploit the existence of query “hot spots”. The current prototype has been evaluated as a component of the Sidera DBMS, a “shared nothing” parallel OLAP server designed for multi-terabyte analytics. Experimental results demonstrate significant performance improvements relative to simpler alternatives.
Multidimensional assessment of self-regulated learning with middle school math students.
Callan, Gregory L; Cleary, Timothy J
2018-03-01
This study examined the convergent and predictive validity of self-regulated learning (SRL) measures situated in mathematics. The sample included 100 eighth graders from a diverse, urban school district. Four measurement formats were examined including, 2 broad-based (i.e., self-report questionnaire and teacher ratings) and 2 task-specific measures (i.e., SRL microanalysis and behavioral traces). Convergent validity was examined across task-difficulty, and the predictive validity was examined across 3 mathematics outcomes: 2 measures of mathematical problem solving skill (i.e., practice session math problems, posttest math problems) and a global measure of mathematical skill (i.e., standardized math test). Correlation analyses were used to examine convergent validity and revealed medium correlations between measures within the same category (i.e., broad-based or task-specific). Relations between measurement classes were not statistically significant. Separate regressions examined the predictive validity of the SRL measures. While controlling all other predictors, a SRL microanalysis metacognitive-monitoring measure emerged as a significant predictor of all 3 outcomes and teacher ratings accounted for unique variance on 2 of the outcomes (i.e., posttest math problems and standardized math test). Results suggest that a multidimensional assessment approach should be considered by school psychologists interested in measuring SRL. (PsycINFO Database Record (c) 2018 APA, all rights reserved).
Kiranyaz, Serkan; Ince, Turker; Pulkkinen, Jenni; Gabbouj, Moncef
2010-01-01
In this paper, we address dynamic clustering in high dimensional data or feature spaces as an optimization problem where multi-dimensional particle swarm optimization (MD PSO) is used to find out the true number of clusters, while fractional global best formation (FGBF) is applied to avoid local optima. Based on these techniques we then present a novel and personalized long-term ECG classification system, which addresses the problem of labeling the beats within a long-term ECG signal, known as Holter register, recorded from an individual patient. Due to the massive amount of ECG beats in a Holter register, visual inspection is quite difficult and cumbersome, if not impossible. Therefore the proposed system helps professionals to quickly and accurately diagnose any latent heart disease by examining only the representative beats (the so called master key-beats) each of which is representing a cluster of homogeneous (similar) beats. We tested the system on a benchmark database where the beats of each Holter register have been manually labeled by cardiologists. The selection of the right master key-beats is the key factor for achieving a highly accurate classification and the proposed systematic approach produced results that were consistent with the manual labels with 99.5% average accuracy, which basically shows the efficiency of the system.
Packer, Tanya L; Fracini, America; Audulv, Åsa; Alizadeh, Neda; van Gaal, Betsie G I; Warner, Grace; Kephart, George
2018-04-01
To identify self-report, self-management measures for adults with chronic conditions, and describe their purpose, theoretical foundation, dimensionality (multi versus uni), and scope (generic versus condition specific). A search of four databases (8479 articles) resulted in a scoping review of 28 self-management measures. Although authors identified tools as measures of self-management, wide variation in constructs measured, purpose, and theoretical foundations existed. Subscales on 13 multidimensional tools collectively measure domains of self-management relevant to clients, however no one tool's subscales cover all domains. Viewing self-management as a complex, multidimensional whole, demonstrated that existing measures assess different, related aspects of self-management. Activities and social roles, though important to patients, are rarely measured. Measures with capacity to quantify and distinguish aspects of self-management may promote tailored patient care. In selecting tools for research or assessment, the reason for development, definitions, and theories underpinning the measure should be scrutinized. Our ability to measure self-management must be rigorously mapped to provide comprehensive and system-wide care for clients with chronic conditions. Viewing self-management as a complex whole will help practitioners to understand the patient perspective and their contribution in supporting each individual patient. Copyright © 2017 Elsevier B.V. All rights reserved.
Numeric invariants from multidimensional persistence
DOE Office of Scientific and Technical Information (OSTI.GOV)
Skryzalin, Jacek; Carlsson, Gunnar
2017-05-19
In this paper, we analyze the space of multidimensional persistence modules from the perspectives of algebraic geometry. We first build a moduli space of a certain subclass of easily analyzed multidimensional persistence modules, which we construct specifically to capture much of the information which can be gained by using multidimensional persistence over one-dimensional persistence. We argue that the global sections of this space provide interesting numeric invariants when evaluated against our subclass of multidimensional persistence modules. Lastly, we extend these global sections to the space of all multidimensional persistence modules and discuss how the resulting numeric invariants might be usedmore » to study data.« less
Lowe, H. J.
1993-01-01
This paper describes Image Engine, an object-oriented, microcomputer-based, multimedia database designed to facilitate the storage and retrieval of digitized biomedical still images, video, and text using inexpensive desktop computers. The current prototype runs on Apple Macintosh computers and allows network database access via peer to peer file sharing protocols. Image Engine supports both free text and controlled vocabulary indexing of multimedia objects. The latter is implemented using the TView thesaurus model developed by the author. The current prototype of Image Engine uses the National Library of Medicine's Medical Subject Headings (MeSH) vocabulary (with UMLS Meta-1 extensions) as its indexing thesaurus. PMID:8130596
BDVC (Bimodal Database of Violent Content): A database of violent audio and video
NASA Astrophysics Data System (ADS)
Rivera Martínez, Jose Luis; Mijes Cruz, Mario Humberto; Rodríguez Vázqu, Manuel Antonio; Rodríguez Espejo, Luis; Montoya Obeso, Abraham; García Vázquez, Mireya Saraí; Ramírez Acosta, Alejandro Álvaro
2017-09-01
Nowadays there is a trend towards the use of unimodal databases for multimedia content description, organization and retrieval applications of a single type of content like text, voice and images, instead bimodal databases allow to associate semantically two different types of content like audio-video, image-text, among others. The generation of a bimodal database of audio-video implies the creation of a connection between the multimedia content through the semantic relation that associates the actions of both types of information. This paper describes in detail the used characteristics and methodology for the creation of the bimodal database of violent content; the semantic relationship is stablished by the proposed concepts that describe the audiovisual information. The use of bimodal databases in applications related to the audiovisual content processing allows an increase in the semantic performance only and only if these applications process both type of content. This bimodal database counts with 580 audiovisual annotated segments, with a duration of 28 minutes, divided in 41 classes. Bimodal databases are a tool in the generation of applications for the semantic web.
Urbanski, William M; Condie, Brian G
2009-12-01
Textpresso Site Specific Recombinases (http://ssrc.genetics.uga.edu/) is a text-mining web server for searching a database of more than 9,000 full-text publications. The papers and abstracts in this database represent a wide range of topics related to site-specific recombinase (SSR) research tools. Included in the database are most of the papers that report the characterization or use of mouse strains that express Cre recombinase as well as papers that describe or analyze mouse lines that carry conditional (floxed) alleles or SSR-activated transgenes/knockins. The database also includes reports describing SSR-based cloning methods such as the Gateway or the Creator systems, papers reporting the development or use of SSR-based tools in systems such as Drosophila, bacteria, parasites, stem cells, yeast, plants, zebrafish, and Xenopus as well as publications that describe the biochemistry, genetics, or molecular structure of the SSRs themselves. Textpresso Site Specific Recombinases is the only comprehensive text-mining resource available for the literature describing the biology and technical applications of SSRs. (c) 2009 Wiley-Liss, Inc.
Text mining for metabolic pathways, signaling cascades, and protein networks.
Hoffmann, Robert; Krallinger, Martin; Andres, Eduardo; Tamames, Javier; Blaschke, Christian; Valencia, Alfonso
2005-05-10
The complexity of the information stored in databases and publications on metabolic and signaling pathways, the high throughput of experimental data, and the growing number of publications make it imperative to provide systems to help the researcher navigate through these interrelated information resources. Text-mining methods have started to play a key role in the creation and maintenance of links between the information stored in biological databases and its original sources in the literature. These links will be extremely useful for database updating and curation, especially if a number of technical problems can be solved satisfactorily, including the identification of protein and gene names (entities in general) and the characterization of their types of interactions. The first generation of openly accessible text-mining systems, such as iHOP (Information Hyperlinked over Proteins), provides additional functions to facilitate the reconstruction of protein interaction networks, combine database and text information, and support the scientist in the formulation of novel hypotheses. The next challenge is the generation of comprehensive information regarding the general function of signaling pathways and protein interaction networks.
Drowning in Data: Sorting through CD ROM and Computer Databases.
ERIC Educational Resources Information Center
Cates, Carl M.; Kaye, Barbara K.
This paper identifies the bibliographic and numeric databases on CD-ROM and computer diskette that should be most useful for investigators in communication, marketing, and communication education. Bibliographic databases are usually found in three formats: citations only, citations and abstracts, and full-text articles. Numeric databases are…
ERIC Educational Resources Information Center
Chen, Ping
2017-01-01
Calibration of new items online has been an important topic in item replenishment for multidimensional computerized adaptive testing (MCAT). Several online calibration methods have been proposed for MCAT, such as multidimensional "one expectation-maximization (EM) cycle" (M-OEM) and multidimensional "multiple EM cycles"…
Best Design for Multidimensional Computerized Adaptive Testing with the Bifactor Model
ERIC Educational Resources Information Center
Seo, Dong Gi; Weiss, David J.
2015-01-01
Most computerized adaptive tests (CATs) have been studied using the framework of unidimensional item response theory. However, many psychological variables are multidimensional and might benefit from using a multidimensional approach to CATs. This study investigated the accuracy, fidelity, and efficiency of a fully multidimensional CAT algorithm…
Semantic Annotation of Complex Text Structures in Problem Reports
NASA Technical Reports Server (NTRS)
Malin, Jane T.; Throop, David R.; Fleming, Land D.
2011-01-01
Text analysis is important for effective information retrieval from databases where the critical information is embedded in text fields. Aerospace safety depends on effective retrieval of relevant and related problem reports for the purpose of trend analysis. The complex text syntax in problem descriptions has limited statistical text mining of problem reports. The presentation describes an intelligent tagging approach that applies syntactic and then semantic analysis to overcome this problem. The tags identify types of problems and equipment that are embedded in the text descriptions. The power of these tags is illustrated in a faceted searching and browsing interface for problem report trending that combines automatically generated tags with database code fields and temporal information.
Array Processing in the Cloud: the rasdaman Approach
NASA Astrophysics Data System (ADS)
Merticariu, Vlad; Dumitru, Alex
2015-04-01
The multi-dimensional array data model is gaining more and more attention when dealing with Big Data challenges in a variety of domains such as climate simulations, geographic information systems, medical imaging or astronomical observations. Solutions provided by classical Big Data tools such as Key-Value Stores and MapReduce, as well as traditional relational databases, proved to be limited in domains associated with multi-dimensional data. This problem has been addressed by the field of array databases, in which systems provide database services for raster data, without imposing limitations on the number of dimensions that a dataset can have. Examples of datasets commonly handled by array databases include 1-dimensional sensor data, 2-D satellite imagery, 3-D x/y/t image time series as well as x/y/z geophysical voxel data, and 4-D x/y/z/t weather data. And this can grow as large as simulations of the whole universe when it comes to astrophysics. rasdaman is a well established array database, which implements many optimizations for dealing with large data volumes and operation complexity. Among those, the latest one is intra-query parallelization support: a network of machines collaborate for answering a single array database query, by dividing it into independent sub-queries sent to different servers. This enables massive processing speed-ups, which promise solutions to research challenges on multi-Petabyte data cubes. There are several correlated factors which influence the speedup that intra-query parallelisation brings: the number of servers, the capabilities of each server, the quality of the network, the availability of the data to the server that needs it in order to compute the result and many more. In the effort of adapting the engine to cloud processing patterns, two main components have been identified: one that handles communication and gathers information about the arrays sitting on every server, and a processing unit responsible with dividing work among available nodes and executing operations on local data. The federation daemon collects and stores statistics from the other network nodes and provides real time updates about local changes. Information exchanged includes available datasets, CPU load and memory usage per host. The processing component is represented by the rasdaman server. Using information from the federation daemon it breaks queries into subqueries to be executed on peer nodes, ships them, and assembles the intermediate results. Thus, we define a rasdaman network node as a pair of a federation daemon and a rasdaman server. Any node can receive a query and will subsequently act as this query's dispatcher, so all peers are at the same level and there is no single point of failure. Should a node become inaccessible then the peers will recognize this and will not any longer consider this peer for distribution. Conversely, a peer at any time can join the network. To assess the feasibility of our approach, we deployed a rasdaman network in the Amazon Elastic Cloud environment on 1001 nodes, and observed that this feature can greatly increase the performance and scalability of the system, offering a large throughput of processed data.
ERIC Educational Resources Information Center
Qin, Jian; Jurisica, Igor; Liddy, Elizabeth D.; Jansen, Bernard J; Spink, Amanda; Priss, Uta; Norton, Melanie J.
2000-01-01
These six articles discuss knowledge discovery in databases (KDD). Topics include data mining; knowledge management systems; applications of knowledge discovery; text and Web mining; text mining and information retrieval; user search patterns through Web log analysis; concept analysis; data collection; and data structure inconsistency. (LRW)
Hong, Mei
2016-01-01
We have determined refined multidimensional chemical shift ranges for intra-residue correlations (13C–13C, 15N–13C, etc.) in proteins, which can be used to gain type-assignment and/or secondary-structure information from experimental NMR spectra. The chemical-shift ranges are the result of a statistical analysis of the PACSY database of >3000 proteins with 3D structures (1,200,207 13C chemical shifts and >3 million chemical shifts in total); these data were originally derived from the Biological Magnetic Resonance Data Bank. Using relatively simple non-parametric statistics to find peak maxima in the distributions of helix, sheet, coil and turn chemical shifts, and without the use of limited “hand-picked” data sets, we show that ~94 % of the 13C NMR data and almost all 15N data are quite accurately referenced and assigned, with smaller standard deviations (0.2 and 0.8 ppm, respectively) than recognized previously. On the other hand, approximately 6 % of the 13C chemical shift data in the PACSY database are shown to be clearly misreferenced, mostly by ca. −2.4 ppm. The removal of the misreferenced data and other outliers by this purging by intrinsic quality criteria (PIQC) allows for reliable identification of secondary maxima in the two-dimensional chemical-shift distributions already pre-separated by secondary structure. We demonstrate that some of these correspond to specific regions in the Ramachandran plot, including left-handed helix dihedral angles, reflect unusual hydrogen bonding, or are due to the influence of a following proline residue. With appropriate smoothing, significantly more tightly defined chemical shift ranges are obtained for each amino acid type in the different secondary structures. These chemical shift ranges, which may be defined at any statistical threshold, can be used for amino-acid type assignment and secondary-structure analysis of chemical shifts from intra-residue cross peaks by inspection or by using a provided command-line Python script (PLUQin), which should be useful in protein structure determination. The refined chemical shift distributions are utilized in a simple quality test (SQAT) that should be applied to new protein NMR data before deposition in a databank, and they could benefit many other chemical-shift based tools. PMID:26787537
Resistance exercise training for fibromyalgia.
Busch, Angela J; Webber, Sandra C; Richards, Rachel S; Bidonde, Julia; Schachter, Candice L; Schafer, Laurel A; Danyliw, Adrienne; Sawant, Anuradha; Dal Bello-Haas, Vanina; Rader, Tamara; Overend, Tom J
2013-12-20
Fibromyalgia is characterized by chronic widespread pain that leads to reduced physical function. Exercise training is commonly recommended as a treatment for management of symptoms. We examined the literature on resistance training for individuals with fibromyalgia. Resistance training is exercise performed against a progressive resistance with the intention of improving muscle strength, muscle endurance, muscle power, or a combination of these. To evaluate the benefits and harms of resistance exercise training in adults with fibromyalgia. We compared resistance training versus control and versus other types of exercise training. We searched nine electronic databases (The Cochrane Library, MEDLINE, EMBASE, CINAHL, PEDro, Dissertation Abstracts, Current Controlled Trials, World Health Organization (WHO) International Clinical Trials Registry Platform, AMED) and other sources for published full-text articles. The date of the last search was 5 March 2013. Two review authors independently screened 1856 citations, 766 abstracts and 156 full-text articles. We included five studies that met our inclusion criteria. Selection criteria included: a) randomized clinical trial, b) diagnosis of fibromyalgia based on published criteria, c) adult sample, d) full-text publication, and e) inclusion of between-group data comparing resistance training versus a control or other physical activity intervention. Pairs of review authors independently assessed risk of bias and extracted intervention and outcome data. We resolved disagreements between the two review authors and questions regarding interpretation of study methods by discussion within the pairs or when necessary the issue was taken to the full team of 11 members. We extracted 21 outcomes of which seven were designated as major outcomes: multidimensional function, self reported physical function, pain, tenderness, muscle strength, attrition rates, and adverse effects. We evaluated benefits and harms of the interventions using standardized mean differences (SMD) or mean differences (MD) or risk ratios or Peto odds ratios and 95% confidence intervals (CI). Where two or more studies provided data for an outcome, we carried out a meta-analysis. The literature search yielded 1865 citations with five studies meeting the selection criteria. One of the studies that had three arms contributed data for two comparisons. In the included studies, there were 219 women participants with fibromyalgia, 95 of whom were assigned to resistance training programs. Three randomized trials compared 16 to 21 weeks of moderate- to high-intensity resistance training versus a control group. Two studies compared eight weeks of progressive resistance training (intensity as tolerated) using free weights or body weight resistance exercise versus aerobic training (ie, progressive treadmill walking, indoor and outdoor walking), and one study compared 12 weeks of low-intensity resistance training using hand weights (1 to 3 lbs (0.45 to 1.36 kg)) and elastic tubing versus flexibility exercise (static stretches to major muscle groups).Statistically significant differences (MD; 95% CI) favoring the resistance training interventions over control group(s) were found in multidimensional function (Fibromyalgia Impact Questionnaire (FIQ) total decreased 16.75 units on a 100-point scale; 95% CI -23.31 to -10.19), self reported physical function (-6.29 units on a 100-point scale; 95% CI -10.45 to -2.13), pain (-3.3 cm on a 10-cm scale; 95% CI -6.35 to -0.26), tenderness (-1.84 out of 18 tender points; 95% CI -2.6 to -1.08), and muscle strength (27.32 kg force on bilateral concentric leg extension; 95% CI 18.28 to 36.36).Differences between the resistance training group(s) and the aerobic training groups were not statistically significant for multidimensional function (5.48 on a 100-point scale; 95% CI -0.92 to 11.88), self reported physical function (-1.48 units on a 100-point scale; 95% CI -6.69 to 3.74) or tenderness (SMD -0.13; 95% CI -0.55 to 0.30). There was a statistically significant reduction in pain (0.99 cm on a 10-cm scale; 95% CI 0.31 to 1.67) favoring the aerobic groups.Statistically significant differences were found between a resistance training group and a flexibility group favoring the resistance training group for multidimensional function (-6.49 FIQ units on a 100-point scale; 95% CI -12.57 to -0.41) and pain (-0.88 cm on a 10-cm scale; 95% CI -1.57 to -0.19), but not for tenderness (-0.46 out of 18 tender points; 95% CI -1.56 to 0.64) or strength (4.77 foot pounds torque on concentric knee extension; 95% CI -2.40 to 11.94). This evidence was classified low quality due to the low number of studies and risk of bias assessment. There were no statistically significant differences in attrition rates between the interventions. In general, adverse effects were poorly recorded, but no serious adverse effects were reported. Assessment of risk of bias was hampered by poor written descriptions (eg, allocation concealment, blinding of outcome assessors). The lack of a priori protocols and lack of care provider blinding were also identified as methodologic concerns. The evidence (rated as low quality) suggested that moderate- and moderate- to high-intensity resistance training improves multidimensional function, pain, tenderness, and muscle strength in women with fibromyalgia. The evidence (rated as low quality) also suggested that eight weeks of aerobic exercise was superior to moderate-intensity resistance training for improving pain in women with fibromyalgia. There was low-quality evidence that 12 weeks of low-intensity resistance training was superior to flexibility exercise training in women with fibromyalgia for improvements in pain and multidimensional function. There was low-quality evidence that women with fibromyalgia can safely perform moderate- to high-resistance training.
Accessing Multi-Dimensional Images and Data Cubes in the Virtual Observatory
NASA Astrophysics Data System (ADS)
Tody, Douglas; Plante, R. L.; Berriman, G. B.; Cresitello-Dittmar, M.; Good, J.; Graham, M.; Greene, G.; Hanisch, R. J.; Jenness, T.; Lazio, J.; Norris, P.; Pevunova, O.; Rots, A. H.
2014-01-01
Telescopes across the spectrum are routinely producing multi-dimensional images and datasets, such as Doppler velocity cubes, polarization datasets, and time-resolved “movies.” Examples of current telescopes producing such multi-dimensional images include the JVLA, ALMA, and the IFU instruments on large optical and near-infrared wavelength telescopes. In the near future, both the LSST and JWST will also produce such multi-dimensional images routinely. High-energy instruments such as Chandra produce event datasets that are also a form of multi-dimensional data, in effect being a very sparse multi-dimensional image. Ensuring that the data sets produced by these telescopes can be both discovered and accessed by the community is essential and is part of the mission of the Virtual Observatory (VO). The Virtual Astronomical Observatory (VAO, http://www.usvao.org/), in conjunction with its international partners in the International Virtual Observatory Alliance (IVOA), has developed a protocol and an initial demonstration service designed for the publication, discovery, and access of arbitrarily large multi-dimensional images. The protocol describing multi-dimensional images is the Simple Image Access Protocol, version 2, which provides the minimal set of metadata required to characterize a multi-dimensional image for its discovery and access. A companion Image Data Model formally defines the semantics and structure of multi-dimensional images independently of how they are serialized, while providing capabilities such as support for sparse data that are essential to deal effectively with large cubes. A prototype data access service has been deployed and tested, using a suite of multi-dimensional images from a variety of telescopes. The prototype has demonstrated the capability to discover and remotely access multi-dimensional data via standard VO protocols. The prototype informs the specification of a protocol that will be submitted to the IVOA for approval, with an operational data cube service to be delivered in mid-2014. An associated user-installable VO data service framework will provide the capabilities required to publish VO-compatible multi-dimensional images or data cubes.
Teaching Advanced SQL Skills: Text Bulk Loading
ERIC Educational Resources Information Center
Olsen, David; Hauser, Karina
2007-01-01
Studies show that advanced database skills are important for students to be prepared for today's highly competitive job market. A common task for database administrators is to insert a large amount of data into a database. This paper illustrates how an up-to-date, advanced database topic, namely bulk insert, can be incorporated into a database…
A User's Applications of Imaging Techniques: The University of Maryland Historic Textile Database.
ERIC Educational Resources Information Center
Anderson, Clarita S.
1991-01-01
Describes the incorporation of textile images into the University of Maryland Historic Textile Database by a computer user rather than a computer expert. Selection of a database management system is discussed, and PICTUREPOWER, a system that integrates photographic quality images with text and numeric information in databases, is described. (three…
ERIC Educational Resources Information Center
Van Deun, Katrijn; Heiser, Willem J.; Delbeke, Luc
2007-01-01
A multidimensional unfolding technique that is not prone to degenerate solutions and is based on multidimensional scaling of a complete data matrix is proposed: distance information about the unfolding data and about the distances both among judges and among objects is included in the complete matrix. The latter information is derived from the…
Arlet, Vincent; Shilt, Jeffrey; Bersusky, Ernesto; Abel, Mark; Ouellet, Jean Albert; Evans, Davis; Menon, K V; Kandziora, Frank; Shen, Frank; Lamartina, Claudio; Adams, Marc; Reddi, Vasantha
2008-11-01
Considerable variability exists in the surgical treatment and outcomes of adolescent idiopathic scoliosis (AIS). This is due to the lack of evidence-based treatment guidelines and outcome measures. Although clinical trials have been extolled as the highest form of evidence for evaluating treatment efficacy, the disadvantage of cost, time, lack of feasibility, and ethical considerations indicate a need for a new paradigm for evidence based research in this spinal deformity. High quality clinical databases offer an alternative approach for evidence-based research in medicine. So, we developed and established Scolisoft, an international, multidimensional and relational database designed to be a repository of surgical cases for AIS, and an active vehicle for standardized surgical information in a format that would permit qualitative and quantitative research and analysis. Here, we describe and discuss the utility of Scolisoft as a new paradigm for evidence-based research on AIS. Scolisoft was developed using dot.net platform and SQL server from Microsoft. All data is deidentified to protect patient privacy. Scolisoft can be accessed at (www.scolisoft.org). Collection of high quality data on surgical cases of AIS is a priority and processes continue to improve the database quality. The database currently has 67 registered users from 21 countries. To date, Scolisoft has 200 detailed surgical cases with pre, post, and follow up data. Scolisoft provides a structured process and practical information for surgeons to benchmark their treatment methods against other like treatments. Scolisoft is multifaceted and its use extends to education of health care providers in training, patients, ability to mine important data to stimulate research and quality improvement initiatives of healthcare organizations.
NASA Astrophysics Data System (ADS)
Montalto, F. A.; Yu, Z.; Soldner, K.; Israel, A.; Fritch, M.; Kim, Y.; White, S.
2017-12-01
Urban stormwater utilities are increasingly using decentralized "green" infrastructure (GI) systems to capture stormwater and achieve compliance with regulations. Because environmental conditions, and design varies by GSI facility, monitoring of GSI systems under a range of conditions is essential. Conventional monitoring efforts can be costly because in-field data logging requires intense data transmission rates. The Internet of Things (IoT) can be used to more cost-effectively collect, store, and publish GSI monitoring data. Using 3G mobile networks, a cloud-based database was built on an Amazon Web Services (AWS) EC2 virtual machine to store and publish data collected with environmental sensors deployed in the field. This database can store multi-dimensional time series data, as well as photos and other observations logged by citizen scientists through a public engagement mobile app through a new Application Programming Interface (API). Also on the AWS EC2 virtual machine, a real-time QAQC flagging algorithm was developed to validate the sensor data streams.
Dziuba, Bartłomiej; Dziuba, Marta
2014-08-20
New peptides with potential antimicrobial activity, encrypted in milk protein sequences, were searched for with the use of bioinformatic tools. The major milk proteins were hydrolyzed in silico by 28 enzymes. The obtained peptides were characterized by the following parameters: molecular weight, isoelectric point, composition and number of amino acid residues, net charge at pH 7.0, aliphatic index, instability index, Boman index, and GRAVY index, and compared with those calculated for known 416 antimicrobial peptides including 59 antimicrobial peptides (AMPs) from milk proteins listed in the BIOPEP database. A simple analysis of physico-chemical properties and the values of biological activity indicators were insufficient to select potentially antimicrobial peptides released in silico from milk proteins by proteolytic enzymes. The final selection was made based on the results of multidimensional statistical analysis such as support vector machines (SVM), random forest (RF), artificial neural networks (ANN) and discriminant analysis (DA) available in the Collection of Anti-Microbial Peptides (CAMP database). Eleven new peptides with potential antimicrobial activity were selected from all peptides released during in silico proteolysis of milk proteins.
Dziuba, Bartłomiej; Dziuba, Marta
2014-01-01
New peptides with potential antimicrobial activity, encrypted in milk protein sequences, were searched for with the use of bioinformatic tools. The major milk proteins were hydrolyzed in silico by 28 enzymes. The obtained peptides were characterized by the following parameters: molecular weight, isoelectric point, composition and number of amino acid residues, net charge at pH 7.0, aliphatic index, instability index, Boman index, and GRAVY index, and compared with those calculated for known 416 antimicrobial peptides including 59 antimicrobial peptides (AMPs) from milk proteins listed in the BIOPEP database. A simple analysis of physico-chemical properties and the values of biological activity indicators were insufficient to select potentially antimicrobial peptides released in silico from milk proteins by proteolytic enzymes. The final selection was made based on the results of multidimensional statistical analysis such as support vector machines (SVM), random forest (RF), artificial neural networks (ANN) and discriminant analysis (DA) available in the Collection of Anti-Microbial Peptides (CAMP database). Eleven new peptides with potential antimicrobial activity were selected from all peptides released during in silico proteolysis of milk proteins. PMID:25141106
NordicDB: a Nordic pool and portal for genome-wide control data.
Leu, Monica; Humphreys, Keith; Surakka, Ida; Rehnberg, Emil; Muilu, Juha; Rosenström, Päivi; Almgren, Peter; Jääskeläinen, Juha; Lifton, Richard P; Kyvik, Kirsten Ohm; Kaprio, Jaakko; Pedersen, Nancy L; Palotie, Aarno; Hall, Per; Grönberg, Henrik; Groop, Leif; Peltonen, Leena; Palmgren, Juni; Ripatti, Samuli
2010-12-01
A cost-efficient way to increase power in a genetic association study is to pool controls from different sources. The genotyping effort can then be directed to large case series. The Nordic Control database, NordicDB, has been set up as a unique resource in the Nordic area and the data are available for authorized users through the web portal (http://www.nordicdb.org). The current version of NordicDB pools together high-density genome-wide SNP information from ∼5000 controls originating from Finnish, Swedish and Danish studies and shows country-specific allele frequencies for SNP markers. The genetic homogeneity of the samples was investigated using multidimensional scaling (MDS) analysis and pairwise allele frequency differences between the studies. The plot of the first two MDS components showed excellent resemblance to the geographical placement of the samples, with a clear NW-SE gradient. We advise researchers to assess the impact of population structure when incorporating NordicDB controls in association studies. This harmonized Nordic database presents a unique genome-wide resource for future genetic association studies in the Nordic countries.
The Star Schema Benchmark and Augmented Fact Table Indexing
NASA Astrophysics Data System (ADS)
O'Neil, Patrick; O'Neil, Elizabeth; Chen, Xuedong; Revilak, Stephen
We provide a benchmark measuring star schema queries retrieving data from a fact table with Where clause column restrictions on dimension tables. Clustering is crucial to performance with modern disk technology, since retrievals with filter factors down to 0.0005 are now performed most efficiently by sequential table search rather than by indexed access. DB2’s Multi-Dimensional Clustering (MDC) provides methods to "dice" the fact table along a number of orthogonal "dimensions", but only when these dimensions are columns in the fact table. The diced cells cluster fact rows on several of these "dimensions" at once so queries restricting several such columns can access crucially localized data, with much faster query response. Unfortunately, columns of dimension tables of a star schema are not usually represented in the fact table. In this paper, we show a simple way to adjoin physical copies of dimension columns to the fact table, dicing data to effectively cluster query retrieval, and explain how such dicing can be achieved on database products other than DB2. We provide benchmark measurements to show successful use of this methodology on three commercial database products.
NordicDB: a Nordic pool and portal for genome-wide control data
Leu, Monica; Humphreys, Keith; Surakka, Ida; Rehnberg, Emil; Muilu, Juha; Rosenström, Päivi; Almgren, Peter; Jääskeläinen, Juha; Lifton, Richard P; Kyvik, Kirsten Ohm; Kaprio, Jaakko; Pedersen, Nancy L; Palotie, Aarno; Hall, Per; Grönberg, Henrik; Groop, Leif; Peltonen, Leena; Palmgren, Juni; Ripatti, Samuli
2010-01-01
A cost-efficient way to increase power in a genetic association study is to pool controls from different sources. The genotyping effort can then be directed to large case series. The Nordic Control database, NordicDB, has been set up as a unique resource in the Nordic area and the data are available for authorized users through the web portal (http://www.nordicdb.org). The current version of NordicDB pools together high-density genome-wide SNP information from ∼5000 controls originating from Finnish, Swedish and Danish studies and shows country-specific allele frequencies for SNP markers. The genetic homogeneity of the samples was investigated using multidimensional scaling (MDS) analysis and pairwise allele frequency differences between the studies. The plot of the first two MDS components showed excellent resemblance to the geographical placement of the samples, with a clear NW–SE gradient. We advise researchers to assess the impact of population structure when incorporating NordicDB controls in association studies. This harmonized Nordic database presents a unique genome-wide resource for future genetic association studies in the Nordic countries. PMID:20664631
Multi-dimensional effects of color on the world wide web
NASA Astrophysics Data System (ADS)
Morton, Jill
2002-06-01
Color is the most powerful building material of visual imagery on the World Wide Web. It must function successfully as it has done historically in traditional two-dimensional media, as well as address new challenges presented by this electronic medium. The psychological, physiological, technical and aesthetic effects of color have been redefined by the unique requirements of the electronic transmission of text and images on the Web. Color simultaneously addresses each of these dimensions in this electronic medium.
PrIMe Next Frontier: Large, Multi-Dimensional Data Sets
2015-07-21
is provided below. 3.4.2 Entities Figure 4 is a diagram that represents the datatypes of objects/instances that are used within the...ajax({ type: ’GET’, url: .html’, dataType : ’text’, success: function(res) { $(’body’).append(res... dataType : ’script’, success: function() { if (++counter == libraries.length) createSpecWindow(callback
ERIC Educational Resources Information Center
Painter, Derrick
1996-01-01
Discussion of dictionaries as databases focuses on the digitizing of The Oxford English dictionary (OED) and the use of Standard Generalized Mark-Up Language (SGML). Topics include the creation of a consortium to digitize the OED, document structure, relational databases, text forms, sequence, and discourse. (LRW)
Ghazizadeh, Mahtab; McDonald, Anthony D; Lee, John D
2014-09-01
This study applies text mining to extract clusters of vehicle problems and associated trends from free-response data in the National Highway Traffic Safety Administration's vehicle owner's complaint database. As the automotive industry adopts new technologies, it is important to systematically assess the effect of these changes on traffic safety. Driving simulators, naturalistic driving data, and crash databases all contribute to a better understanding of how drivers respond to changing vehicle technology, but other approaches, such as automated analysis of incident reports, are needed. Free-response data from incidents representing two severity levels (fatal incidents and incidents involving injury) were analyzed using a text mining approach: latent semantic analysis (LSA). LSA and hierarchical clustering identified clusters of complaints for each severity level, which were compared and analyzed across time. Cluster analysis identified eight clusters of fatal incidents and six clusters of incidents involving injury. Comparisons showed that although the airbag clusters across the two severity levels have the same most frequent terms, the circumstances around the incidents differ. The time trends show clear increases in complaints surrounding the Ford/Firestone tire recall and the Toyota unintended acceleration recall. Increases in complaints may be partially driven by these recall announcements and the associated media attention. Text mining can reveal useful information from free-response databases that would otherwise be prohibitively time-consuming and difficult to summarize manually. Text mining can extend human analysis capabilities for large free-response databases to support earlier detection of problems and more timely safety interventions.
Influencing factors on high-risk sexual behaviors in young people: an ecological perspective.
Arabi-Mianrood, Hoda; Hamzehgardeshi, Zeinab; Khoori, Elham; Moosazadeh, Mahmood; Shahhosseini, Zohreh
2017-04-19
Background In recent years, high-risk sexual behaviors due to their negative consequences both for the individual and society have received more attention than other high-risk behaviors. Objective The aim of this study was to review the influencing factors of high-risk sexual behaviors among young people from an ecological point of view. Methods This review was conducted through searching databases including PubMed, Web of Science, Scopus, Google Scholar and the Cochrane Library with keywords such as sexual risk-taking behavior, high-risk sex, unprotected sex and unsafe sex. The relevant papers published between 1995 and 2016 were extracted. After reviewing the abstract and full text of the articles, 45 papers were used to write this article. Results From an ecological theory approach, factors which influence high-risk sexual behaviors are divided into three categories - the microsystem, the mesosystem and the macrosystem. The microsystem includes factors such as age, gender, race, marital status, place of residence, religion, level of education, personality traits, psychological problems, childhood experiences, body image and coincidence of high-risk behaviors; the mesosystem includes factors such as family structure, peers and sex education; in the macrosystem, the impact of culture and traditions of the society, economic status and the media are presented. Conclusion Given that high-risk sexual behaviors often have multiple causes, it seems that health policymakers must consider multi-dimensional interventions to influence high-risk sexual behaviors based on the ecological approach.
The National State Policy Database. Quick Turn Around (QTA).
ERIC Educational Resources Information Center
Ahearn, Eileen; Jackson, Terry
This paper describes the National State Policy Database (NSPD), a full-text searchable database of state and federal education regulations for special education. It summarizes the history of the NSPD and reports on a survey of state directors or their designees as to their use of the database and their suggestions for its future expansion. The…
Selecting Data-Base Management Software for Microcomputers in Libraries and Information Units.
ERIC Educational Resources Information Center
Pieska, K. A. O.
1986-01-01
Presents a model for the evaluation of database management systems software from the viewpoint of librarians and information specialists. The properties of data management systems, database management systems, and text retrieval systems are outlined and compared. (10 references) (CLB)
ASM Based Synthesis of Handwritten Arabic Text Pages
Al-Hamadi, Ayoub; Elzobi, Moftah; El-etriby, Sherif; Ghoneim, Ahmed
2015-01-01
Document analysis tasks, as text recognition, word spotting, or segmentation, are highly dependent on comprehensive and suitable databases for training and validation. However their generation is expensive in sense of labor and time. As a matter of fact, there is a lack of such databases, which complicates research and development. This is especially true for the case of Arabic handwriting recognition, that involves different preprocessing, segmentation, and recognition methods, which have individual demands on samples and ground truth. To bypass this problem, we present an efficient system that automatically turns Arabic Unicode text into synthetic images of handwritten documents and detailed ground truth. Active Shape Models (ASMs) based on 28046 online samples were used for character synthesis and statistical properties were extracted from the IESK-arDB database to simulate baselines and word slant or skew. In the synthesis step ASM based representations are composed to words and text pages, smoothed by B-Spline interpolation and rendered considering writing speed and pen characteristics. Finally, we use the synthetic data to validate a segmentation method. An experimental comparison with the IESK-arDB database encourages to train and test document analysis related methods on synthetic samples, whenever no sufficient natural ground truthed data is available. PMID:26295059
ASM Based Synthesis of Handwritten Arabic Text Pages.
Dinges, Laslo; Al-Hamadi, Ayoub; Elzobi, Moftah; El-Etriby, Sherif; Ghoneim, Ahmed
2015-01-01
Document analysis tasks, as text recognition, word spotting, or segmentation, are highly dependent on comprehensive and suitable databases for training and validation. However their generation is expensive in sense of labor and time. As a matter of fact, there is a lack of such databases, which complicates research and development. This is especially true for the case of Arabic handwriting recognition, that involves different preprocessing, segmentation, and recognition methods, which have individual demands on samples and ground truth. To bypass this problem, we present an efficient system that automatically turns Arabic Unicode text into synthetic images of handwritten documents and detailed ground truth. Active Shape Models (ASMs) based on 28046 online samples were used for character synthesis and statistical properties were extracted from the IESK-arDB database to simulate baselines and word slant or skew. In the synthesis step ASM based representations are composed to words and text pages, smoothed by B-Spline interpolation and rendered considering writing speed and pen characteristics. Finally, we use the synthetic data to validate a segmentation method. An experimental comparison with the IESK-arDB database encourages to train and test document analysis related methods on synthetic samples, whenever no sufficient natural ground truthed data is available.
Image query and indexing for digital x rays
NASA Astrophysics Data System (ADS)
Long, L. Rodney; Thoma, George R.
1998-12-01
The web-based medical information retrieval system (WebMIRS) allows interned access to databases containing 17,000 digitized x-ray spine images and associated text data from National Health and Nutrition Examination Surveys (NHANES). WebMIRS allows SQL query of the text, and viewing of the returned text records and images using a standard browser. We are now working (1) to determine utility of data directly derived from the images in our databases, and (2) to investigate the feasibility of computer-assisted or automated indexing of the images to support image retrieval of images of interest to biomedical researchers in the field of osteoarthritis. To build an initial database based on image data, we are manually segmenting a subset of the vertebrae, using techniques from vertebral morphometry. From this, we will derive and add to the database vertebral features. This image-derived data will enhance the user's data access capability by enabling the creation of combined SQL/image-content queries.
Tebel, Katrin; Boldt, Vivien; Steininger, Anne; Port, Matthias; Ebert, Grit; Ullmann, Reinhard
2017-01-06
The analysis of DNA copy number variants (CNV) has increasing impact in the field of genetic diagnostics and research. However, the interpretation of CNV data derived from high resolution array CGH or NGS platforms is complicated by the considerable variability of the human genome. Therefore, tools for multidimensional data analysis and comparison of patient cohorts are needed to assist in the discrimination of clinically relevant CNVs from others. We developed GenomeCAT, a standalone Java application for the analysis and integrative visualization of CNVs. GenomeCAT is composed of three modules dedicated to the inspection of single cases, comparative analysis of multidimensional data and group comparisons aiming at the identification of recurrent aberrations in patients sharing the same phenotype, respectively. Its flexible import options ease the comparative analysis of own results derived from microarray or NGS platforms with data from literature or public depositories. Multidimensional data obtained from different experiment types can be merged into a common data matrix to enable common visualization and analysis. All results are stored in the integrated MySQL database, but can also be exported as tab delimited files for further statistical calculations in external programs. GenomeCAT offers a broad spectrum of visualization and analysis tools that assist in the evaluation of CNVs in the context of other experiment data and annotations. The use of GenomeCAT does not require any specialized computer skills. The various R packages implemented for data analysis are fully integrated into GenomeCATs graphical user interface and the installation process is supported by a wizard. The flexibility in terms of data import and export in combination with the ability to create a common data matrix makes the program also well suited as an interface between genomic data from heterogeneous sources and external software tools. Due to the modular architecture the functionality of GenomeCAT can be easily extended by further R packages or customized plug-ins to meet future requirements.
Whitehead, Lisa
2009-01-01
Fatigue is a common symptom associated with a wide range of chronic diseases. A large number of instruments have been developed to measure fatigue. An assessment regarding the reliability, validity, and utility of fatigue measures is time-consuming for the clinician and researcher, and few reviews exist on which to draw such information. The aim of this article is to present a critical review of fatigue measures, the populations in which the scales have been used, and the extent to which the psychometric properties of each instrument have been evaluated to provide clinicians and researchers with information on which to base decisions. Seven databases were searched for all articles that measured fatigue and offered an insight into the psychometric properties of the scales used over the period 1980-2007. Criteria for judging the "ideal" measure were developed to encompass scale usability, clinical/research utility, and the robustness of psychometric properties. Twenty-two fatigue measures met the inclusion criteria and were evaluated. A further 17 measures met some of the criteria, but have not been tested beyond initial development, and are reviewed briefly at the end of the article. The review did not identify any instrument that met all the criteria of an ideal instrument. However, a small number of short instruments demonstrated good psychometric properties (Fatigue Severity Scale [FSS], Fatigue Impact Scale [FIS], and Brief Fatigue Inventory [BFI]), and three comprehensive instruments demonstrated the same (Fatigue Symptom Inventory [FSI], Multidimensional Assessment of Fatigue [MAF], and Multidimensional Fatigue Symptom Inventory [MFSI]). Only four measures (BFI, FSS, FSI, and MAF) demonstrated the ability to detect change over time. The clinician and researcher also should consider the populations in which the scale has been used previously to assess its validity with their own patient group, and assess the content of a scale to ensure that the key qualitative aspects of fatigue of the population of interest are covered.
Generalized entropies and the similarity of texts
NASA Astrophysics Data System (ADS)
Altmann, Eduardo G.; Dias, Laércio; Gerlach, Martin
2017-01-01
We show how generalized Gibbs-Shannon entropies can provide new insights on the statistical properties of texts. The universal distribution of word frequencies (Zipf’s law) implies that the generalized entropies, computed at the word level, are dominated by words in a specific range of frequencies. Here we show that this is the case not only for the generalized entropies but also for the generalized (Jensen-Shannon) divergences, used to compute the similarity between different texts. This finding allows us to identify the contribution of specific words (and word frequencies) for the different generalized entropies and also to estimate the size of the databases needed to obtain a reliable estimation of the divergences. We test our results in large databases of books (from the google n-gram database) and scientific papers (indexed by Web of Science).
DOE Office of Scientific and Technical Information (OSTI.GOV)
Jasper, Ahren
2015-04-14
The appropriateness of treating crossing seams of electronic states of different spins as nonadiabatic transition states in statistical calculations of spin-forbidden reaction rates is considered. We show that the spin-forbidden reaction coordinate, the nuclear coordinate perpendicular to the crossing seam, is coupled to the remaining nuclear degrees of freedom. We found that this coupling gives rise to multidimensional effects that are not typically included in statistical treatments of spin-forbidden kinetics. Three qualitative categories of multidimensional effects may be identified: static multidimensional effects due to the geometry-dependence of the local shape of the crossing seam and of the spin–orbit coupling, dynamicalmore » multidimensional effects due to energy exchange with the reaction coordinate during the seam crossing, and nonlocal(history-dependent) multidimensional effects due to interference of the electronic variables at second, third, and later seam crossings. Nonlocal multidimensional effects are intimately related to electronic decoherence, where electronic dephasing acts to erase the history of the system. A semiclassical model based on short-time full-dimensional trajectories that includes all three multidimensional effects as well as a model for electronic decoherence is presented. The results of this multidimensional nonadiabatic statistical theory (MNST) for the 3O + CO → CO 2 reaction are compared with the results of statistical theories employing one-dimensional (Landau–Zener and weak coupling) models for the transition probability and with those calculated previously using multistate trajectories. The MNST method is shown to accurately reproduce the multistate decay-of-mixing trajectory results, so long as consistent thresholds are used. Furthermore, the MNST approach has several advantages over multistate trajectory approaches and is more suitable in chemical kinetics calculations at low temperatures and for complex systems. The error in statistical calculations that neglect multidimensional effects is shown to be as large as a factor of 2 for this system, with static multidimensional effects identified as the largest source of error.« less
Mining Quality Phrases from Massive Text Corpora
Liu, Jialu; Shang, Jingbo; Wang, Chi; Ren, Xiang; Han, Jiawei
2015-01-01
Text data are ubiquitous and play an essential role in big data applications. However, text data are mostly unstructured. Transforming unstructured text into structured units (e.g., semantically meaningful phrases) will substantially reduce semantic ambiguity and enhance the power and efficiency at manipulating such data using database technology. Thus mining quality phrases is a critical research problem in the field of databases. In this paper, we propose a new framework that extracts quality phrases from text corpora integrated with phrasal segmentation. The framework requires only limited training but the quality of phrases so generated is close to human judgment. Moreover, the method is scalable: both computation time and required space grow linearly as corpus size increases. Our experiments on large text corpora demonstrate the quality and efficiency of the new method. PMID:26705375
On the Need for Multidimensional Stirling Simulations
NASA Technical Reports Server (NTRS)
Dyson, Rodger W.; Wilson, Scott D.; Tew, Roy C.; Demko, Rikako
2005-01-01
Given the cost and complication of simulating Stirling convertors, do we really need multidimensional modeling when one-dimensional capabilities exist? This paper provides a comprehensive description of when and why multidimensional simulation is needed.
Crawford, Jarret T; Brandt, Mark J; Inbar, Yoel; Chambers, John R; Motyl, Matt
2017-03-01
Liberals and conservatives both express prejudice toward ideologically dissimilar others (Brandt et al., 2014). Previous work on ideological prejudice did not take advantage of evidence showing that ideology is multidimensional, with social and economic ideologies representing related but separable belief systems. In 5 studies (total N = 4912), we test 3 competing hypotheses of a multidimensional account of ideological prejudice. The dimension-specific symmetry hypothesis predicts that social and economic ideologies differentially predict prejudice against targets who are perceived to vary on the social and economic political dimensions, respectively. The social primacy hypothesis predicts that such ideological worldview conflict is experienced more strongly along the social than economic dimension. The social-specific asymmetry hypothesis predicts that social conservatives will be more prejudiced than social liberals, with no specific hypotheses for the economic dimension. Using multiple target groups, multiple prejudice measures (e.g., global evaluations, behavior), and multiple social and economic ideology measures (self-placement, issue positions), we found relatively consistent support for the dimension-specific symmetry and social primacy hypotheses, and no support for the social-specific asymmetry hypothesis. These results suggest that worldview conflict and negative intergroup attitudes and behaviors are dimension-specific, but that the social dimension appears to inspire more political conflict than the economic dimension. (PsycINFO Database Record (c) 2017 APA, all rights reserved).
A multidimensional model of police legitimacy: A cross-cultural assessment.
Tankebe, Justice; Reisig, Michael D; Wang, Xia
2016-02-01
This study used survey data from cross-sectional, university-based samples of young adults in different cultural settings (i.e., the United States and Ghana) to accomplish 2 main objectives: (1) to construct a 4-dimensional police legitimacy scale, and (2) to assess the relationship that police legitimacy and feelings of obligation to obey the police have with 2 outcome measures. The fit statistics for the second-order confirmatory factor models indicated that the 4-dimensional police legitimacy model is reasonably consistent with the data in both samples. Results from the linear regression analyses showed that the police legitimacy scale is related to cooperation with the police, and that the observed association is attenuated when the obligation to obey scale is included in the model specification in both the United States and Ghana data. A similar pattern emerged in the U.S. sample when estimating compliance with the law models. However, although police legitimacy was associated with compliance in the Ghana sample, this relationship along with the test statistic for the sense of obligation to obey estimate were both null in the fully saturated equation. The findings provide support for the Bottoms and Tankebe's (2012) argument that legitimacy is multidimensional, comprising police lawfulness, distributive fairness, procedural fairness, and effectiveness. However, the link between police legitimacy and social order appears to be culturally variable. PsycINFO Database Record (c) 2016 APA, all rights reserved.
van Dongen, J J M; Lhermitte, L; Böttcher, S; Almeida, J; van der Velden, V H J; Flores-Montero, J; Rawstron, A; Asnafi, V; Lécrevisse, Q; Lucio, P; Mejstrikova, E; Szczepański, T; Kalina, T; de Tute, R; Brüggemann, M; Sedek, L; Cullen, M; Langerak, A W; Mendonça, A; Macintyre, E; Martin-Ayuso, M; Hrusak, O; Vidriales, M B; Orfao, A
2012-01-01
Most consensus leukemia & lymphoma antibody panels consist of lists of markers based on expert opinions, but they have not been validated. Here we present the validated EuroFlow 8-color antibody panels for immunophenotyping of hematological malignancies. The single-tube screening panels and multi-tube classification panels fit into the EuroFlow diagnostic algorithm with entries defined by clinical and laboratory parameters. The panels were constructed in 2–7 sequential design–evaluation–redesign rounds, using novel Infinicyt software tools for multivariate data analysis. Two groups of markers are combined in each 8-color tube: (i) backbone markers to identify distinct cell populations in a sample, and (ii) markers for characterization of specific cell populations. In multi-tube panels, the backbone markers were optimally placed at the same fluorochrome position in every tube, to provide identical multidimensional localization of the target cell population(s). The characterization markers were positioned according to the diagnostic utility of the combined markers. Each proposed antibody combination was tested against reference databases of normal and malignant cells from healthy subjects and WHO-based disease entities, respectively. The EuroFlow studies resulted in validated and flexible 8-color antibody panels for multidimensional identification and characterization of normal and aberrant cells, optimally suited for immunophenotypic screening and classification of hematological malignancies. PMID:22552007
Bidonde, Julia; Busch, Angela Jean; Bath, Brenna; Milosavljevic, Stephan
2014-01-01
The objective of this umbrella systematic review was to identify, evaluate, and synthesize systematic reviews of physical activity interventions for adults with fibromyalgia (FM) focussing on four outcomes: pain, multidimensional function (wellness or quality of life), physical function (self-reported physical function or measured physical fitness) and adverse effects. A further objective was to link these outcomes with details of the interventions so as to guide and shape future practice and research. Electronic databases including Medline, EMBASE, CINAHL, AMED, the Cochrane Library, and DARE, were searched for the January 1(st) 2007 to March 31(st) 2013 period. Nine systematic reviews (60 RCTs with 3816 participants) were included. Meta-analysis was not conducted due to the heterogeneity of the sample. We found positive results of diverse exercise interventions on pain, multidimensional function, and self-reported physical function, and no supporting evidence for new (to FM) interventions (i.e., qigong, tai chi). There were no serious adverse effects reported. The variability of the interventions in the reviews prevented us from answering important clinical questions to guide practical decisions about optimal modes or dosages (i.e., frequency, intensity, duration). Finally, the number of review articles is proliferating, leading researchers and reviewers to consider the rigor and quality of the information being reviewed. As well, consumers of these reviews (i.e., clinicians, individuals with FM) should not rely on them without careful consideration.
Tomlinson, Deborah; Hinds, Pamela S; Ethier, Marie-Chantal; Ness, Kirsten K; Zupanec, Sue; Sung, Lillian
2013-01-01
Despite the recognized distressing symptom of fatigue in children with cancer, little information is available to assist in the selection of an instrument to be used to measure fatigue. The objectives of this study were to 1) describe the instruments that have been used to measure cancer-related fatigue in children and adolescents and 2) summarize the psychometric properties of the most commonly used instruments used to measure fatigue in children and adolescents with cancer. Five major electronic databases were systematically searched for studies using a fatigue measurement scale in a population of children or adolescents with cancer. Fatigue scales used in those studies were included in the review. From a total of 1753 articles, 25 were included. We identified two main fatigue measurement instruments used in a pediatric oncology population: 1) the Fatigue Scale-Child/Fatigue Scale-Adolescent and the proxy report versions for parents and staff and 2) the PedsQL™ Multidimensional Fatigue Scale. These two scales show similar attributes with reasonably good internal consistency and responsiveness. Either the Fatigue Scale or PedsQL Multidimensional Fatigue Scale can be incorporated into clinical research. Future research should focus on identifying specific fatigue measures more suited to different purposes such as comparative trials or identification of high-risk groups. Copyright © 2013 U.S. Cancer Pain Relief Committee. Published by Elsevier Inc. All rights reserved.
Measuring adult picky eating: The development of a multidimensional self-report instrument.
Ellis, Jordan M; Galloway, Amy T; Webb, Rose Mary; Martz, Denise M
2017-08-01
A brief multidimensional measure of adult picky eating (PE) behavior was developed using a large U.S. adult sample. In addition, the study explored associations between specific aspects of adult PE behavior and psychosocial impairment in effort to support the inclusion of adults in the Diagnostic and Statistical Manual for Mental Disorders-Fifth Edition (DSM-5) avoidant-restrictive food intake disorder (ARFID). The study included 3 phases of qualitative and quantitative data collection. Participants were 1,663 U.S. adults who completed online surveys. Exploratory and confirmatory factor analyses were used to develop PE subscales. Associations among the PE subscales and measures of psychosocial impairment were examined. Exploratory and confirmatory factor analysis supported a 16-item 4-factor model of adult PE that included subscales of meal presentation, food variety, meal disengagement, and taste aversion. The measure also demonstrated convergence with previous measures of PE. The meal disengagement and meal presentation subscales demonstrated significant associations with social eating anxiety, anxiety sensitivity, eating related quality of life (QOL), and psychological flexibility. Meal disengagement alone was significantly associated with depressive symptoms. The Adult Picky Eating Questionnaire (APEQ) demonstrated sound psychometric properties and may be used to further investigate adult PE behavior. The relationships between adult PE and psychological impairment, particularly social anxiety, support the inclusion of ARFID in the DSM-5. (PsycINFO Database Record (c) 2017 APA, all rights reserved).
Posttraumatic growth in bereaved parents: A multidimensional model of associated factors.
Albuquerque, Sara; Narciso, Isabel; Pereira, Marco
2018-03-01
Although the death of a child is a devastating event, recent evidence shows that personal growth is a relevant outcome of parents' grief. This study aimed to examine the factors associated with posttraumatic growth (PTG) and to propose a multidimensional model consisting of sociodemographic, situational, and intrapersonal and interpersonal factors. A sample (N = 197; 89.8% female; mean age = 39.44 years) of bereaved parents completed the Post-Traumatic Growth Inventory-Short Form, the 14-Item Resilience Scale, the Continuing Bonds Scale, and the Dyadic Coping Inventory. The final model consisted of sociodemographic, situational, intrapersonal, and interpersonal factors of PTG, which accounted for 36.7% of the variance. Higher levels of PTG were generally associated with female sex, younger age of the child, higher levels of resilience, higher levels of internalized continuing bonds (i.e., internal representation of the child, maintaining psychological proximity), and higher levels of stress communication by the partner (communicating the stress experience and requesting emotional or practical support). In clinical practice, health professionals assisting bereaved parents should pay attention to men and parents of older children, who might be at higher risk of difficulties in developing PTG. Additionally, promoting a more internalized bond with the child, resilience and dyadic coping, especially stress communication, can constitute important therapeutic goals. (PsycINFO Database Record (c) 2018 APA, all rights reserved).
Vanasse, A; Courteau, M; Ethier, J-F
2018-04-01
To synthesize concepts and approaches related to the analysis of patterns or processes of care and patient's outcomes into a comprehensive model of care trajectories, focusing on hospital readmissions for patients with chronic ambulatory care sensitive conditions (ACSCs). Narrative literature review. Published studies between January 2000 and November 2017, using the concepts of 'continuity', 'pathway', 'episode', and 'trajectory', and focused on readmissions and chronic ACSCs, were collected in electronic databases. Qualitative content analysis was performed with emphasis on key constituents to build a comprehensive model. Specific common constituents are shared by the concepts reviewed: they focus on the patient, aim to measure and improve outcomes, follow specific periods of time and consider other factors related to care providers, care units, care settings, and treatments. Using these common denominators, the comprehensive '6W' multidimensional model of care trajectories was created. Considering patients' attributes and their chronic ACSCs illness course ('who' and 'why' dimensions), this model reflects their patterns of health care use across care providers ('which'), care units ('where'), and treatments ('what'), at specific periods of time ('when'). The '6W' model of care trajectories could provide valuable information on 'missed opportunities' to reduce readmission rates and improve quality of both ambulatory and inpatient care. Copyright © 2018 The Authors. Published by Elsevier Ltd.. All rights reserved.
Multidimensional Scaling in the Poincare Disk
2011-05-01
REPORT Multidimensional Scaling in the Poincare Dis 14. ABSTRACT 16. SECURITY CLASSIFICATION OF: Multidimensional scaling (MDS) is a class of projective...DATES COVERED (From - To) Standard Form 298 (Rev 8/98) Prescribed by ANSI Std. Z39.18 - Multidimensional Scaling in the Poincare Dis Report Title... plane . Our construction is based on an approximate hyperbolic line search and exempli?es some of the particulars that need to be addressed when
Solomon, Nancy Pearl; Dietsch, Angela M; Dietrich-Burns, Katie E; Styrmisdottir, Edda L; Armao, Christopher S
2016-05-01
This report describes the development and preliminary analysis of a database for traumatically injured military service members with dysphagia. A multidimensional database was developed to capture clinical variables related to swallowing. Data were derived from clinical records and instrumental swallow studies, and ranged from demographics, injury characteristics, swallowing biomechanics, medications, and standardized tools (e.g., Glasgow Coma Scale, Penetration-Aspiration Scale). Bayesian Belief Network modeling was used to analyze the data at intermediate points, guide data collection, and predict outcomes. Predictive models were validated with independent data via receiver operating characteristic curves. The first iteration of the model (n = 48) revealed variables that could be collapsed for the second model (n = 96). The ability to predict recovery from dysphagia improved from the second to third models (area under the curve = 0.68 to 0.86). The third model, based on 161 cases, revealed "initial diet restrictions" as first-degree, and "Glasgow Coma Scale, intubation history, and diet change" as second-degree associates for diet restrictions at discharge. This project demonstrates the potential for bioinformatics to advance understanding of dysphagia. This database in concert with Bayesian Belief Network modeling makes it possible to explore predictive relationships between injuries and swallowing function, individual variability in recovery, and appropriate treatment options. Reprint & Copyright © 2016 Association of Military Surgeons of the U.S.
Scientific Data Services -- A High-Performance I/O System with Array Semantics
DOE Office of Scientific and Technical Information (OSTI.GOV)
Wu, Kesheng; Byna, Surendra; Rotem, Doron
2011-09-21
As high-performance computing approaches exascale, the existing I/O system design is having trouble keeping pace in both performance and scalability. We propose to address this challenge by adopting database principles and techniques in parallel I/O systems. First, we propose to adopt an array data model because many scientific applications represent their data in arrays. This strategy follows a cardinal principle from database research, which separates the logical view from the physical layout of data. This high-level data model gives the underlying implementation more freedom to optimize the physical layout and to choose the most effective way of accessing the data.more » For example, knowing that a set of write operations is working on a single multi-dimensional array makes it possible to keep the subarrays in a log structure during the write operations and reassemble them later into another physical layout as resources permit. While maintaining the high-level view, the storage system could compress the user data to reduce the physical storage requirement, collocate data records that are frequently used together, or replicate data to increase availability and fault-tolerance. Additionally, the system could generate secondary data structures such as database indexes and summary statistics. We expect the proposed Scientific Data Services approach to create a “live” storage system that dynamically adjusts to user demands and evolves with the massively parallel storage hardware.« less
DEXTER: Disease-Expression Relation Extraction from Text.
Gupta, Samir; Dingerdissen, Hayley; Ross, Karen E; Hu, Yu; Wu, Cathy H; Mazumder, Raja; Vijay-Shanker, K
2018-01-01
Gene expression levels affect biological processes and play a key role in many diseases. Characterizing expression profiles is useful for clinical research, and diagnostics and prognostics of diseases. There are currently several high-quality databases that capture gene expression information, obtained mostly from large-scale studies, such as microarray and next-generation sequencing technologies, in the context of disease. The scientific literature is another rich source of information on gene expression-disease relationships that not only have been captured from large-scale studies but have also been observed in thousands of small-scale studies. Expression information obtained from literature through manual curation can extend expression databases. While many of the existing databases include information from literature, they are limited by the time-consuming nature of manual curation and have difficulty keeping up with the explosion of publications in the biomedical field. In this work, we describe an automated text-mining tool, Disease-Expression Relation Extraction from Text (DEXTER) to extract information from literature on gene and microRNA expression in the context of disease. One of the motivations in developing DEXTER was to extend the BioXpress database, a cancer-focused gene expression database that includes data derived from large-scale experiments and manual curation of publications. The literature-based portion of BioXpress lags behind significantly compared to expression information obtained from large-scale studies and can benefit from our text-mined results. We have conducted two different evaluations to measure the accuracy of our text-mining tool and achieved average F-scores of 88.51 and 81.81% for the two evaluations, respectively. Also, to demonstrate the ability to extract rich expression information in different disease-related scenarios, we used DEXTER to extract information on differential expression information for 2024 genes in lung cancer, 115 glycosyltransferases in 62 cancers and 826 microRNA in 171 cancers. All extractions using DEXTER are integrated in the literature-based portion of BioXpress.Database URL: http://biotm.cis.udel.edu/DEXTER.
Scaling Laws for the Multidimensional Burgers Equation with Quadratic External Potential
NASA Astrophysics Data System (ADS)
Leonenko, N. N.; Ruiz-Medina, M. D.
2006-07-01
The reordering of the multidimensional exponential quadratic operator in coordinate-momentum space (see X. Wang, C.H. Oh and L.C. Kwek (1998). J. Phys. A.: Math. Gen. 31:4329-4336) is applied to derive an explicit formulation of the solution to the multidimensional heat equation with quadratic external potential and random initial conditions. The solution to the multidimensional Burgers equation with quadratic external potential under Gaussian strongly dependent scenarios is also obtained via the Hopf-Cole transformation. The limiting distributions of scaling solutions to the multidimensional heat and Burgers equations with quadratic external potential are then obtained under such scenarios.
NASA Astrophysics Data System (ADS)
Gromov, Yu Yu; Minin, Yu V.; Ivanova, O. G.; Morozova, O. N.
2018-03-01
Multidimensional discrete distributions of probabilities of independent random values were received. Their one-dimensional distribution is widely used in probability theory. Producing functions of those multidimensional distributions were also received.
Library Instruction and Online Database Searching.
ERIC Educational Resources Information Center
Mercado, Heidi
1999-01-01
Reviews changes in online database searching in academic libraries. Topics include librarians conducting all searches; the advent of end-user searching and the need for user instruction; compact disk technology; online public catalogs; the Internet; full text databases; electronic information literacy; user education and the remote library user;…
A New Time-varying Concept of Risk in a Changing Climate.
Sarhadi, Ali; Ausín, María Concepción; Wiper, Michael P
2016-10-20
In a changing climate arising from anthropogenic global warming, the nature of extreme climatic events is changing over time. Existing analytical stationary-based risk methods, however, assume multi-dimensional extreme climate phenomena will not significantly vary over time. To strengthen the reliability of infrastructure designs and the management of water systems in the changing environment, multidimensional stationary risk studies should be replaced with a new adaptive perspective. The results of a comparison indicate that current multi-dimensional stationary risk frameworks are no longer applicable to projecting the changing behaviour of multi-dimensional extreme climate processes. Using static stationary-based multivariate risk methods may lead to undesirable consequences in designing water system infrastructures. The static stationary concept should be replaced with a flexible multi-dimensional time-varying risk framework. The present study introduces a new multi-dimensional time-varying risk concept to be incorporated in updating infrastructure design strategies under changing environments arising from human-induced climate change. The proposed generalized time-varying risk concept can be applied for all stochastic multi-dimensional systems that are under the influence of changing environments.
Information Extraction for Clinical Data Mining: A Mammography Case Study
Nassif, Houssam; Woods, Ryan; Burnside, Elizabeth; Ayvaci, Mehmet; Shavlik, Jude; Page, David
2013-01-01
Breast cancer is the leading cause of cancer mortality in women between the ages of 15 and 54. During mammography screening, radiologists use a strict lexicon (BI-RADS) to describe and report their findings. Mammography records are then stored in a well-defined database format (NMD). Lately, researchers have applied data mining and machine learning techniques to these databases. They successfully built breast cancer classifiers that can help in early detection of malignancy. However, the validity of these models depends on the quality of the underlying databases. Unfortunately, most databases suffer from inconsistencies, missing data, inter-observer variability and inappropriate term usage. In addition, many databases are not compliant with the NMD format and/or solely consist of text reports. BI-RADS feature extraction from free text and consistency checks between recorded predictive variables and text reports are crucial to addressing this problem. We describe a general scheme for concept information retrieval from free text given a lexicon, and present a BI-RADS features extraction algorithm for clinical data mining. It consists of a syntax analyzer, a concept finder and a negation detector. The syntax analyzer preprocesses the input into individual sentences. The concept finder uses a semantic grammar based on the BI-RADS lexicon and the experts’ input. It parses sentences detecting BI-RADS concepts. Once a concept is located, a lexical scanner checks for negation. Our method can handle multiple latent concepts within the text, filtering out ultrasound concepts. On our dataset, our algorithm achieves 97.7% precision, 95.5% recall and an F1-score of 0.97. It outperforms manual feature extraction at the 5% statistical significance level. PMID:23765123
Information Extraction for Clinical Data Mining: A Mammography Case Study.
Nassif, Houssam; Woods, Ryan; Burnside, Elizabeth; Ayvaci, Mehmet; Shavlik, Jude; Page, David
2009-01-01
Breast cancer is the leading cause of cancer mortality in women between the ages of 15 and 54. During mammography screening, radiologists use a strict lexicon (BI-RADS) to describe and report their findings. Mammography records are then stored in a well-defined database format (NMD). Lately, researchers have applied data mining and machine learning techniques to these databases. They successfully built breast cancer classifiers that can help in early detection of malignancy. However, the validity of these models depends on the quality of the underlying databases. Unfortunately, most databases suffer from inconsistencies, missing data, inter-observer variability and inappropriate term usage. In addition, many databases are not compliant with the NMD format and/or solely consist of text reports. BI-RADS feature extraction from free text and consistency checks between recorded predictive variables and text reports are crucial to addressing this problem. We describe a general scheme for concept information retrieval from free text given a lexicon, and present a BI-RADS features extraction algorithm for clinical data mining. It consists of a syntax analyzer, a concept finder and a negation detector. The syntax analyzer preprocesses the input into individual sentences. The concept finder uses a semantic grammar based on the BI-RADS lexicon and the experts' input. It parses sentences detecting BI-RADS concepts. Once a concept is located, a lexical scanner checks for negation. Our method can handle multiple latent concepts within the text, filtering out ultrasound concepts. On our dataset, our algorithm achieves 97.7% precision, 95.5% recall and an F 1 -score of 0.97. It outperforms manual feature extraction at the 5% statistical significance level.
Multidimensional Knowledge Structures.
ERIC Educational Resources Information Center
Schuh, Kathy L.
Multidimensional knowledge structures, described from a constructivist perspective and aligned with the "Mind as Rhizome" metaphor, provide support for constructivist learning strategies. This qualitative study was conducted to seek empirical support for a description of multidimensional knowledge structures, focusing on the…
Multidimensional quantum entanglement with large-scale integrated optics.
Wang, Jianwei; Paesani, Stefano; Ding, Yunhong; Santagati, Raffaele; Skrzypczyk, Paul; Salavrakos, Alexia; Tura, Jordi; Augusiak, Remigiusz; Mančinska, Laura; Bacco, Davide; Bonneau, Damien; Silverstone, Joshua W; Gong, Qihuang; Acín, Antonio; Rottwitt, Karsten; Oxenløwe, Leif K; O'Brien, Jeremy L; Laing, Anthony; Thompson, Mark G
2018-04-20
The ability to control multidimensional quantum systems is central to the development of advanced quantum technologies. We demonstrate a multidimensional integrated quantum photonic platform able to generate, control, and analyze high-dimensional entanglement. A programmable bipartite entangled system is realized with dimensions up to 15 × 15 on a large-scale silicon photonics quantum circuit. The device integrates more than 550 photonic components on a single chip, including 16 identical photon-pair sources. We verify the high precision, generality, and controllability of our multidimensional technology, and further exploit these abilities to demonstrate previously unexplored quantum applications, such as quantum randomness expansion and self-testing on multidimensional states. Our work provides an experimental platform for the development of multidimensional quantum technologies. Copyright © 2018 The Authors, some rights reserved; exclusive licensee American Association for the Advancement of Science. No claim to original U.S. Government Works.
Duchrow, Timo; Shtatland, Timur; Guettler, Daniel; Pivovarov, Misha; Kramer, Stefan; Weissleder, Ralph
2009-01-01
Background The breadth of biological databases and their information content continues to increase exponentially. Unfortunately, our ability to query such sources is still often suboptimal. Here, we introduce and apply community voting, database-driven text classification, and visual aids as a means to incorporate distributed expert knowledge, to automatically classify database entries and to efficiently retrieve them. Results Using a previously developed peptide database as an example, we compared several machine learning algorithms in their ability to classify abstracts of published literature results into categories relevant to peptide research, such as related or not related to cancer, angiogenesis, molecular imaging, etc. Ensembles of bagged decision trees met the requirements of our application best. No other algorithm consistently performed better in comparative testing. Moreover, we show that the algorithm produces meaningful class probability estimates, which can be used to visualize the confidence of automatic classification during the retrieval process. To allow viewing long lists of search results enriched by automatic classifications, we added a dynamic heat map to the web interface. We take advantage of community knowledge by enabling users to cast votes in Web 2.0 style in order to correct automated classification errors, which triggers reclassification of all entries. We used a novel framework in which the database "drives" the entire vote aggregation and reclassification process to increase speed while conserving computational resources and keeping the method scalable. In our experiments, we simulate community voting by adding various levels of noise to nearly perfectly labelled instances, and show that, under such conditions, classification can be improved significantly. Conclusion Using PepBank as a model database, we show how to build a classification-aided retrieval system that gathers training data from the community, is completely controlled by the database, scales well with concurrent change events, and can be adapted to add text classification capability to other biomedical databases. The system can be accessed at . PMID:19799796
Developing a Large Lexical Database for Information Retrieval, Parsing, and Text Generation Systems.
ERIC Educational Resources Information Center
Conlon, Sumali Pin-Ngern; And Others
1993-01-01
Important characteristics of lexical databases and their applications in information retrieval and natural language processing are explained. An ongoing project using various machine-readable sources to build a lexical database is described, and detailed designs of individual entries with examples are included. (Contains 66 references.) (EAM)
Multidimensional Perfectionism and the Self
ERIC Educational Resources Information Center
Ward, Andrew M.; Ashby, Jeffrey S.
2008-01-01
This study examined multidimensional perfectionism and self-development. Two hundred seventy-one undergraduates completed a measure of multidimensional perfectionism and two Kohutian measures designed to measure aspects of self-development including social connectedness, social assurance, goal instability (idealization), and grandiosity. The…
de la Vega de León, Antonio; Bajorath, Jürgen
2016-09-01
The concept of chemical space is of fundamental relevance for medicinal chemistry and chemical informatics. Multidimensional chemical space representations are coordinate-based. Chemical space networks (CSNs) have been introduced as a coordinate-free representation. A computational approach is presented for the transformation of multidimensional chemical space into CSNs. The design of transformation CSNs (TRANS-CSNs) is based upon a similarity function that directly reflects distance relationships in original multidimensional space. TRANS-CSNs provide an immediate visualization of coordinate-based chemical space and do not require the use of dimensionality reduction techniques. At low network density, TRANS-CSNs are readily interpretable and make it possible to evaluate structure-activity relationship information originating from multidimensional chemical space.
Text mining for the biocuration workflow
Hirschman, Lynette; Burns, Gully A. P. C; Krallinger, Martin; Arighi, Cecilia; Cohen, K. Bretonnel; Valencia, Alfonso; Wu, Cathy H.; Chatr-Aryamontri, Andrew; Dowell, Karen G.; Huala, Eva; Lourenço, Anália; Nash, Robert; Veuthey, Anne-Lise; Wiegers, Thomas; Winter, Andrew G.
2012-01-01
Molecular biology has become heavily dependent on biological knowledge encoded in expert curated biological databases. As the volume of biological literature increases, biocurators need help in keeping up with the literature; (semi-) automated aids for biocuration would seem to be an ideal application for natural language processing and text mining. However, to date, there have been few documented successes for improving biocuration throughput using text mining. Our initial investigations took place for the workshop on ‘Text Mining for the BioCuration Workflow’ at the third International Biocuration Conference (Berlin, 2009). We interviewed biocurators to obtain workflows from eight biological databases. This initial study revealed high-level commonalities, including (i) selection of documents for curation; (ii) indexing of documents with biologically relevant entities (e.g. genes); and (iii) detailed curation of specific relations (e.g. interactions); however, the detailed workflows also showed many variabilities. Following the workshop, we conducted a survey of biocurators. The survey identified biocurator priorities, including the handling of full text indexed with biological entities and support for the identification and prioritization of documents for curation. It also indicated that two-thirds of the biocuration teams had experimented with text mining and almost half were using text mining at that time. Analysis of our interviews and survey provide a set of requirements for the integration of text mining into the biocuration workflow. These can guide the identification of common needs across curated databases and encourage joint experimentation involving biocurators, text mining developers and the larger biomedical research community. PMID:22513129
Text mining for the biocuration workflow.
Hirschman, Lynette; Burns, Gully A P C; Krallinger, Martin; Arighi, Cecilia; Cohen, K Bretonnel; Valencia, Alfonso; Wu, Cathy H; Chatr-Aryamontri, Andrew; Dowell, Karen G; Huala, Eva; Lourenço, Anália; Nash, Robert; Veuthey, Anne-Lise; Wiegers, Thomas; Winter, Andrew G
2012-01-01
Molecular biology has become heavily dependent on biological knowledge encoded in expert curated biological databases. As the volume of biological literature increases, biocurators need help in keeping up with the literature; (semi-) automated aids for biocuration would seem to be an ideal application for natural language processing and text mining. However, to date, there have been few documented successes for improving biocuration throughput using text mining. Our initial investigations took place for the workshop on 'Text Mining for the BioCuration Workflow' at the third International Biocuration Conference (Berlin, 2009). We interviewed biocurators to obtain workflows from eight biological databases. This initial study revealed high-level commonalities, including (i) selection of documents for curation; (ii) indexing of documents with biologically relevant entities (e.g. genes); and (iii) detailed curation of specific relations (e.g. interactions); however, the detailed workflows also showed many variabilities. Following the workshop, we conducted a survey of biocurators. The survey identified biocurator priorities, including the handling of full text indexed with biological entities and support for the identification and prioritization of documents for curation. It also indicated that two-thirds of the biocuration teams had experimented with text mining and almost half were using text mining at that time. Analysis of our interviews and survey provide a set of requirements for the integration of text mining into the biocuration workflow. These can guide the identification of common needs across curated databases and encourage joint experimentation involving biocurators, text mining developers and the larger biomedical research community.
Mohanty, Sanjay K; Agrawal, Nand Kishor; Mahapatra, Bidhubhusan; Choudhury, Dhrupad; Tuladhar, Sabarnee; Holmgren, E Valdemar
2017-01-18
Economic burden to households due to out-of-pocket expenditure (OOPE) is large in many Asian countries. Though studies suggest increasing household poverty due to high OOPE in developing countries, studies on association of multidimensional poverty and household health spending is limited. This paper tests the hypothesis that the multidimensionally poor are more likely to incur catastrophic health spending cutting across countries. Data from the Poverty and Vulnerability Assessment (PVA) Survey carried out by the International Center for Integrated Mountain Development (ICIMOD) has been used in the analyses. The PVA survey was a comprehensive household survey that covered the mountainous regions of India, Nepal and Myanmar. A total of 2647 households from India, 2310 households in Nepal and 4290 households in Myanmar covered under the PVA survey. Poverty is measured in a multidimensional framework by including the dimensions of education, income and energy, water and sanitation using the Alkire and Foster method. Health shock is measured using the frequency of illness, family sickness and death of any family member in a reference period of one year. Catastrophic health expenditure is defined as 40% above the household's capacity to pay. Results suggest that about three-fifths of the population in Myanmar, two-fifths of the population in Nepal and one-third of the population in India are multidimensionally poor. About 47% of the multidimensionally poor in India had incurred catastrophic health spending compared to 35% of the multidimensionally non-poor and the pattern was similar in both Nepal and Myanmar. The odds of incurring catastrophic health spending was 56% more among the multidimensionally poor than among the multidimensionally non-poor [95% CI: 1.35-1.76]. While health shocks to households are consistently significant predictors of catastrophic health spending cutting across country of residence, the educational attainment of the head of the household is not significant. The multidimensionally poor in the poorer regions are more likely to face health shocks and are less likely to afford professional health services. Increasing government spending on health and increasing households' access to health insurance can reduce catastrophic health spending and multidimensional poverty.
Thematic clustering of text documents using an EM-based approach
2012-01-01
Clustering textual contents is an important step in mining useful information on the web or other text-based resources. The common task in text clustering is to handle text in a multi-dimensional space, and to partition documents into groups, where each group contains documents that are similar to each other. However, this strategy lacks a comprehensive view for humans in general since it cannot explain the main subject of each cluster. Utilizing semantic information can solve this problem, but it needs a well-defined ontology or pre-labeled gold standard set. In this paper, we present a thematic clustering algorithm for text documents. Given text, subject terms are extracted and used for clustering documents in a probabilistic framework. An EM approach is used to ensure documents are assigned to correct subjects, hence it converges to a locally optimal solution. The proposed method is distinctive because its results are sufficiently explanatory for human understanding as well as efficient for clustering performance. The experimental results show that the proposed method provides a competitive performance compared to other state-of-the-art approaches. We also show that the extracted themes from the MEDLINE® dataset represent the subjects of clusters reasonably well. PMID:23046528
An introduction to multidimensional measurement using Rasch models.
Briggs, Derek C; Wilson, Mark
2003-01-01
The act of constructing a measure requires a number of important assumptions. Principle among these assumptions is that the construct is unidimensional. In practice there are many instances when the assumption of unidimensionality does not hold, and where the application of a multidimensional measurement model is both technically appropriate and substantively advantageous. In this paper we illustrate the usefulness of a multidimensional approach to measurement with the Multidimensional Random Coefficient Multinomial Logit (MRCML) model, an extension of the unidimensional Rasch model. An empirical example is taken from a collection of embedded assessments administered to 541 students enrolled in middle school science classes with a hands-on science curriculum. Student achievement on these assessments are multidimensional in nature, but can also be treated as consecutive unidimensional estimates, or as is most common, as a composite unidimensional estimate. Structural parameters are estimated for each model using ConQuest, and model fit is compared. Student achievement in science is also compared across models. The multidimensional approach has the best fit to the data, and provides more reliable estimates of student achievement than under the consecutive unidimensional approach. Finally, at an interpretational level, the multidimensional approach may well provide richer information to the classroom teacher about the nature of student achievement.
DeTEXT: A Database for Evaluating Text Extraction from Biomedical Literature Figures
Yin, Xu-Cheng; Yang, Chun; Pei, Wei-Yi; Man, Haixia; Zhang, Jun; Learned-Miller, Erik; Yu, Hong
2015-01-01
Hundreds of millions of figures are available in biomedical literature, representing important biomedical experimental evidence. Since text is a rich source of information in figures, automatically extracting such text may assist in the task of mining figure information. A high-quality ground truth standard can greatly facilitate the development of an automated system. This article describes DeTEXT: A database for evaluating text extraction from biomedical literature figures. It is the first publicly available, human-annotated, high quality, and large-scale figure-text dataset with 288 full-text articles, 500 biomedical figures, and 9308 text regions. This article describes how figures were selected from open-access full-text biomedical articles and how annotation guidelines and annotation tools were developed. We also discuss the inter-annotator agreement and the reliability of the annotations. We summarize the statistics of the DeTEXT data and make available evaluation protocols for DeTEXT. Finally we lay out challenges we observed in the automated detection and recognition of figure text and discuss research directions in this area. DeTEXT is publicly available for downloading at http://prir.ustb.edu.cn/DeTEXT/. PMID:25951377
Methods to determine the growth domain in a multidimensional environmental space.
Le Marc, Yvan; Pin, Carmen; Baranyi, József
2005-04-15
Data from a database on microbial responses to the food environment (ComBase, see www.combase.cc) were used to study the boundary of growth several pathogens (Aeromonas hydrophila, Escherichia coli, Listeria monocytogenes, Yersinia enterocolitica). Two methods were used to evaluate the growth/no growth interface. The first one is an application of the Minimum Convex Polyhedron (MCP) introduced by Baranyi et al. [Baranyi, J., Ross, T., McMeekin, T., Roberts, T.A., 1996. The effect of parameterisation on the performance of empirical models used in Predictive Microbiology. Food Microbiol. 13, 83-91.]. The second method applies logistic regression to define the boundary of growth. The combination of these two different techniques can be a useful tool to handle the problem of extrapolation of predictive models at the growth limits.
Universalism Versus Particularism Through the European Social Survey Lenses
NASA Astrophysics Data System (ADS)
Nawojczyk, M.
2006-11-01
The cultural variation of economic activity is wide and multidimensional. In my presentation I will refer to the analyses of the culture of capitalism provided by Alfons Trompenaars and Charles Hampden-Turner. According to them there are seven processes and related dilemmas which are important in analyzing the construction of a cultural system of economy. I will focus only on one of them, universalism {ITALIC versus} particularism. Using the database of Trompenaars and Hampden-Turner I will show how this dilemma was solved by managers from different European countries. That will be starting point for my analysis of universalism-particularism attitudes of respondents of European Social Survey (ESS). I will be particularly interested in verification of hypothesis on the place of Poland on the mosaic of European cultures of capitalism.
NASA Astrophysics Data System (ADS)
Romero, A.; Sol, D.
2017-09-01
Collecting data by crowdsourcing is an explored trend to support database population and update. This kind of data is unstructured and comes from text, in particular text in social networks. Geographic database is a particular case of database that can be populated by crowdsourcing which can be done when people report some urban event in a social network by writing a short message. An event can describe an accident or a non-functioning device in the urban area. The authorities then need to read and to interpret the message to provide some help for injured people or to fix a problem in a device installed in the urban area like a light or a problem on road. Our main interest is located on working with short messages organized in a collection. Most of the messages do not have geographical coordinates. The messages can then be classified by text patterns describing a location. In fact, people use a text pattern to describe an urban location. Our work tries to identify patterns inside a short text and to indicate when it describes a location. When a pattern is identified our approach look to describe the place where the event is located. The source messages used are tweets reporting events from several Mexican cities.
Blake, M.C.; Jones, D.L.; Graymer, R.W.; digital database by Soule, Adam
2000-01-01
This digital map database, compiled from previously published and unpublished data, and new mapping by the authors, represents the general distribution of bedrock and surficial deposits in the mapped area. Together with the accompanying text file (mageo.txt, mageo.pdf, or mageo.ps), it provides current information on the geologic structure and stratigraphy of the area covered. The database delineates map units that are identified by general age and lithology following the stratigraphic nomenclature of the U.S. Geological Survey. The scale of the source maps limits the spatial resolution (scale) of the database to 1:62,500 or smaller general distribution of bedrock and surficial deposits in the mapped area. Together with the accompanying text file (mageo.txt, mageo.pdf, or mageo.ps), it provides current information on the geologic structure and stratigraphy of the area covered. The database delineates map units that are identified by general age and lithology following the stratigraphic nomenclature of the U.S. Geological Survey. The scale of the source maps limits the spatial resolution (scale) of the database to 1:62,500 or smaller.
Search extension transforms Wiki into a relational system: a case for flavonoid metabolite database.
Arita, Masanori; Suwa, Kazuhiro
2008-09-17
In computer science, database systems are based on the relational model founded by Edgar Codd in 1970. On the other hand, in the area of biology the word 'database' often refers to loosely formatted, very large text files. Although such bio-databases may describe conflicts or ambiguities (e.g. a protein pair do and do not interact, or unknown parameters) in a positive sense, the flexibility of the data format sacrifices a systematic query mechanism equivalent to the widely used SQL. To overcome this disadvantage, we propose embeddable string-search commands on a Wiki-based system and designed a half-formatted database. As proof of principle, a database of flavonoid with 6902 molecular structures from over 1687 plant species was implemented on MediaWiki, the background system of Wikipedia. Registered users can describe any information in an arbitrary format. Structured part is subject to text-string searches to realize relational operations. The system was written in PHP language as the extension of MediaWiki. All modifications are open-source and publicly available. This scheme benefits from both the free-formatted Wiki style and the concise and structured relational-database style. MediaWiki supports multi-user environments for document management, and the cost for database maintenance is alleviated.
Search extension transforms Wiki into a relational system: A case for flavonoid metabolite database
Arita, Masanori; Suwa, Kazuhiro
2008-01-01
Background In computer science, database systems are based on the relational model founded by Edgar Codd in 1970. On the other hand, in the area of biology the word 'database' often refers to loosely formatted, very large text files. Although such bio-databases may describe conflicts or ambiguities (e.g. a protein pair do and do not interact, or unknown parameters) in a positive sense, the flexibility of the data format sacrifices a systematic query mechanism equivalent to the widely used SQL. Results To overcome this disadvantage, we propose embeddable string-search commands on a Wiki-based system and designed a half-formatted database. As proof of principle, a database of flavonoid with 6902 molecular structures from over 1687 plant species was implemented on MediaWiki, the background system of Wikipedia. Registered users can describe any information in an arbitrary format. Structured part is subject to text-string searches to realize relational operations. The system was written in PHP language as the extension of MediaWiki. All modifications are open-source and publicly available. Conclusion This scheme benefits from both the free-formatted Wiki style and the concise and structured relational-database style. MediaWiki supports multi-user environments for document management, and the cost for database maintenance is alleviated. PMID:18822113
DrugQuest - a text mining workflow for drug association discovery.
Papanikolaou, Nikolas; Pavlopoulos, Georgios A; Theodosiou, Theodosios; Vizirianakis, Ioannis S; Iliopoulos, Ioannis
2016-06-06
Text mining and data integration methods are gaining ground in the field of health sciences due to the exponential growth of bio-medical literature and information stored in biological databases. While such methods mostly try to extract bioentity associations from PubMed, very few of them are dedicated in mining other types of repositories such as chemical databases. Herein, we apply a text mining approach on the DrugBank database in order to explore drug associations based on the DrugBank "Description", "Indication", "Pharmacodynamics" and "Mechanism of Action" text fields. We apply Name Entity Recognition (NER) techniques on these fields to identify chemicals, proteins, genes, pathways, diseases, and we utilize the TextQuest algorithm to find additional biologically significant words. Using a plethora of similarity and partitional clustering techniques, we group the DrugBank records based on their common terms and investigate possible scenarios why these records are clustered together. Different views such as clustered chemicals based on their textual information, tag clouds consisting of Significant Terms along with the terms that were used for clustering are delivered to the user through a user-friendly web interface. DrugQuest is a text mining tool for knowledge discovery: it is designed to cluster DrugBank records based on text attributes in order to find new associations between drugs. The service is freely available at http://bioinformatics.med.uoc.gr/drugquest .
The Tunneling Method for Global Optimization in Multidimensional Scaling.
ERIC Educational Resources Information Center
Groenen, Patrick J. F.; Heiser, Willem J.
1996-01-01
A tunneling method for global minimization in multidimensional scaling is introduced and adjusted for multidimensional scaling with general Minkowski distances. The method alternates a local search step with a tunneling step in which a different configuration is sought with the same STRESS implementation. (SLD)
Multidimensional Poverty and Health Status as a Predictor of Chronic Income Poverty.
Callander, Emily J; Schofield, Deborah J
2015-12-01
Longitudinal analysis of Wave 5 to 10 of the nationally representative Household, Income and Labour Dynamics in Australia dataset was undertaken to assess whether multidimensional poverty status can predict chronic income poverty. Of those who were multidimensionally poor (low income plus poor health or poor health and insufficient education attainment) in 2007, and those who were in income poverty only (no other forms of disadvantage) in 2007, a greater proportion of those in multidimensional poverty continued to be in income poverty for the subsequent 5 years through to 2012. People who were multidimensionally poor in 2007 had 2.17 times the odds of being in income poverty each year through to 2012 than those who were in income poverty only in 2005 (95% CI: 1.23-3.83). Multidimensional poverty measures are a useful tool for policymakers to identify target populations for policies aiming to improve equity and reduce chronic disadvantage. Copyright © 2014 John Wiley & Sons, Ltd.
NASA Astrophysics Data System (ADS)
Paardekooper, S.-J.
2017-08-01
We present a new method for numerical hydrodynamics which uses a multidimensional generalization of the Roe solver and operates on an unstructured triangular mesh. The main advantage over traditional methods based on Riemann solvers, which commonly use one-dimensional flux estimates as building blocks for a multidimensional integration, is its inherently multidimensional nature, and as a consequence its ability to recognize multidimensional stationary states that are not hydrostatic. A second novelty is the focus on graphics processing units (GPUs). By tailoring the algorithms specifically to GPUs, we are able to get speedups of 100-250 compared to a desktop machine. We compare the multidimensional upwind scheme to a traditional, dimensionally split implementation of the Roe solver on several test problems, and we find that the new method significantly outperforms the Roe solver in almost all cases. This comes with increased computational costs per time-step, which makes the new method approximately a factor of 2 slower than a dimensionally split scheme acting on a structured grid.
WebCSD: the online portal to the Cambridge Structural Database
Thomas, Ian R.; Bruno, Ian J.; Cole, Jason C.; Macrae, Clare F.; Pidcock, Elna; Wood, Peter A.
2010-01-01
WebCSD, a new web-based application developed by the Cambridge Crystallographic Data Centre, offers fast searching of the Cambridge Structural Database using only a standard internet browser. Search facilities include two-dimensional substructure, molecular similarity, text/numeric and reduced cell searching. Text, chemical diagrams and three-dimensional structural information can all be studied in the results browser using the efficient entry summaries and embedded three-dimensional viewer. PMID:22477776
Knudsen, Keith L.; Noller, Jay S.; Sowers, Janet M.; Lettis, William R.
1997-01-01
This Open-File report is a digital geologic map database. This pamphlet serves to introduce and describe the digital data. There are no paper maps included in the Open-File report. The report does include, however, PostScript plot files containing the images of the geologic map sheets with explanations, as well as the accompanying text describing the geology of the area. For those interested in a paper plot of information contained in the database or in obtaining the PostScript plot files, please see the section entitled 'For Those Who Aren't Familiar With Digital Geologic Map Databases' below. This digital map database, compiled from previously unpublished data, and new mapping by the authors, represents the general distribution of surficial deposits in the San Francisco bay region. Together with the accompanying text file (sf_geo.txt or sf_geo.pdf), it provides current information on Quaternary geology and liquefaction susceptibility of the San Francisco, California, 1:100,000 quadrangle. The database delineates map units that are identified by general age and lithology following the stratigraphic nomenclature of the U.S. Geological Survey. The scale of the source maps limits the spatial resolution (scale) of the database to 1:100,000 or smaller. The content and character of the database, as well as three methods of obtaining the database, are described below.
Classroom Laboratory Report: Using an Image Database System in Engineering Education.
ERIC Educational Resources Information Center
Alam, Javed; And Others
1991-01-01
Describes an image database system assembled using separate computer components that was developed to overcome text-only computer hardware storage and retrieval limitations for a pavement design class. (JJK)
De-identifying an EHR database - anonymity, correctness and readability of the medical record.
Pantazos, Kostas; Lauesen, Soren; Lippert, Soren
2011-01-01
Electronic health records (EHR) contain a large amount of structured data and free text. Exploring and sharing clinical data can improve healthcare and facilitate the development of medical software. However, revealing confidential information is against ethical principles and laws. We de-identified a Danish EHR database with 437,164 patients. The goal was to generate a version with real medical records, but related to artificial persons. We developed a de-identification algorithm that uses lists of named entities, simple language analysis, and special rules. Our algorithm consists of 3 steps: collect lists of identifiers from the database and external resources, define a replacement for each identifier, and replace identifiers in structured data and free text. Some patient records could not be safely de-identified, so the de-identified database has 323,122 patient records with an acceptable degree of anonymity, readability and correctness (F-measure of 95%). The algorithm has to be adjusted for each culture, language and database.
Computerized lung cancer malignancy level analysis using 3D texture features
NASA Astrophysics Data System (ADS)
Sun, Wenqing; Huang, Xia; Tseng, Tzu-Liang; Zhang, Jianying; Qian, Wei
2016-03-01
Based on the likelihood of malignancy, the nodules are classified into five different levels in Lung Image Database Consortium (LIDC) database. In this study, we tested the possibility of using threedimensional (3D) texture features to identify the malignancy level of each nodule. Five groups of features were implemented and tested on 172 nodules with confident malignancy levels from four radiologists. These five feature groups are: grey level co-occurrence matrix (GLCM) features, local binary pattern (LBP) features, scale-invariant feature transform (SIFT) features, steerable features, and wavelet features. Because of the high dimensionality of our proposed features, multidimensional scaling (MDS) was used for dimension reduction. RUSBoost was applied for our extracted features for classification, due to its advantages in handling imbalanced dataset. Each group of features and the final combined features were used to classify nodules highly suspicious for cancer (level 5) and moderately suspicious (level 4). The results showed that the area under the curve (AUC) and accuracy are 0.7659 and 0.8365 when using the finalized features. These features were also tested on differentiating benign and malignant cases, and the reported AUC and accuracy were 0.8901 and 0.9353.
DIRT: The Dust InfraRed Toolbox
NASA Astrophysics Data System (ADS)
Pound, M. W.; Wolfire, M. G.; Mundy, L. G.; Teuben, P. J.; Lord, S.
We present DIRT, a Java applet geared toward modeling a variety of processes in envelopes of young and evolved stars. Users can automatically and efficiently search grids of pre-calculated models to fit their data. A large set of physical parameters and dust types are included in the model database, which contains over 500,000 models. The computing cluster for the database is described in the accompanying paper by Teuben et al. (2000). A typical user query will return about 50-100 models, which the user can then interactively filter as a function of 8 model parameters (e.g., extinction, size, flux, luminosity). A flexible, multi-dimensional plotter (Figure 1) allows users to view the models, rotate them, tag specific parameters with color or symbol size, and probe individual model points. For any given model, auxiliary plots such as dust grain properties, radial intensity profiles, and the flux as a function of wavelength and beamsize can be viewed. The user can fit observed data to several models simultaneously and see the results of the fit; the best fit is automatically selected for plotting. The URL for this project is http://dustem.astro.umd.edu.
[Terrorism and human behavior].
Leistedt, S J
2018-04-01
Theories of religion are essential for understanding current trends in terrorist activities. The aim of this work is to clarify religion's role in facilitating terror and outline in parallel with recent theoretical developments on terrorism and human behaviour. Several databases were used such as PubCentral, Scopus, Medline and Science Direct. The search terms "terrorism", "social psychology", "religion", "evolution", and "cognition" were used to identify relevant studies in the databases. This work examines, in a multidimensional way, how terrorists employ these features of religion to achieve their goals. In the same way, it describes how terrorists use rituals to conditionally associate emotions with sanctified symbols that are emotionally evocative and motivationally powerful, fostering group solidarity, trust, and cooperation. Religious beliefs, including promised rewards in the afterlife, further serve to facilitate cooperation by altering the perceived payoffs of costly actions, including suicide bombing. The adolescent pattern of brain development is unique, and young adulthood presents an ideal developmental stage to attract recruits and enlist them in high-risk behaviors. This work offers insights, based on this translational analysis, concerning the links between religion, terrorism and human behavior. Copyright © 2017 L'Encéphale, Paris. Published by Elsevier Masson SAS. All rights reserved.
Enhancing Student Motivation and Engagement: The Effects of a Multidimensional Intervention
ERIC Educational Resources Information Center
Martin, Andrew J.
2008-01-01
The present study sought to investigate the effects of a multidimensional educational intervention on high school students' motivation and engagement. The intervention incorporated: (a) multidimensional targets of motivation and engagement, (b) empirically derived intervention methodology, (c) research-based risk and protective factors, (d)…
Data Compression in Full-Text Retrieval Systems.
ERIC Educational Resources Information Center
Bell, Timothy C.; And Others
1993-01-01
Describes compression methods for components of full-text systems such as text databases on CD-ROM. Topics discussed include storage media; structures for full-text retrieval, including indexes, inverted files, and bitmaps; compression tools; memory requirements during retrieval; and ranking and information retrieval. (Contains 53 references.)…
Heteronuclear Multidimensional Protein NMR in a Teaching Laboratory
ERIC Educational Resources Information Center
Wright, Nathan T.
2016-01-01
Heteronuclear multidimensional NMR techniques are commonly used to study protein structure, function, and dynamics, yet they are rarely taught at the undergraduate level. Here, we describe a senior undergraduate laboratory where students collect, process, and analyze heteronuclear multidimensional NMR experiments using an unstudied Ig domain (Ig2…
Compressed Continuous Computation v. 12/20/2016
DOE Office of Scientific and Technical Information (OSTI.GOV)
Gorodetsky, Alex
2017-02-17
A library for performing numerical computation with low-rank functions. The (C3) library enables performing continuous linear and multilinear algebra with multidimensional functions. Common tasks include taking "matrix" decompositions of vector- or matrix-valued functions, approximating multidimensional functions in low-rank format, adding or multiplying functions together, integrating multidimensional functions.
The Discriminating Power of Items that Measure More than One Dimension.
ERIC Educational Resources Information Center
Reckase, Mark D.
The work presented in this paper defined conceptually the concepts of multidimensional discrimination and information, derived mathematical expressions for the concepts for a particular multidimensional item response theory (IRT) model, and applied the concepts to actual test data. Multidimensional discrimination was defined as a function of the…
Multidimensional Computerized Adaptive Testing for Indonesia Junior High School Biology
ERIC Educational Resources Information Center
Kuo, Bor-Chen; Daud, Muslem; Yang, Chih-Wei
2015-01-01
This paper describes a curriculum-based multidimensional computerized adaptive test that was developed for Indonesia junior high school Biology. In adherence to the Indonesian curriculum of different Biology dimensions, 300 items was constructed, and then tested to 2238 students. A multidimensional random coefficients multinomial logit model was…
Supervised and Unsupervised Learning of Multidimensional Acoustic Categories
ERIC Educational Resources Information Center
Goudbeek, Martijn; Swingley, Daniel; Smits, Roel
2009-01-01
Learning to recognize the contrasts of a language-specific phonemic repertoire can be viewed as forming categories in a multidimensional psychophysical space. Research on the learning of distributionally defined visual categories has shown that categories defined over 1 dimension are easy to learn and that learning multidimensional categories is…
Health, Wealth and Wisdom: Exploring Multidimensional Inequality in a Developing Country
ERIC Educational Resources Information Center
Nilsson, Therese
2010-01-01
Despite a broad theoretical literature on multidimensional inequality and a widespread belief that welfare is not synonymous to income--not the least in a developing context--empirical inequality examinations rarely includes several welfare attributes. We explore three techniques on how to evaluate multidimensional inequality using Zambian…
Multidimensional Physical Self-Concept of Athletes with Physical Disabilities
ERIC Educational Resources Information Center
Shapiro, Deborah R.; Martin, Jeffrey J.
2010-01-01
The purposes of this investigation were first to predict reported PA (physical activity) behavior and self-esteem using a multidimensional physical self-concept model and second to describe perceptions of multidimensional physical self-concept (e.g., strength, endurance, sport competence) among athletes with physical disabilities. Athletes (N =…
Meyers, Charles E.; Davidson, George S.; Johnson, David K.; Hendrickson, Bruce A.; Wylie, Brian N.
1999-01-01
A method of data mining represents related items in a multidimensional space. Distance between items in the multidimensional space corresponds to the extent of relationship between the items. The user can select portions of the space to perceive. The user also can interact with and control the communication of the space, focusing attention on aspects of the space of most interest. The multidimensional spatial representation allows more ready comprehension of the structure of the relationships among the items.
A Conceptual Model for Multidimensional Analysis of Documents
NASA Astrophysics Data System (ADS)
Ravat, Franck; Teste, Olivier; Tournier, Ronan; Zurlfluh, Gilles
Data warehousing and OLAP are mainly used for the analysis of transactional data. Nowadays, with the evolution of Internet, and the development of semi-structured data exchange format (such as XML), it is possible to consider entire fragments of data such as documents as analysis sources. As a consequence, an adapted multidimensional analysis framework needs to be provided. In this paper, we introduce an OLAP multidimensional conceptual model without facts. This model is based on the unique concept of dimensions and is adapted for multidimensional document analysis. We also provide a set of manipulation operations.
A review of snapshot multidimensional optical imaging: measuring photon tags in parallel
Gao, Liang; Wang, Lihong V.
2015-01-01
Multidimensional optical imaging has seen remarkable growth in the past decade. Rather than measuring only the two-dimensional spatial distribution of light, as in conventional photography, multidimensional optical imaging captures light in up to nine dimensions, providing unprecedented information about incident photons’ spatial coordinates, emittance angles, wavelength, time, and polarization. Multidimensional optical imaging can be accomplished either by scanning or parallel acquisition. Compared with scanning-based imagers, parallel acquisition—also dubbed snapshot imaging—has a prominent advantage in maximizing optical throughput, particularly when measuring a datacube of high dimensions. Here, we first categorize snapshot multidimensional imagers based on their acquisition and image reconstruction strategies, then highlight the snapshot advantage in the context of optical throughput, and finally we discuss their state-of-the-art implementations and applications. PMID:27134340
A multidimensional subdiffusion model: An arbitrage-free market
NASA Astrophysics Data System (ADS)
Li, Guo-Hua; Zhang, Hong; Luo, Mao-Kang
2012-12-01
To capture the subdiffusive characteristics of financial markets, the subordinated process, directed by the inverse α-stale subordinator Sα(t) for 0 < α < 1, has been employed as the model of asset prices. In this article, we introduce a multidimensional subdiffusion model that has a bond and K correlated stocks. The stock price process is a multidimensional subdiffusion process directed by the inverse α-stable subordinator. This model describes the period of stagnation for each stock and the behavior of the dependency between multiple stocks. Moreover, we derive the multidimensional fractional backward Kolmogorov equation for the subordinated process using the Laplace transform technique. Finally, using a martingale approach, we prove that the multidimensional subdiffusion model is arbitrage-free, and also gives an arbitrage-free pricing rule for contingent claims associated with the martingale measure.
Progress in multi-dimensional upwind differencing
NASA Technical Reports Server (NTRS)
Vanleer, Bram
1992-01-01
Multi-dimensional upwind-differencing schemes for the Euler equations are reviewed. On the basis of the first-order upwind scheme for a one-dimensional convection equation, the two approaches to upwind differencing are discussed: the fluctuation approach and the finite-volume approach. The usual extension of the finite-volume method to the multi-dimensional Euler equations is not entirely satisfactory, because the direction of wave propagation is always assumed to be normal to the cell faces. This leads to smearing of shock and shear waves when these are not grid-aligned. Multi-directional methods, in which upwind-biased fluxes are computed in a frame aligned with a dominant wave, overcome this problem, but at the expense of robustness. The same is true for the schemes incorporating a multi-dimensional wave model not based on multi-dimensional data but on an 'educated guess' of what they could be. The fluctuation approach offers the best possibilities for the development of genuinely multi-dimensional upwind schemes. Three building blocks are needed for such schemes: a wave model, a way to achieve conservation, and a compact convection scheme. Recent advances in each of these components are discussed; putting them all together is the present focus of a worldwide research effort. Some numerical results are presented, illustrating the potential of the new multi-dimensional schemes.
OSTMED.DR®, an Osteopathic Medicine Digital Library.
Fitterling, Lori; Powers, Elaine; Vardell, Emily
2018-01-01
The OSTMED.DR® database provides access to both citation and full-text osteopathic literature, including the Journal of the American Osteopathic Association. Currently, it is a free database searchable using basic and advanced search features.
An Update on Electronic Information Sources.
ERIC Educational Resources Information Center
Ackerman, Katherine
1987-01-01
This review of new developments and products in online services discusses trends in travel related services; full text databases; statistical source databases; an emphasis on regional and international business news; and user friendly systems. (Author/CLB)
ERIC Educational Resources Information Center
Li, Yiu-On; Leung, Shirley W.
2001-01-01
Discussion of aggregator databases focuses on a project at the Hong Kong Baptist University library to integrate full-text electronic journal titles from three unstable aggregator databases into its online public access catalog (OPAC). Explains the development of the electronic journal computer program (EJCOP) to generate MARC records for…
Chen, Josephine; Zhao, Po; Massaro, Donald; Clerch, Linda B; Almon, Richard R; DuBois, Debra C; Jusko, William J; Hoffman, Eric P
2004-01-01
Publicly accessible DNA databases (genome browsers) are rapidly accelerating post-genomic research (see http://www.genome.ucsc.edu/), with integrated genomic DNA, gene structure, EST/ splicing and cross-species ortholog data. DNA databases have relatively low dimensionality; the genome is a linear code that anchors all associated data. In contrast, RNA expression and protein databases need to be able to handle very high dimensional data, with time, tissue, cell type and genes, as interrelated variables. The high dimensionality of microarray expression profile data, and the lack of a standard experimental platform have complicated the development of web-accessible databases and analytical tools. We have designed and implemented a public resource of expression profile data containing 1024 human, mouse and rat Affymetrix GeneChip expression profiles, generated in the same laboratory, and subject to the same quality and procedural controls (Public Expression Profiling Resource; PEPR). Our Oracle-based PEPR data warehouse includes a novel time series query analysis tool (SGQT), enabling dynamic generation of graphs and spreadsheets showing the action of any transcript of interest over time. In this report, we demonstrate the utility of this tool using a 27 time point, in vivo muscle regeneration series. This data warehouse and associated analysis tools provides access to multidimensional microarray data through web-based interfaces, both for download of all types of raw data for independent analysis, and also for straightforward gene-based queries. Planned implementations of PEPR will include web-based remote entry of projects adhering to quality control and standard operating procedure (QC/SOP) criteria, and automated output of alternative probe set algorithms for each project (see http://microarray.cnmcresearch.org/pgadatatable.asp).
Context indexing of digital cardiac ultrasound records in PACS
NASA Astrophysics Data System (ADS)
Lobodzinski, S. Suave; Meszaros, Georg N.
1998-07-01
Recent wide adoption of the DICOM 3.0 standard by ultrasound equipment vendors created a need for practical clinical implementations of cardiac imaging study visualization, management and archiving, DICOM 3.0 defines only a logical and physical format for exchanging image data (still images, video, patient and study demographics). All DICOM compliant imaging studies must presently be archived on a 650 Mb recordable compact disk. This is a severe limitation for ultrasound applications where studies of 3 to 10 minutes long are a common practice. In addition, DICOM digital echocardiography objects require physiological signal indexing, content segmentation and characterization. Since DICOM 3.0 is an interchange standard only, it does not define how to database composite video objects. The goal of this research was therefore to address the issues of efficient storage, retrieval and management of DICOM compliant cardiac video studies in a distributed PACS environment. Our Web based implementation has the advantage of accommodating both DICOM defined entity-relation modules (equipment data, patient data, video format, etc.) in standard relational database tables and digital indexed video with its attributes in an object relational database. Object relational data model facilitates content indexing of full motion cardiac imaging studies through bi-directional hyperlink generation that tie searchable video attributes and related objects to individual video frames in the temporal domain. Benefits realized from use of bi-directionally hyperlinked data models in an object relational database include: (1) real time video indexing during image acquisition, (2) random access and frame accurate instant playback of previously recorded full motion imaging data, and (3) time savings from faster and more accurate access to data through multiple navigation mechanisms such as multidimensional queries on an index, queries on a hyperlink attribute, free search and browsing.
Chen, Josephine; Zhao, Po; Massaro, Donald; Clerch, Linda B.; Almon, Richard R.; DuBois, Debra C.; Jusko, William J.; Hoffman, Eric P.
2004-01-01
Publicly accessible DNA databases (genome browsers) are rapidly accelerating post-genomic research (see http://www.genome.ucsc.edu/), with integrated genomic DNA, gene structure, EST/ splicing and cross-species ortholog data. DNA databases have relatively low dimensionality; the genome is a linear code that anchors all associated data. In contrast, RNA expression and protein databases need to be able to handle very high dimensional data, with time, tissue, cell type and genes, as interrelated variables. The high dimensionality of microarray expression profile data, and the lack of a standard experimental platform have complicated the development of web-accessible databases and analytical tools. We have designed and implemented a public resource of expression profile data containing 1024 human, mouse and rat Affymetrix GeneChip expression profiles, generated in the same laboratory, and subject to the same quality and procedural controls (Public Expression Profiling Resource; PEPR). Our Oracle-based PEPR data warehouse includes a novel time series query analysis tool (SGQT), enabling dynamic generation of graphs and spreadsheets showing the action of any transcript of interest over time. In this report, we demonstrate the utility of this tool using a 27 time point, in vivo muscle regeneration series. This data warehouse and associated analysis tools provides access to multidimensional microarray data through web-based interfaces, both for download of all types of raw data for independent analysis, and also for straightforward gene-based queries. Planned implementations of PEPR will include web-based remote entry of projects adhering to quality control and standard operating procedure (QC/SOP) criteria, and automated output of alternative probe set algorithms for each project (see http://microarray.cnmcresearch.org/pgadatatable.asp). PMID:14681485
NASA Technical Reports Server (NTRS)
2002-01-01
A system that retrieves problem reports from a NASA database is described. The database is queried with natural language questions. Part-of-speech tags are first assigned to each word in the question using a rule based tagger. A partial parse of the question is then produced with independent sets of deterministic finite state a utomata. Using partial parse information, a look up strategy searches the database for problem reports relevant to the question. A bigram stemmer and irregular verb conjugates have been incorporated into the system to improve accuracy. The system is evaluated by a set of fifty five questions posed by NASA engineers. A discussion of future research is also presented.
How to locate and appraise qualitative research in complementary and alternative medicine
2013-01-01
Background The aim of this publication is to present a case study of how to locate and appraise qualitative studies for the conduct of a meta-ethnography in the field of complementary and alternative medicine (CAM). CAM is commonly associated with individualized medicine. However, one established scientific approach to the individual, qualitative research, thus far has been explicitly used very rarely. This article demonstrates a case example of how qualitative research in the field of CAM studies was identified and critically appraised. Methods Several search terms and techniques were tested for the identification and appraisal of qualitative CAM research in the conduct of a meta-ethnography. Sixty-seven electronic databases were searched for the identification of qualitative CAM trials, including CAM databases, nursing, nutrition, psychological, social, medical databases, the Cochrane Library and DIMDI. Results 9578 citations were screened, 223 articles met the pre-specified inclusion criteria, 63 full text publications were reviewed, 38 articles were appraised qualitatively and 30 articles were included. The search began with PubMed, yielding 87% of the included publications of all databases with few additional relevant findings in the specific databases. CINHAL and DIMDI also revealed a high number of precise hits. Although CAMbase and CAM-QUEST® focus on CAM research only, almost no hits of qualitative trials were found there. Searching with broad text terms was the most effective search strategy in all databases. Conclusions This publication presents a case study on how to locate and appraise qualitative studies in the field of CAM. The example shows that the literature search for qualitative studies in the field of CAM is most effective when the search is begun in PubMed followed by CINHAL or DIMDI using broad text terms. Exclusive CAM databases delivered no additional findings to locate qualitative CAM studies. PMID:23731997
How to locate and appraise qualitative research in complementary and alternative medicine.
Franzel, Brigitte; Schwiegershausen, Martina; Heusser, Peter; Berger, Bettina
2013-06-03
The aim of this publication is to present a case study of how to locate and appraise qualitative studies for the conduct of a meta-ethnography in the field of complementary and alternative medicine (CAM). CAM is commonly associated with individualized medicine. However, one established scientific approach to the individual, qualitative research, thus far has been explicitly used very rarely. This article demonstrates a case example of how qualitative research in the field of CAM studies was identified and critically appraised. Several search terms and techniques were tested for the identification and appraisal of qualitative CAM research in the conduct of a meta-ethnography. Sixty-seven electronic databases were searched for the identification of qualitative CAM trials, including CAM databases, nursing, nutrition, psychological, social, medical databases, the Cochrane Library and DIMDI. 9578 citations were screened, 223 articles met the pre-specified inclusion criteria, 63 full text publications were reviewed, 38 articles were appraised qualitatively and 30 articles were included. The search began with PubMed, yielding 87% of the included publications of all databases with few additional relevant findings in the specific databases. CINHAL and DIMDI also revealed a high number of precise hits. Although CAMbase and CAM-QUEST® focus on CAM research only, almost no hits of qualitative trials were found there. Searching with broad text terms was the most effective search strategy in all databases. This publication presents a case study on how to locate and appraise qualitative studies in the field of CAM. The example shows that the literature search for qualitative studies in the field of CAM is most effective when the search is begun in PubMed followed by CINHAL or DIMDI using broad text terms. Exclusive CAM databases delivered no additional findings to locate qualitative CAM studies.
Aquatic exercise training for fibromyalgia.
Bidonde, Julia; Busch, Angela J; Webber, Sandra C; Schachter, Candice L; Danyliw, Adrienne; Overend, Tom J; Richards, Rachel S; Rader, Tamara
2014-10-28
Exercise training is commonly recommended for individuals with fibromyalgia. This review examined the effects of supervised group aquatic training programs (led by an instructor). We defined aquatic training as exercising in a pool while standing at waist, chest, or shoulder depth. This review is part of the update of the 'Exercise for treating fibromyalgia syndrome' review first published in 2002, and previously updated in 2007. The objective of this systematic review was to evaluate the benefits and harms of aquatic exercise training in adults with fibromyalgia. We searched The Cochrane Library 2013, Issue 2 (Cochrane Database of Systematic Reviews, Database of Abstracts of Reviews of Effects, Cochrane Central Register of Controlled Trials, Health Technology Assessment Database, NHS Economic Evaluation Database), MEDLINE, EMBASE, CINAHL, PEDro, Dissertation Abstracts, WHO international Clinical Trials Registry Platform, and AMED, as well as other sources (i.e., reference lists from key journals, identified articles, meta-analyses, and reviews of all types of treatment for fibromyalgia) from inception to October 2013. Using Cochrane methods, we screened citations, abstracts, and full-text articles. Subsequently, we identified aquatic exercise training studies. Selection criteria were: a) full-text publication of a randomized controlled trial (RCT) in adults diagnosed with fibromyalgia based on published criteria, and b) between-group data for an aquatic intervention and a control or other intervention. We excluded studies if exercise in water was less than 50% of the full intervention. We independently assessed risk of bias and extracted data (24 outcomes), of which we designated seven as major outcomes: multidimensional function, self reported physical function, pain, stiffness, muscle strength, submaximal cardiorespiratory function, withdrawal rates and adverse effects. We resolved discordance through discussion. We evaluated interventions using mean differences (MD) or standardized mean differences (SMD) and 95% confidence intervals (95% CI). Where two or more studies provided data for an outcome, we carried out meta-analysis. In addition, we set and used a 15% threshold for calculation of clinically relevant differences. We included 16 aquatic exercise training studies (N = 881; 866 women and 15 men). Nine studies compared aquatic exercise to control, five studies compared aquatic to land-based exercise, and two compared aquatic exercise to a different aquatic exercise program.We rated the risk of bias related to random sequence generation (selection bias), incomplete outcome data (attrition bias), selective reporting (reporting bias), blinding of outcome assessors (detection bias), and other bias as low. We rated blinding of participants and personnel (selection and performance bias) and allocation concealment (selection bias) as low risk and unclear. The assessment of the evidence showed limitations related to imprecision, high statistical heterogeneity, and wide confidence intervals. Aquatic versus controlWe found statistically significant improvements (P value < 0.05) in all of the major outcomes. Based on a 100-point scale, multidimensional function improved by six units (MD -5.97, 95% CI -9.06 to -2.88; number needed to treat (NNT) 5, 95% CI 3 to 9), self reported physical function by four units (MD -4.35, 95% CI -7.77 to -0.94; NNT 6, 95% CI 3 to 22), pain by seven units (MD -6.59, 95% CI -10.71 to -2.48; NNT 5, 95% CI 3 to 8), and stiffness by 18 units (MD -18.34, 95% CI -35.75 to -0.93; NNT 3, 95% CI 2 to 24) more in the aquatic than the control groups. The SMD for muscle strength as measured by knee extension and hand grip was 0.63 standard deviations higher compared to the control group (SMD 0.63, 95% CI 0.20 to 1.05; NNT 4, 95% CI 3 to 12) and cardiovascular submaximal function improved by 37 meters on six-minute walk test (95% CI 4.14 to 69.92). Only two major outcomes, stiffness and muscle strength, met the 15% threshold for clinical relevance (improved by 27% and 37% respectively). Withdrawals were similar in the aquatic and control groups and adverse effects were poorly reported, with no serious adverse effects reported. Aquatic versus land-basedThere were no statistically significant differences between interventions for multidimensional function, self reported physical function, pain or stiffness: 0.91 units (95% CI -4.01 to 5.83), -5.85 units (95% CI -12.33 to 0.63), -0.75 units (95% CI -10.72 to 9.23), and two units (95% CI -8.88 to 1.28) respectively (all based on a 100-point scale), or in submaximal cardiorespiratory function (three seconds on a 100-meter walk test, 95% CI -1.77 to 7.77). We found a statistically significant difference between interventions for strength, favoring land-based training (2.40 kilo pascals grip strength, 95% CI 4.52 to 0.28). None of the outcomes in the aquatic versus land comparison reached clinically relevant differences of 15%. Withdrawals were similar in the aquatic and land groups and adverse effects were poorly reported, with no serious adverse effects in either group. Aquatic versus aquatic (Ai Chi versus stretching in the water, exercise in pool water versus exercise in sea water)Among the major outcomes the only statistically significant difference between interventions was for stiffness, favoring Ai Chi (1.00 on a 100-point scale, 95% CI 0.31 to 1.69). Low to moderate quality evidence relative to control suggests that aquatic training is beneficial for improving wellness, symptoms, and fitness in adults with fibromyalgia. Very low to low quality evidence suggests that there are benefits of aquatic and land-based exercise, except in muscle strength (very low quality evidence favoring land). No serious adverse effects were reported.
Beyond solicitousness: a comprehensive review on informal pain-related social support.
Bernardes, Sónia F; Forgeron, Paula; Fournier, Karine; Reszel, Jessica
2017-11-01
Adults with chronic pain cite social support (SS) as an important resource. Research has mostly focused on general SS or pain-specific solicitousness, resulting in a limited understanding of the role of SS in pain experiences. Drawing on SS theoretical models, this review aimed to understand how pain-related SS has been conceptualized and measured and how its relationship with pain experiences has been investigated. Arksey and O'Malley scoping review framework guided the study. A database search (2000-2015) was conducted in PsycINFO, CINAHL, MEDLINE, and EMBASE using a combination of subject headings/keywords on pain and SS; 3864 citations were screened; 101 full texts were assessed for eligibility; references of 52 papers were hand searched. Fifty-three studies were included. Most studies were either a-theoretical or drew upon the operant conditioning model. There are several self-report measures and observational systems to operationalize pain-related SS. However, the Multidimensional Pain Inventory remains the most often used, accounting for the centrality of the concept of solicitousness in the literature. Most studies focused on individuals with chronic pain self-report of spousal pain-related SS and investigated its main effects on pain outcomes. Only a minority investigated the role of pain SS within the stress and coping process (as a buffer or mediator). Little is known about mediating pathways, contextual modulation of the effectiveness of SS exchanges, and there are practically no SS-based intervention studies. Drawing on general SS models, the main gaps in pain-related SS research are discussed and research directions for moving this literature beyond solicitousness are proposed.
Gálvez, Carmen
2016-12-01
Identifying research lines is essential to understand the knowledge structure of a scientific domain. The aim of this study was to identify the main research topics of within the domain of public health, in the Revista Española de Saslud Pública during 2006-2015. Original articles included in the Social Sciences Citation Index (SSCI) database, available online through the Web of Science (WoS), were selected. The analysis units used were the keywords, KeyWords Plus (KW+), extracted automatically by SSCI. With KW+ obtained bibliometric, maps were created using a methodology based on the combination of co-word analysis, co-word analysis, clustering techniques and visualization techniques. We analyzed 512 documents, of which 176 KW+ were obtained with a frequency greater than or equal to 3. The results were bidimensional bibliometric maps with thematic groupings of KW+, representing the main research fronts: i) epidemiology, risk control programs disease and, in general, service organization and health policies; ii) infectious diseases, principally HIV; iii) a progressive increase in several lines interrelated with cardiovascular diseases (CVD); iv) a line multidimensional dedicated to different aspects associated to the quality of life related to health (HRQoL); and v) an emerging line linked to binge drinking. For the multidisciplinary and multidimensional nature of public health, the construction of bibliometric maps is an appropriate methodology to understand the knowledge structure of this scientific domain.
Gerpott, Fabiola H; Balliet, Daniel; Columbus, Simon; Molho, Catherine; de Vries, Reinout E
2017-09-04
Interdependence is a fundamental characteristic of social interactions. Interdependence Theory states that 6 dimensions describe differences between social situations. Here we examine if these 6 dimensions describe how people think about their interdependence with others in a situation. We find that people (in situ and ex situ) can reliably differentiate situations according to 5, but not 6, dimensions of interdependence: (a) mutual dependence, (b) power, (c) conflict, (d) future interdependence, and (e) information certainty. This model offers a unique framework for understanding how people think about social situations compared to another recent model of situation construal (DIAMONDS). Furthermore, we examine factors that are theorized to shape perceptions of interdependence, such as situational cues (e.g., nonverbal behavior) and personality (e.g., HEXACO and Social Value Orientation). We also study the implications of subjective interdependence for emotions and cooperative behavior during social interactions. This model of subjective interdependence explains substantial variation in the emotions people experience in situations (i.e., happiness, sadness, anger, and disgust), and explains 24% of the variance in cooperation, above and beyond the DIAMONDS model. Throughout these studies, we develop and validate a multidimensional measure of subjective outcome interdependence that can be used in diverse situations and relationships-the Situational Interdependence Scale (SIS). We discuss how this model of interdependence can be used to better understand how people think about social situations encountered in close relationships, organizations, and society. (PsycINFO Database Record (c) 2017 APA, all rights reserved).
Future perspective and healthy lifestyle choices in adulthood.
Tasdemir-Ozdes, Aylin; Strickland-Hughes, Carla M; Bluck, Susan; Ebner, Natalie C
2016-09-01
Regardless of age, making healthy lifestyle choices is prudent. Despite that, individuals of all ages sometimes have difficulty choosing the healthy option. We argue that individuals' view of the future and position in the life span affects their current lifestyle choices. We capture the multidimensionality of future thinking by assessing 3 types of future perspective. Younger and older men and women (N = 127) reported global future time perspective, future health perspective, and perceived importance of future health-related events. They also rated their likelihood of making healthy lifestyle choices. As predicted, older participants indicated greater intention to make healthy choices in their current life than did younger participants. Compared to younger participants, older participants reported shorter global future time perspective and anticipated worse future health but perceived future health-related events as more important. Having a positive view of one's future health and seeing future health-related events as important were related to greater intention to make healthy lifestyle choices, but greater global future time perspective was not directly related to healthy choices. However, follow-up analyses suggested that greater global future time perspective indirectly affected healthy choices via a more positive view of future health. None of these relations were moderated by age. Individuals' perspective on the future is shown to be an important multidimensional construct affecting everyday healthy lifestyle choices for both younger and older adults. Implications for encouraging healthy choices across the adult life span are discussed. (PsycINFO Database Record (c) 2016 APA, all rights reserved).
An online analytical processing multi-dimensional data warehouse for malaria data
Madey, Gregory R; Vyushkov, Alexander; Raybaud, Benoit; Burkot, Thomas R; Collins, Frank H
2017-01-01
Abstract Malaria is a vector-borne disease that contributes substantially to the global burden of morbidity and mortality. The management of malaria-related data from heterogeneous, autonomous, and distributed data sources poses unique challenges and requirements. Although online data storage systems exist that address specific malaria-related issues, a globally integrated online resource to address different aspects of the disease does not exist. In this article, we describe the design, implementation, and applications of a multi-dimensional, online analytical processing data warehouse, named the VecNet Data Warehouse (VecNet-DW). It is the first online, globally-integrated platform that provides efficient search, retrieval and visualization of historical, predictive, and static malaria-related data, organized in data marts. Historical and static data are modelled using star schemas, while predictive data are modelled using a snowflake schema. The major goals, characteristics, and components of the DW are described along with its data taxonomy and ontology, the external data storage systems and the logical modelling and physical design phases. Results are presented as screenshots of a Dimensional Data browser, a Lookup Tables browser, and a Results Viewer interface. The power of the DW emerges from integrated querying of the different data marts and structuring those queries to the desired dimensions, enabling users to search, view, analyse, and store large volumes of aggregated data, and responding better to the increasing demands of users. Database URL https://dw.vecnet.org/datawarehouse/ PMID:29220463
A theoretical framework for the associations between identity and psychopathology.
Klimstra, Theo A; Denissen, Jaap J A
2017-11-01
Identity research largely emerged from clinical observations. Decades of empirical work advanced the field in refining existing approaches and adding new approaches. Furthermore, the existence of linkages of identity with psychopathology is now well established. Unfortunately, both the directionality of effects between identity aspects and psychopathology symptoms, and the mechanisms underlying associations are unclear. In the present paper, we present a new framework to inspire hypothesis-driven empirical research to overcome this limitation. The framework has a basic resemblance to theoretical models for the study of personality and psychopathology, so we provide examples of how these might apply to the study of identity. Next, we explain that unique features of identity may come into play in individuals suffering from psychopathology that are mostly related to the content of one's identity. These include pros and cons of identifying with one's diagnostic label. Finally, inspired by Hermans' dialogical self theory and principles derived from Piaget's, Swann's and Kelly's work, we delineate a framework with identity at the core of an individual multidimensional space. In this space, psychopathology symptoms have a known distance (representing relevance) to one's identity, and individual multidimensional spaces are connected to those of other individuals in one's social network. We discuss methodological (quantitative and qualitative, idiographic and nomothetic) and statistical procedures (multilevel models and network models) to test the framework. Resulting evidence can boost the field of identity research in demonstrating its high practical relevance for the emergence and conservation of psychopathology. (PsycINFO Database Record (c) 2017 APA, all rights reserved).
From Population Databases to Research and Informed Health Decisions and Policy.
Machluf, Yossy; Tal, Orna; Navon, Amir; Chaiter, Yoram
2017-01-01
In the era of big data, the medical community is inspired to maximize the utilization and processing of the rapidly expanding medical datasets for clinical-related and policy-driven research. This requires a medical database that can be aggregated, interpreted, and integrated at both the individual and population levels. Policymakers seek data as a lever for wise, evidence-based decision-making and information-driven policy. Yet, bridging the gap between data collection, research, and policymaking, is a major challenge. To bridge this gap, we propose a four-step model: (A) creating a conjoined task force of all relevant parties to declare a national program to promote collaborations; (B) promoting a national digital records project, or at least a network of synchronized and integrated databases, in an accessible transparent manner; (C) creating an interoperative national research environment to enable the analysis of the organized and integrated data and to generate evidence; and (D) utilizing the evidence to improve decision-making, to support a wisely chosen national policy. For the latter purpose, we also developed a novel multidimensional set of criteria to illuminate insights and estimate the risk for future morbidity based on current medical conditions. Used by policymakers, providers of health plans, caregivers, and health organizations, we presume this model will assist transforming evidence generation to support the design of health policy and programs, as well as improved decision-making about health and health care, at all levels: individual, communal, organizational, and national.
Designing a framework of intelligent information processing for dentistry administration data.
Amiri, N; Matthews, D C; Gao, Q
2005-07-01
This study was designed to test a cumulative view of current data in the clinical database at the Faculty of Dentistry, Dalhousie University. We planned to examine associations among demographic factors and treatments. Three tables were selected from the database of the faculty: patient, treatment and procedures. All fields and record numbers in each table were documented. Data was explored using SQL server and Visual Basic and then cleaned by removing incongruent fields. After transformation, a data warehouse was created. This was imported to SQL analysis services manager to create an OLAP (Online Analytic Process) cube. The multidimensional model used for access to data was created using a star schema. Treatment count was the measurement variable. Five dimensions--date, postal code, gender, age group and treatment categories--were used to detect associations. Another data warehouse of 8 tables (international tooth code # 1-8) was created and imported to SAS enterprise miner to complete data mining. Association nodes were used for each table to find sequential associations and minimum criteria were set to 2% of cases. Findings of this study confirmed most assumptions of treatment planning procedures. There were some small unexpected patterns of clinical interest. Further developments are recommended to create predictive models. Recent improvements in information technology offer numerous advantages for conversion of raw data from faculty databases to information and subsequently to knowledge. This knowledge can be used by decision makers, managers, and researchers to answer clinical questions, affect policy change and determine future research needs.
Application Program Interface for the Orion Aerodynamics Database
NASA Technical Reports Server (NTRS)
Robinson, Philip E.; Thompson, James
2013-01-01
The Application Programming Interface (API) for the Crew Exploration Vehicle (CEV) Aerodynamic Database has been developed to provide the developers of software an easily implemented, fully self-contained method of accessing the CEV Aerodynamic Database for use in their analysis and simulation tools. The API is programmed in C and provides a series of functions to interact with the database, such as initialization, selecting various options, and calculating the aerodynamic data. No special functions (file read/write, table lookup) are required on the host system other than those included with a standard ANSI C installation. It reads one or more files of aero data tables. Previous releases of aerodynamic databases for space vehicles have only included data tables and a document of the algorithm and equations to combine them for the total aerodynamic forces and moments. This process required each software tool to have a unique implementation of the database code. Errors or omissions in the documentation, or errors in the implementation, led to a lengthy and burdensome process of having to debug each instance of the code. Additionally, input file formats differ for each space vehicle simulation tool, requiring the aero database tables to be reformatted to meet the tool s input file structure requirements. Finally, the capabilities for built-in table lookup routines vary for each simulation tool. Implementation of a new database may require an update to and verification of the table lookup routines. This may be required if the number of dimensions of a data table exceeds the capability of the simulation tools built-in lookup routines. A single software solution was created to provide an aerodynamics software model that could be integrated into other simulation and analysis tools. The highly complex Orion aerodynamics model can then be quickly included in a wide variety of tools. The API code is written in ANSI C for ease of portability to a wide variety of systems. The input data files are in standard formatted ASCII, also for improved portability. The API contains its own implementation of multidimensional table reading and lookup routines. The same aerodynamics input file can be used without modification on all implementations. The turnaround time from aerodynamics model release to a working implementation is significantly reduced
National Water Quality Standards Database (NWQSD)
The National Water Quality Standards Database (WQSDB) provides access to EPA and state water quality standards (WQS) information in text, tables, and maps. This data source was last updated in December 2007 and will no longer be updated.
A Comparison of Global Indexing Schemes to Facilitate Earth Science Data Management
NASA Astrophysics Data System (ADS)
Griessbaum, N.; Frew, J.; Rilee, M. L.; Kuo, K. S.
2017-12-01
Recent advances in database technology have led to systems optimized for managing petabyte-scale multidimensional arrays. These array databases are a good fit for subsets of the Earth's surface that can be projected into a rectangular coordinate system with acceptable geometric fidelity. However, for global analyses, array databases must address the same distortions and discontinuities that apply to map projections in general. The array database SciDB supports enormous databases spread across thousands of computing nodes. Additionally, the following SciDB characteristics are particularly germane to the coordinate system problem: SciDB efficiently stores and manipulates sparse (i.e. mostly empty) arrays. SciDB arrays have 64-bit indexes. SciDB supports user-defined data types, functions, and operators. We have implemented two geospatial indexing schemes in SciDB. The simplest uses two array dimensions to represent longitude and latitude. For representation as 64-bit integers, the coordinates are multiplied by a scale factor large enough to yield an appropriate Earth surface resolution (e.g., a scale factor of 100,000 yields a resolution of approximately 1m at the equator). Aside from the longitudinal discontinuity, the principal disadvantage of this scheme is its fixed scale factor. The second scheme uses a single array dimension to represent the bit-codes for locations in a hierarchical triangular mesh (HTM) coordinate system. A HTM maps the Earth's surface onto an octahedron, and then recursively subdivides each triangular face to the desired resolution. Earth surface locations are represented as the concatenation of an octahedron face code and a quadtree code within the face. Unlike our integerized lat-lon scheme, the HTM allow for objects of different size (e.g., pixels with differing resolutions) to be represented in the same indexing scheme. We present an evaluation of the relative utility of these two schemes for managing and analyzing MODIS swath data.
ERIC Educational Resources Information Center
Lee, Eunjung
2013-01-01
The purpose of this research was to compare the equating performance of various equating procedures for the multidimensional tests. To examine the various equating procedures, simulated data sets were used that were generated based on a multidimensional item response theory (MIRT) framework. Various equating procedures were examined, including…
Effect Size Measures for Differential Item Functioning in a Multidimensional IRT Model
ERIC Educational Resources Information Center
Suh, Youngsuk
2016-01-01
This study adapted an effect size measure used for studying differential item functioning (DIF) in unidimensional tests and extended the measure to multidimensional tests. Two effect size measures were considered in a multidimensional item response theory model: signed weighted P-difference and unsigned weighted P-difference. The performance of…
On Multidimensional Item Response Theory: A Coordinate-Free Approach. Research Report. ETS RR-07-30
ERIC Educational Resources Information Center
Antal, Tamás
2007-01-01
A coordinate-free definition of complex-structure multidimensional item response theory (MIRT) for dichotomously scored items is presented. The point of view taken emphasizes the possibilities and subtleties of understanding MIRT as a multidimensional extension of the classical unidimensional item response theory models. The main theorem of the…
The Definition of Difficulty and Discrimination for Multidimensional Item Response Theory Models.
ERIC Educational Resources Information Center
Reckase, Mark D.; McKinley, Robert L.
A study was undertaken to develop guidelines for the interpretation of the parameters of three multidimensional item response theory models and to determine the relationship between the parameters and traditional concepts of item difficulty and discrimination. The three models considered were multidimensional extensions of the one-, two-, and…
Bifactor Approach to Modeling Multidimensionality of Physical Self-Perception Profile
ERIC Educational Resources Information Center
Chung, ChihMing; Liao, Xiaolan; Song, Hairong; Lee, Taehun
2016-01-01
The multi-dimensionality of Physical Self-Perception Profile (PSPP) has been acknowledged by the use of correlated-factor model and second-order model. In this study, the authors critically endorse the bifactor model, as a substitute to address the multi-dimensionality of PSPP. To cross-validate the models, analyses are conducted first in…
Item Vector Plots for the Multidimensional Three-Parameter Logistic Model
ERIC Educational Resources Information Center
Bryant, Damon; Davis, Larry
2011-01-01
This brief technical note describes how to construct item vector plots for dichotomously scored items fitting the multidimensional three-parameter logistic model (M3PLM). As multidimensional item response theory (MIRT) shows promise of being a very useful framework in the test development life cycle, graphical tools that facilitate understanding…
ERIC Educational Resources Information Center
Coromaldi, Manuela; Zoli, Mariangela
2012-01-01
Theoretical and empirical studies have recently adopted a multidimensional concept of poverty. There is considerable debate about the most appropriate degree of multidimensionality to retain in the analysis. In this work we add to the received literature in two ways. First, we derive indicators of multiple deprivation by applying a particular…
Evaluating Item Fit for Multidimensional Item Response Models
ERIC Educational Resources Information Center
Zhang, Bo; Stone, Clement A.
2008-01-01
This research examines the utility of the s-x[superscript 2] statistic proposed by Orlando and Thissen (2000) in evaluating item fit for multidimensional item response models. Monte Carlo simulation was conducted to investigate both the Type I error and statistical power of this fit statistic in analyzing two kinds of multidimensional test…
ERIC Educational Resources Information Center
Toro, Maritsa
2011-01-01
The statistical assessment of dimensionality provides evidence of the underlying constructs measured by a survey or test instrument. This study focuses on educational measurement, specifically tests comprised of items described as multidimensional. That is, items that require examinee proficiency in multiple content areas and/or multiple cognitive…
Perceptual Salience and Children's Multidimensional Problem Solving
ERIC Educational Resources Information Center
Odom, Richard D.; Corbin, David W.
1973-01-01
Uni- and multidimensional processing of 6- to 9-year olds was studied using recall tasks in which an array of stimuli was reconstructed to match a model array. Results indicated that both age groups were able to solve multidimensional problems, but that solution rate was retarded by the unidimensional processing of highly salient dimensions.…
ERIC Educational Resources Information Center
Findler, Liora; Vilchinsky, Noa; Werner, Shirli
2007-01-01
This study presents the development of a new instrument, the "Multidimensional Attitudes Scale Toward Persons With Disabilities" (MAS). Based on the multidimensional approach, it posits that attitudes are composed of three dimensions: affect, cognition, and behavior. The scale was distributed to a sample of 132 people along with a…
The Extraction of One-Dimensional Flow Properties from Multi-Dimensional Data Sets
NASA Technical Reports Server (NTRS)
Baurle, Robert A.; Gaffney, Richard L., Jr.
2007-01-01
The engineering design and analysis of air-breathing propulsion systems relies heavily on zero- or one-dimensional properties (e.g. thrust, total pressure recovery, mixing and combustion efficiency, etc.) for figures of merit. The extraction of these parameters from experimental data sets and/or multi-dimensional computational data sets is therefore an important aspect of the design process. A variety of methods exist for extracting performance measures from multi-dimensional data sets. Some of the information contained in the multi-dimensional flow is inevitably lost when any one-dimensionalization technique is applied. Hence, the unique assumptions associated with a given approach may result in one-dimensional properties that are significantly different than those extracted using alternative approaches. The purpose of this effort is to examine some of the more popular methods used for the extraction of performance measures from multi-dimensional data sets, reveal the strengths and weaknesses of each approach, and highlight various numerical issues that result when mapping data from a multi-dimensional space to a space of one dimension.
The Art of Extracting One-Dimensional Flow Properties from Multi-Dimensional Data Sets
NASA Technical Reports Server (NTRS)
Baurle, R. A.; Gaffney, R. L.
2007-01-01
The engineering design and analysis of air-breathing propulsion systems relies heavily on zero- or one-dimensional properties (e:g: thrust, total pressure recovery, mixing and combustion efficiency, etc.) for figures of merit. The extraction of these parameters from experimental data sets and/or multi-dimensional computational data sets is therefore an important aspect of the design process. A variety of methods exist for extracting performance measures from multi-dimensional data sets. Some of the information contained in the multi-dimensional flow is inevitably lost when any one-dimensionalization technique is applied. Hence, the unique assumptions associated with a given approach may result in one-dimensional properties that are significantly different than those extracted using alternative approaches. The purpose of this effort is to examine some of the more popular methods used for the extraction of performance measures from multi-dimensional data sets, reveal the strengths and weaknesses of each approach, and highlight various numerical issues that result when mapping data from a multi-dimensional space to a space of one dimension.
The Cognitive Visualization System with the Dynamic Projection of Multidimensional Data
NASA Astrophysics Data System (ADS)
Gorohov, V.; Vitkovskiy, V.
2008-08-01
The phenomenon of cognitive machine drawing consists in the generation on the screen the special graphic representations, which create in the brain of human operator entertainment means. These means seem man by aesthetically attractive and, thus, they stimulate its descriptive imagination, closely related to the intuitive mechanisms of thinking. The essence of cognitive effect lies in the fact that man receives the moving projection as pseudo-three-dimensional object characterizing multidimensional means in the multidimensional space. After the thorough qualitative study of the visual aspects of multidimensional means with the aid of the enumerated algorithms appears the possibility, using algorithms of standard machine drawing to paint the interesting user separate objects or the groups of objects. Then it is possible to again return to the dynamic behavior of the rotation of means for the purpose of checking the intuitive ideas of user about the clusters and the connections in multidimensional data. Is possible the development of the methods of cognitive machine drawing in combination with other information technologies, first of all with the packets of digital processing of images and multidimensional statistical analysis.
Geology of Point Reyes National Seashore and vicinity, California: a digital database
Clark, Jospeh C.; Brabb, Earl E.
1997-01-01
This Open-File report is a digital geologic map database. This pamphlet serves to introduce and describe the digital data. There is no paper map included in the Open-File report. The report does include, however, a PostScript plot file containing an image of the geologic map sheet with explanation, as well as the accompanying text describing the geology of the area. For those interested in a paper plot of information contained in the database or in obtaining the PostScript plot files, please see the section entitled 'For Those Who Aren't Familiar With Digital Geologic Map Databases' below. This digital map database, compiled from previously published and unpublished data and new mapping by the authors, represents the general distribution of surficial deposits and rock units in Point Reyes and surrounding areas. Together with the accompanying text file (pr-geo.txt or pr-geo.ps), it provides current information on the stratigraphy and structural geology of the area covered. The database delineates map units that are identified by general age and lithology following the stratigraphic nomenclature of the U.S. Geological Survey. The scale of the source maps limits the spatial resolution (scale) of the database to 1:48,000 or smaller.
Kupferberg, Natalie; Jones Hartel, Lynda
2004-01-01
The purpose of this study is to assess the usefulness of five full-text drug databases as evaluated by medical librarians, pharmacy faculty, and pharmacy students at an academic health center. Study findings and recommendations are offered as guidance to librarians responsible for purchasing decisions. Four pharmacy students, four pharmacy faculty members, and four medical librarians answered ten drug information questions using the databases AHFS Drug Information (STAT!Ref); DRUGDEX (Micromedex); eFacts (Drug Facts and Comparisons); Lexi-Drugs Online (Lexi-Comp); and the PDR Electronic Library (Micromedex). Participants noted whether each database contained answers to the questions and evaluated each database on ease of navigation, screen readability, overall satisfaction, and product recommendation. While each study group found that DRUGDEX provided the most direct answers to the ten questions, faculty members gave Lexi-Drugs the highest overall rating. Students favored eFacts. The faculty and students found the PDR least useful. Librarians ranked DRUGDEX the highest and AHFS the lowest. The comments of pharmacy faculty and students show that these groups preferred concise, easy-to-use sources; librarians focused on the comprehensiveness, layout, and supporting references of the databases. This study demonstrates the importance of consulting with primary clientele before purchasing databases. Although there are many online drug databases to consider, present findings offer strong support for eFacts, Lexi-Drugs, and DRUGDEX.
Kupferberg, Natalie; Hartel, Lynda Jones
2004-01-01
Objectives: The purpose of this study is to assess the usefulness of five full-text drug databases as evaluated by medical librarians, pharmacy faculty, and pharmacy students at an academic health center. Study findings and recommendations are offered as guidance to librarians responsible for purchasing decisions. Methods: Four pharmacy students, four pharmacy faculty members, and four medical librarians answered ten drug information questions using the databases AHFS Drug Information (STAT!Ref); DRUGDEX (Micromedex); eFacts (Drug Facts and Comparisons); Lexi-Drugs Online (Lexi-Comp); and the PDR Electronic Library (Micromedex). Participants noted whether each database contained answers to the questions and evaluated each database on ease of navigation, screen readability, overall satisfaction, and product recommendation. Results: While each study group found that DRUGDEX provided the most direct answers to the ten questions, faculty members gave Lexi-Drugs the highest overall rating. Students favored eFacts. The faculty and students found the PDR least useful. Librarians ranked DRUGDEX the highest and AHFS the lowest. The comments of pharmacy faculty and students show that these groups preferred concise, easy-to-use sources; librarians focused on the comprehensiveness, layout, and supporting references of the databases. Conclusion: This study demonstrates the importance of consulting with primary clientele before purchasing databases. Although there are many online drug databases to consider, present findings offer strong support for eFacts, Lexi-Drugs, and DRUGDEX. PMID:14762464
Large-Scale Event Extraction from Literature with Multi-Level Gene Normalization
Wei, Chih-Hsuan; Hakala, Kai; Pyysalo, Sampo; Ananiadou, Sophia; Kao, Hung-Yu; Lu, Zhiyong; Salakoski, Tapio; Van de Peer, Yves; Ginter, Filip
2013-01-01
Text mining for the life sciences aims to aid database curation, knowledge summarization and information retrieval through the automated processing of biomedical texts. To provide comprehensive coverage and enable full integration with existing biomolecular database records, it is crucial that text mining tools scale up to millions of articles and that their analyses can be unambiguously linked to information recorded in resources such as UniProt, KEGG, BioGRID and NCBI databases. In this study, we investigate how fully automated text mining of complex biomolecular events can be augmented with a normalization strategy that identifies biological concepts in text, mapping them to identifiers at varying levels of granularity, ranging from canonicalized symbols to unique gene and proteins and broad gene families. To this end, we have combined two state-of-the-art text mining components, previously evaluated on two community-wide challenges, and have extended and improved upon these methods by exploiting their complementary nature. Using these systems, we perform normalization and event extraction to create a large-scale resource that is publicly available, unique in semantic scope, and covers all 21.9 million PubMed abstracts and 460 thousand PubMed Central open access full-text articles. This dataset contains 40 million biomolecular events involving 76 million gene/protein mentions, linked to 122 thousand distinct genes from 5032 species across the full taxonomic tree. Detailed evaluations and analyses reveal promising results for application of this data in database and pathway curation efforts. The main software components used in this study are released under an open-source license. Further, the resulting dataset is freely accessible through a novel API, providing programmatic and customized access (http://www.evexdb.org/api/v001/). Finally, to allow for large-scale bioinformatic analyses, the entire resource is available for bulk download from http://evexdb.org/download/, under the Creative Commons – Attribution – Share Alike (CC BY-SA) license. PMID:23613707
Liu, Yifeng; Liang, Yongjie; Wishart, David
2015-07-01
PolySearch2 (http://polysearch.ca) is an online text-mining system for identifying relationships between biomedical entities such as human diseases, genes, SNPs, proteins, drugs, metabolites, toxins, metabolic pathways, organs, tissues, subcellular organelles, positive health effects, negative health effects, drug actions, Gene Ontology terms, MeSH terms, ICD-10 medical codes, biological taxonomies and chemical taxonomies. PolySearch2 supports a generalized 'Given X, find all associated Ys' query, where X and Y can be selected from the aforementioned biomedical entities. An example query might be: 'Find all diseases associated with Bisphenol A'. To find its answers, PolySearch2 searches for associations against comprehensive collections of free-text collections, including local versions of MEDLINE abstracts, PubMed Central full-text articles, Wikipedia full-text articles and US Patent application abstracts. PolySearch2 also searches 14 widely used, text-rich biological databases such as UniProt, DrugBank and Human Metabolome Database to improve its accuracy and coverage. PolySearch2 maintains an extensive thesaurus of biological terms and exploits the latest search engine technology to rapidly retrieve relevant articles and databases records. PolySearch2 also generates, ranks and annotates associative candidates and present results with relevancy statistics and highlighted key sentences to facilitate user interpretation. © The Author(s) 2015. Published by Oxford University Press on behalf of Nucleic Acids Research.
Liu, Yifeng; Liang, Yongjie; Wishart, David
2015-01-01
PolySearch2 (http://polysearch.ca) is an online text-mining system for identifying relationships between biomedical entities such as human diseases, genes, SNPs, proteins, drugs, metabolites, toxins, metabolic pathways, organs, tissues, subcellular organelles, positive health effects, negative health effects, drug actions, Gene Ontology terms, MeSH terms, ICD-10 medical codes, biological taxonomies and chemical taxonomies. PolySearch2 supports a generalized ‘Given X, find all associated Ys’ query, where X and Y can be selected from the aforementioned biomedical entities. An example query might be: ‘Find all diseases associated with Bisphenol A’. To find its answers, PolySearch2 searches for associations against comprehensive collections of free-text collections, including local versions of MEDLINE abstracts, PubMed Central full-text articles, Wikipedia full-text articles and US Patent application abstracts. PolySearch2 also searches 14 widely used, text-rich biological databases such as UniProt, DrugBank and Human Metabolome Database to improve its accuracy and coverage. PolySearch2 maintains an extensive thesaurus of biological terms and exploits the latest search engine technology to rapidly retrieve relevant articles and databases records. PolySearch2 also generates, ranks and annotates associative candidates and present results with relevancy statistics and highlighted key sentences to facilitate user interpretation. PMID:25925572
NASA Astrophysics Data System (ADS)
Kim, Jong-Ahn; Bae, Eui Won; Kim, Soo Hyun; Kwak, Yoon Keun
2001-09-01
Precision actuators, such as pick-up actuators for HDDs or CD-ROMs, mostly show multidimensional motion. So, to evaluate them more completely, multidimensional measurement is required. Through structural variation and optimization of the design index, the performance of a measurement system can be improved to satisfy the requirement of this application, and so the resolution of each axis is higher than 0.1 μm for translation and 0.5 arcsec for rotation. Using this measurement system, the multidimensional motion and frequency transfer functions of a bimorph-type piezoelectric actuator are obtained.
Strandroos, Lisa; Antelius, Eleonor
2017-09-01
Previous research concerning bilingual people with a dementia disease has mainly focused on the importance of sharing a spoken language with caregivers. While acknowledging this, this article addresses the multidimensional character of communication and interaction. As using spoken language is made difficult as a consequence of the dementia disease, this multidimensionality becomes particularly important. The article is based on a qualitative analysis of ethnographic fieldwork at a dementia care facility. It presents ethnographic examples of different communicative forms, with particular focus on bilingual interactions. Interaction is understood as a collective and collaborative activity. The text finds that a shared spoken language is advantageous, but is not the only source of, nor a guarantee for, creating common ground and understanding. Communicative resources other than spoken language are for example body language, embodiment, artefacts and time. Furthermore, forms of communication are not static but develop, change and are created over time. Ability to communicate is thus not something that one has or has not, but is situationally and collaboratively created. To facilitate this, time and familiarity are central resources, and the results indicate the importance of continuity in interpersonal relations.
Overview of Historical Earthquake Document Database in Japan and Future Development
NASA Astrophysics Data System (ADS)
Nishiyama, A.; Satake, K.
2014-12-01
In Japan, damage and disasters from historical large earthquakes have been documented and preserved. Compilation of historical earthquake documents started in the early 20th century and 33 volumes of historical document source books (about 27,000 pages) have been published. However, these source books are not effectively utilized for researchers due to a contamination of low-reliability historical records and a difficulty for keyword searching by characters and dates. To overcome these problems and to promote historical earthquake studies in Japan, construction of text database started in the 21 century. As for historical earthquakes from the beginning of the 7th century to the early 17th century, "Online Database of Historical Documents in Japanese Earthquakes and Eruptions in the Ancient and Medieval Ages" (Ishibashi, 2009) has been already constructed. They investigated the source books or original texts of historical literature, emended the descriptions, and assigned the reliability of each historical document on the basis of written age. Another database compiled the historical documents for seven damaging earthquakes occurred along the Sea of Japan coast in Honshu, central Japan in the Edo period (from the beginning of the 17th century to the middle of the 19th century) and constructed text database and seismic intensity data base. These are now publicized on the web (written only in Japanese). However, only about 9 % of the earthquake source books have been digitized so far. Therefore, we plan to digitize all of the remaining historical documents by the research-program which started in 2014. The specification of the data base will be similar for previous ones. We also plan to combine this database with liquefaction traces database, which will be constructed by other research program, by adding the location information described in historical documents. Constructed database would be utilized to estimate the distributions of seismic intensities and tsunami heights.
Graphics-based intelligent search and abstracting using Data Modeling
NASA Astrophysics Data System (ADS)
Jaenisch, Holger M.; Handley, James W.; Case, Carl T.; Songy, Claude G.
2002-11-01
This paper presents an autonomous text and context-mining algorithm that converts text documents into point clouds for visual search cues. This algorithm is applied to the task of data-mining a scriptural database comprised of the Old and New Testaments from the Bible and the Book of Mormon, Doctrine and Covenants, and the Pearl of Great Price. Results are generated which graphically show the scripture that represents the average concept of the database and the mining of the documents down to the verse level.
Multidimensional poverty and child survival in India.
Mohanty, Sanjay K
2011-01-01
Though the concept of multidimensional poverty has been acknowledged cutting across the disciplines (among economists, public health professionals, development thinkers, social scientists, policy makers and international organizations) and included in the development agenda, its measurement and application are still limited. OBJECTIVES AND METHODOLOGY: Using unit data from the National Family and Health Survey 3, India, this paper measures poverty in multidimensional space and examine the linkages of multidimensional poverty with child survival. The multidimensional poverty is measured in the dimension of knowledge, health and wealth and the child survival is measured with respect to infant mortality and under-five mortality. Descriptive statistics, principal component analyses and the life table methods are used in the analyses. The estimates of multidimensional poverty are robust and the inter-state differentials are large. While infant mortality rate and under-five mortality rate are disproportionately higher among the abject poor compared to the non-poor, there are no significant differences in child survival among educationally, economically and health poor at the national level. State pattern in child survival among the education, economical and health poor are mixed. Use of multidimensional poverty measures help to identify abject poor who are unlikely to come out of poverty trap. The child survival is significantly lower among abject poor compared to moderate poor and non-poor. We urge to popularize the concept of multiple deprivations in research and program so as to reduce poverty and inequality in the population.
Naeem, S.; Prager, Case; Weeks, Brian; Varga, Alex; Flynn, Dan F. B.; Griffin, Kevin; Muscarella, Robert; Palmer, Matthew; Wood, Stephen; Schuster, William
2016-01-01
Biodiversity is inherently multidimensional, encompassing taxonomic, functional, phylogenetic, genetic, landscape and many other elements of variability of life on the Earth. However, this fundamental principle of multidimensionality is rarely applied in research aimed at understanding biodiversity's value to ecosystem functions and the services they provide. This oversight means that our current understanding of the ecological and environmental consequences of biodiversity loss is limited primarily to what unidimensional studies have revealed. To address this issue, we review the literature, develop a conceptual framework for multidimensional biodiversity research based on this review and provide a case study to explore the framework. Our case study specifically examines how herbivory by whitetail deer (Odocoileus virginianus) alters the multidimensional influence of biodiversity on understory plant cover at Black Rock Forest, New York. Using three biodiversity dimensions (taxonomic, functional and phylogenetic diversity) to explore our framework, we found that herbivory alters biodiversity's multidimensional influence on plant cover; an effect not observable through a unidimensional approach. Although our review, framework and case study illustrate the advantages of multidimensional over unidimensional approaches, they also illustrate the statistical and empirical challenges such work entails. Meeting these challenges, however, where data and resources permit, will be important if we are to better understand and manage the consequences we face as biodiversity continues to decline in the foreseeable future. PMID:27928041
Naeem, S; Prager, Case; Weeks, Brian; Varga, Alex; Flynn, Dan F B; Griffin, Kevin; Muscarella, Robert; Palmer, Matthew; Wood, Stephen; Schuster, William
2016-12-14
Biodiversity is inherently multidimensional, encompassing taxonomic, functional, phylogenetic, genetic, landscape and many other elements of variability of life on the Earth. However, this fundamental principle of multidimensionality is rarely applied in research aimed at understanding biodiversity's value to ecosystem functions and the services they provide. This oversight means that our current understanding of the ecological and environmental consequences of biodiversity loss is limited primarily to what unidimensional studies have revealed. To address this issue, we review the literature, develop a conceptual framework for multidimensional biodiversity research based on this review and provide a case study to explore the framework. Our case study specifically examines how herbivory by whitetail deer (Odocoileus virginianus) alters the multidimensional influence of biodiversity on understory plant cover at Black Rock Forest, New York. Using three biodiversity dimensions (taxonomic, functional and phylogenetic diversity) to explore our framework, we found that herbivory alters biodiversity's multidimensional influence on plant cover; an effect not observable through a unidimensional approach. Although our review, framework and case study illustrate the advantages of multidimensional over unidimensional approaches, they also illustrate the statistical and empirical challenges such work entails. Meeting these challenges, however, where data and resources permit, will be important if we are to better understand and manage the consequences we face as biodiversity continues to decline in the foreseeable future. © 2016 The Authors.
Pinilla-Roncancio, Mónica
2017-12-30
Disability and poverty are interconnected and although this relationship has been recognised, there is a lack of empirical evidence to support any possible causal relationship in this topic, particularly in the context of Latin America (LA). This study tests the hypothesis "Disability increases the risk of multidimensional poverty of people living with disabilities and their families". Using national census data from Brazil, Chile, Colombia, Costa Rica and Mexico, the Global Multidimensional Poverty Index (Global MPI) was calculated with the aim of measuring and comparing the levels of multidimensional poverty of people living in households with and without disabled members in the five countries. We found that in the five countries people with disabilities and their families had higher incidence, intensity and levels of multidimensional poverty compared with people living in other households. Their levels of deprivation were also higher for all the indicators included in the Global MPI and the contribution of this group to the national MPI was higher than their share of the population, thus people with disabilities and their families are overrepresented in those living in multidimensional poverty. People with disabilities and their families are in worse conditions than poor households without disabled members and social policies should aim to reduce their high levels of multidimensional poverty and deprivation. Copyright © 2017 Elsevier Inc. All rights reserved.
Massively Scalable Near Duplicate Detection in Streams of Documents using MDSH
DOE Office of Scientific and Technical Information (OSTI.GOV)
Bogen, Paul Logasa; Symons, Christopher T; McKenzie, Amber T
2013-01-01
In a world where large-scale text collections are not only becoming ubiquitous but also are growing at increasing rates, near duplicate documents are becoming a growing concern that has the potential to hinder many different information filtering tasks. While others have tried to address this problem, prior techniques have only been used on limited collection sizes and static cases. We will briefly describe the problem in the context of Open Source Intelligence (OSINT) along with our additional constraints for performance. In this work we propose two variations on Multi-dimensional Spectral Hash (MDSH) tailored for working on extremely large, growing setsmore » of text documents. We analyze the memory and runtime characteristics of our techniques and provide an informal analysis of the quality of the near-duplicate clusters produced by our techniques.« less
ERIC Educational Resources Information Center
Yao, Lihua; Schwarz, Richard D.
2006-01-01
Multidimensional item response theory (IRT) models have been proposed for better understanding the dimensional structure of data or to define diagnostic profiles of student learning. A compensatory multidimensional two-parameter partial credit model (M-2PPC) for constructed-response items is presented that is a generalization of those proposed to…
ERIC Educational Resources Information Center
Arens, A. Katrin; Jansen, Malte
2016-01-01
Academic self-concept has been conceptualized as a multidimensional and hierarchical construct. Previous research has mostly focused on its multidimensionality, distinguishing between verbal and mathematical self-concept domains, and only a few studies have examined the factorial structure within specific self-concept domains. The present study…
Assessing Construct Validity Using Multidimensional Item Response Theory.
ERIC Educational Resources Information Center
Ackerman, Terry A.
The concept of a user-specified validity sector is discussed. The idea of the validity sector combines the work of M. D. Reckase (1986) and R. Shealy and W. Stout (1991). Reckase developed a methodology to represent an item in a multidimensional latent space as a vector. Item vectors are computed using multidimensional item response theory item…
ERIC Educational Resources Information Center
Martin, Andrew J.
2009-01-01
This investigation conducts measurement and evaluation of a multidimensional model of workplace motivation and engagement from a construct validation perspective. Two studies were conducted, one using the multi-item multidimensional Motivation and Engagement Scale-Work (N = 637 school personnel) and one using a parallel short form (N = 574 school…
Effects of Multidimensional Concept Maps on Fourth Graders' Learning in Web-Based Computer Course
ERIC Educational Resources Information Center
Huang, Hwa-Shan; Chiou, Chei-Chang; Chiang, Heien-Kun; Lai, Sung-Hsi; Huang, Chiun-Yen; Chou, Yin-Yu
2012-01-01
This study explores the effect of multidimensional concept mapping instruction on students' learning performance in a web-based computer course. The subjects consisted of 103 fourth graders from an elementary school in central Taiwan. They were divided into three groups: multidimensional concept map (MCM) instruction group, Novak concept map (NCM)…
ERIC Educational Resources Information Center
Kim, Se-Kang
2010-01-01
The aim of the current study is to validate the invariance of major profile patterns derived from multidimensional scaling (MDS) by bootstrapping. Profile Analysis via Multidimensional Scaling (PAMS) was employed to obtain profiles and bootstrapping was used to construct the sampling distributions of the profile coordinates and the empirical…
ERIC Educational Resources Information Center
Hawkins, Colleen C.; Watt, Helen M. G.; Sinclair, Kenneth E.
2006-01-01
The psychometric properties of the Frost, Marten, Lahart, and Rosenblate Multidimensional Perfectionism Scale (1990) are investigated to determine its usefulness as a measurement of perfectionism with Australian secondary school girls and to find empirical support for the existence of both healthy and unhealthy types of perfectionist students.…
The Flip Sides of Full-Text: Superindex and the Harvard Business Review/Online.
ERIC Educational Resources Information Center
Dadlez, Eva M.
1984-01-01
This article illustrates similarities between two different types of full-text databases--Superindex, Harvard Business Review/Online--and uses them as arena to demonstrate search and display applications of full-text. The selection of logical operators, full-text search strategies, and keywords and Bibliographic Retrieval Service's Occurrence…
Panepinto, Julie A; Torres, Sylvia; Bendo, Cristiane B; McCavit, Timothy L; Dinu, Bogdan; Sherman-Bien, Sandra; Bemrich-Stolz, Christy; Varni, James W
2014-01-01
Sickle cell disease (SCD) is an inherited blood disorder characterized by a chronic hemolytic anemia that can contribute to fatigue and global cognitive impairment in patients. The study objective was to report on the feasibility, reliability, and validity of the PedsQL™ Multidimensional Fatigue Scale in SCD for pediatric patient self-report ages 5-18 years and parent proxy-report for ages 2-18 years. This was a cross-sectional multi-site study whereby 240 pediatric patients with SCD and 303 parents completed the 18-item PedsQL™ Multidimensional Fatigue Scale. Participants also completed the PedsQL™ 4.0 Generic Core Scales. The PedsQL™ Multidimensional Fatigue Scale evidenced excellent feasibility, excellent reliability for the Total Scale Scores (patient self-report α = 0.90; parent proxy-report α = 0.95), and acceptable reliability for the three individual scales (patient self-report α = 0.77-0.84; parent proxy-report α = 0.90-0.97). Intercorrelations of the PedsQL™ Multidimensional Fatigue Scale with the PedsQL™ Generic Core Scales were predominantly in the large (≥0.50) range, supporting construct validity. PedsQL™ Multidimensional Fatigue Scale Scores were significantly worse with large effects sizes (≥0.80) for patients with SCD than for a comparison sample of healthy children, supporting known-groups discriminant validity. Confirmatory factor analysis demonstrated an acceptable to excellent model fit in SCD. The PedsQL™ Multidimensional Fatigue Scale demonstrated acceptable to excellent measurement properties in SCD. The results demonstrate the relative severity of fatigue symptoms in pediatric patients with SCD, indicating the potential clinical utility of multidimensional assessment of fatigue in patients with SCD in clinical research and practice. © 2013 Wiley Periodicals, Inc.
PedsQL™ Multidimensional Fatigue Scale in Sickle Cell Disease: Feasibility, Reliability and Validity
Panepinto, Julie A.; Torres, Sylvia; Bendo, Cristiane B.; McCavit, Timothy L.; Dinu, Bogdan; Sherman-Bien, Sandra; Bemrich-Stolz, Christy; Varni, James W.
2013-01-01
Background Sickle cell disease (SCD) is an inherited blood disorder characterized by a chronic hemolytic anemia that can contribute to fatigue and global cognitive impairment in patients. The study objective was to report on the feasibility, reliability, and validity of the PedsQL™ Multidimensional Fatigue Scale in SCD for pediatric patient self-report ages 5–18 years and parent proxy-report for ages 2–18 years. Procedure This was a cross-sectional multi-site study whereby 240 pediatric patients with SCD and 303 parents completed the 18-item PedsQL™ Multidimensional Fatigue Scale. Participants also completed the PedsQL™ 4.0 Generic Core Scales. Results The PedsQL™ Multidimensional Fatigue Scale evidenced excellent feasibility, excellent reliability for the Total Scale Scores (patient self-report α = 0.90; parent proxy-report α = 0.95), and acceptable reliability for the three individual scales (patient self-report α = 0.77–0.84; parent proxy-report α = 0.90–0.97). Intercorrelations of the PedsQL™ Multidimensional Fatigue Scale with the PedsQL™ Generic Core Scales were predominantly in the large (≥ 0.50) range, supporting construct validity. PedsQL™ Multidimensional Fatigue Scale Scores were significantly worse with large effects sizes (≥0.80) for patients with SCD than for a comparison sample of healthy children, supporting known-groups discriminant validity. Confirmatory factor analysis demonstrated an acceptable to excellent model fit in SCD. Conclusions The PedsQL™ Multidimensional Fatigue Scale demonstrated acceptable to excellent measurement properties in SCD. The results demonstrate the relative severity of fatigue symptoms in pediatric patients with SCD, indicating the potential clinical utility of multidimensional assessment of fatigue in patients with SCD in clinical research and practice. PMID:24038960
Development of multi-dimensional body image scale for malaysian female adolescents
Taib, Mohd Nasir Mohd; Shariff, Zalilah Mohd; Khor, Geok Lin
2008-01-01
The present study was conducted to develop a Multi-dimensional Body Image Scale for Malaysian female adolescents. Data were collected among 328 female adolescents from a secondary school in Kuantan district, state of Pahang, Malaysia by using a self-administered questionnaire and anthropometric measurements. The self-administered questionnaire comprised multiple measures of body image, Eating Attitude Test (EAT-26; Garner & Garfinkel, 1979) and Rosenberg Self-esteem Inventory (Rosenberg, 1965). The 152 items from selected multiple measures of body image were examined through factor analysis and for internal consistency. Correlations between Multi-dimensional Body Image Scale and body mass index (BMI), risk of eating disorders and self-esteem were assessed for construct validity. A seven factor model of a 62-item Multi-dimensional Body Image Scale for Malaysian female adolescents with construct validity and good internal consistency was developed. The scale encompasses 1) preoccupation with thinness and dieting behavior, 2) appearance and body satisfaction, 3) body importance, 4) muscle increasing behavior, 5) extreme dieting behavior, 6) appearance importance, and 7) perception of size and shape dimensions. Besides, a multidimensional body image composite score was proposed to screen negative body image risk in female adolescents. The result found body image was correlated with BMI, risk of eating disorders and self-esteem in female adolescents. In short, the present study supports a multi-dimensional concept for body image and provides a new insight into its multi-dimensionality in Malaysian female adolescents with preliminary validity and reliability of the scale. The Multi-dimensional Body Image Scale can be used to identify female adolescents who are potentially at risk of developing body image disturbance through future intervention programs. PMID:20126371
Development of multi-dimensional body image scale for malaysian female adolescents.
Chin, Yit Siew; Taib, Mohd Nasir Mohd; Shariff, Zalilah Mohd; Khor, Geok Lin
2008-01-01
The present study was conducted to develop a Multi-dimensional Body Image Scale for Malaysian female adolescents. Data were collected among 328 female adolescents from a secondary school in Kuantan district, state of Pahang, Malaysia by using a self-administered questionnaire and anthropometric measurements. The self-administered questionnaire comprised multiple measures of body image, Eating Attitude Test (EAT-26; Garner & Garfinkel, 1979) and Rosenberg Self-esteem Inventory (Rosenberg, 1965). The 152 items from selected multiple measures of body image were examined through factor analysis and for internal consistency. Correlations between Multi-dimensional Body Image Scale and body mass index (BMI), risk of eating disorders and self-esteem were assessed for construct validity. A seven factor model of a 62-item Multi-dimensional Body Image Scale for Malaysian female adolescents with construct validity and good internal consistency was developed. The scale encompasses 1) preoccupation with thinness and dieting behavior, 2) appearance and body satisfaction, 3) body importance, 4) muscle increasing behavior, 5) extreme dieting behavior, 6) appearance importance, and 7) perception of size and shape dimensions. Besides, a multidimensional body image composite score was proposed to screen negative body image risk in female adolescents. The result found body image was correlated with BMI, risk of eating disorders and self-esteem in female adolescents. In short, the present study supports a multi-dimensional concept for body image and provides a new insight into its multi-dimensionality in Malaysian female adolescents with preliminary validity and reliability of the scale. The Multi-dimensional Body Image Scale can be used to identify female adolescents who are potentially at risk of developing body image disturbance through future intervention programs.
Fisher, Jolene H; Al-Hejaili, Faris; Kandel, Sonja; Hirji, Alim; Shapera, Shane; Mura, Marco
2017-04-01
The heterogeneous progression of idiopathic pulmonary fibrosis (IPF) makes prognostication difficult and contributes to high mortality on the waitlist for lung transplantation (LTx). Multi-dimensional scores (Composite Physiologic index [CPI], [Gender-Age-Physiology [GAP]; RIsk Stratification scorE [RISE]) demonstrated enhanced predictive power towards outcome in IPF. The lung allocation score (LAS) is a multi-dimensional tool commonly used to stratify patients assessed for LTx. We sought to investigate whether IPF-specific multi-dimensional scores predict mortality in patients with IPF assessed for LTx. The study included 302 patients with IPF who underwent a LTx assessment (2003-2014). Multi-dimensional scores were calculated. The primary outcome was 12-month mortality after assessment. LTx was considered as competing event in all analyses. At the end of the observation period, there were 134 transplants, 63 deaths, and 105 patients were alive without LTx. Multi-dimensional scores predicted mortality with accuracy similar to LAS, and superior to that of individual variables: area under the curve (AUC) for LAS was 0.78 (sensitivity 71%, specificity 86%); CPI 0.75 (sensitivity 67%, specificity 82%); GAP 0.67 (sensitivity 59%, specificity 74%); RISE 0.78 (sensitivity 71%, specificity 84%). A separate analysis conducted only in patients actively listed for LTx (n = 247; 50 deaths) yielded similar results. In patients with IPF assessed for LTx as well as in those actually listed, multi-dimensional scores predict mortality better than individual variables, and with accuracy similar to the LAS. If validated, multi-dimensional scores may serve as inexpensive tools to guide decisions on the timing of referral and listing for LTx. Copyright © 2017 Elsevier Ltd. All rights reserved.
Numeric invariants from multidimensional persistence
Skryzalin, Jacek; Carlsson, Gunnar
2017-05-19
Topological data analysis is the study of data using techniques from algebraic topology. Often, one begins with a finite set of points representing data and a “filter” function which assigns a real number to each datum. Using both the data and the filter function, one can construct a filtered complex for further analysis. For example, applying the homology functor to the filtered complex produces an algebraic object known as a “one-dimensional persistence module”, which can often be interpreted as a finite set of intervals representing various geometric features in the data. If one runs the above process incorporating multiple filtermore » functions simultaneously, one instead obtains a multidimensional persistence module. Unfortunately, these are much more difficult to interpret. In this article, we analyze the space of multidimensional persistence modules from the perspective of algebraic geometry. First we build a moduli space of a certain subclass of easily analyzed multidimensional persistence modules, which we construct specifically to capture much of the information which can be gained by using multidimensional persistence instead of one-dimensional persistence. Fruthermore, we argue that the global sections of this space provide interesting numeric invariants when evaluated against our subclass of multidimensional persistence modules. Finally, we extend these global sections to the space of all multidimensional persistence modules and discuss how the resulting numeric invariants might be used to study data. This paper extends the results of Adcock et al. (Homol Homotopy Appl 18(1), 381–402, 2016) by constructing numeric invariants from the computation of a multidimensional persistence module as given by Carlsson et al. (J Comput Geom 1(1), 72–100, 2010).« less
Mewes, Janne C.; IJzerman, Maarten J.; van Harten, Wim H.
2012-01-01
Introduction. Many cancer survivors suffer from a combination of disease- and treatment-related morbidities and complaints after primary treatment. There is a growing evidence base for the effectiveness of monodimensional rehabilitation interventions; in practice, however, patients often participate in multidimensional programs. This study systematically reviews evidence regarding effectiveness of multidimensional rehabilitation programs for cancer survivors and cost-effectiveness of cancer rehabilitation in general. Methods. The published literature was systematically reviewed. Data were extracted using standardized forms and were summarized narratively. Results. Sixteen effectiveness and six cost-effectiveness studies were included. Multidimensional rehabilitation programs were found to be effective, but not more effective than monodimensional interventions, and not on all outcome measures. Effect sizes for quality of life were in the range of −0.12 (95% confidence interval [CI], −0.45–0.20) to 0.98 (95% CI, 0.69–1.29). Incremental cost-effectiveness ratios ranged from −€16,976, indicating cost savings, to €11,057 per quality-adjusted life year. Conclusions. The evidence for multidimensional interventions and the economic impact of rehabilitation studies is scarce and dominated by breast cancer studies. Studies published so far report statistically significant benefits for multidimensional interventions over usual care, most notably for the outcomes fatigue and physical functioning. An additional benefit of multidimensional over monodimensional rehabilitation was not found, but this was also sparsely reported on. Available economic evaluations assessed very different rehabilitation interventions. Yet, despite low comparability, all showed favorable cost-effectiveness ratios. Future studies should focus their designs on the comparative effectiveness and cost-effectiveness of multidimensional programs. PMID:22982580
NASA Astrophysics Data System (ADS)
Kiessling, Wolfgang; Kumar Pandey, Dhirendra; Schemm-Gregory, Mena; Mewis, Heike; Aberhan, Martin
2011-02-01
We present the first modern description of corals, brachiopods and bivalves from the Antalo Limestone in the Mekele Outlier of northern Ethiopia. This fauna is largely of Oxfordian age and lived in shallow subtidal environments and in small patch reefs. In combining our new data with fossil occurrence data from the Paleobiology Database, we conducted multidimensional scaling analyses to assess biogeographic patterns and the delineation of the Ethiopian Province for the Callovian to Kimmeridgian stages. Results suggest that an Ethiopian Province is indeed evident for our focal groups, but this is more confined than traditionally assumed. The so defined Ethiopian Province includes Tunisia, the Levant, Arabia and much of East Africa, but excludes Tanzania and India. The special status of India and Tanzania is perhaps due to latitudinal gradients in faunal composition.
Standard model of knowledge representation
NASA Astrophysics Data System (ADS)
Yin, Wensheng
2016-09-01
Knowledge representation is the core of artificial intelligence research. Knowledge representation methods include predicate logic, semantic network, computer programming language, database, mathematical model, graphics language, natural language, etc. To establish the intrinsic link between various knowledge representation methods, a unified knowledge representation model is necessary. According to ontology, system theory, and control theory, a standard model of knowledge representation that reflects the change of the objective world is proposed. The model is composed of input, processing, and output. This knowledge representation method is not a contradiction to the traditional knowledge representation method. It can express knowledge in terms of multivariate and multidimensional. It can also express process knowledge, and at the same time, it has a strong ability to solve problems. In addition, the standard model of knowledge representation provides a way to solve problems of non-precision and inconsistent knowledge.
The Multidimensional Assessment of Interoceptive Awareness (MAIA)
Mehling, Wolf E.; Price, Cynthia; Daubenmier, Jennifer J.; Acree, Mike; Bartmess, Elizabeth; Stewart, Anita
2012-01-01
This paper describes the development of a multidimensional self-report measure of interoceptive body awareness. The systematic mixed-methods process involved reviewing the current literature, specifying a multidimensional conceptual framework, evaluating prior instruments, developing items, and analyzing focus group responses to scale items by instructors and patients of body awareness-enhancing therapies. Following refinement by cognitive testing, items were field-tested in students and instructors of mind-body approaches. Final item selection was achieved by submitting the field test data to an iterative process using multiple validation methods, including exploratory cluster and confirmatory factor analyses, comparison between known groups, and correlations with established measures of related constructs. The resulting 32-item multidimensional instrument assesses eight concepts. The psychometric properties of these final scales suggest that the Multidimensional Assessment of Interoceptive Awareness (MAIA) may serve as a starting point for research and further collaborative refinement. PMID:23133619
Benefits of Multidimensional Measures of Child Well Being in China
Gatenio Gabel, Shirley; Zhang, Yiwei
2017-01-01
In recent decades, measures of child well-being have evolved from single dimension to multidimensional measures. Multi-dimensional measures deepen and broaden our understanding of child well-being and inform us of areas of neglect. Child well-being in China today is measured through proxy measures of household need. This paper discusses the evolution of child well-being measures more generally, explores the benefits of positive indicators and multiple dimensions in formulating policy, and then reviews efforts to date by the Chinese government, researchers, and non-governmental and intergovernmental organizations to develop comprehensive multidimensional measures of child well-being in China. The domains and their potential interactions, as well as data sources and availability, are presented. The authors believe that child well-being in China would benefit from the development of a multidimensional index and that there is sufficient data to develop such an index. PMID:29113121
Positivity-preserving numerical schemes for multidimensional advection
NASA Technical Reports Server (NTRS)
Leonard, B. P.; Macvean, M. K.; Lock, A. P.
1993-01-01
This report describes the construction of an explicit, single time-step, conservative, finite-volume method for multidimensional advective flow, based on a uniformly third-order polynomial interpolation algorithm (UTOPIA). Particular attention is paid to the problem of flow-to-grid angle-dependent, anisotropic distortion typical of one-dimensional schemes used component-wise. The third-order multidimensional scheme automatically includes certain cross-difference terms that guarantee good isotropy (and stability). However, above first-order, polynomial-based advection schemes do not preserve positivity (the multidimensional analogue of monotonicity). For this reason, a multidimensional generalization of the first author's universal flux-limiter is sought. This is a very challenging problem. A simple flux-limiter can be found; but this introduces strong anisotropic distortion. A more sophisticated technique, limiting part of the flux and then restoring the isotropy-maintaining cross-terms afterwards, gives more satisfactory results. Test cases are confined to two dimensions; three-dimensional extensions are briefly discussed.
Benefits of Multidimensional Measures of Child Well Being in China.
Gatenio Gabel, Shirley; Zhang, Yiwei
2017-11-06
In recent decades, measures of child well-being have evolved from single dimension to multidimensional measures. Multi-dimensional measures deepen and broaden our understanding of child well-being and inform us of areas of neglect. Child well-being in China today is measured through proxy measures of household need. This paper discusses the evolution of child well-being measures more generally, explores the benefits of positive indicators and multiple dimensions in formulating policy, and then reviews efforts to date by the Chinese government, researchers, and non-governmental and intergovernmental organizations to develop comprehensive multidimensional measures of child well-being in China. The domains and their potential interactions, as well as data sources and availability, are presented. The authors believe that child well-being in China would benefit from the development of a multidimensional index and that there is sufficient data to develop such an index.
Testing the multidimensionality of the inventory of school motivation in a Dutch student sample.
Korpershoek, Hanke; Xu, Kun; Mok, Magdalena Mo Ching; McInerney, Dennis M; van der Werf, Greetje
2015-01-01
A factor analytic and a Rasch measurement approach were applied to evaluate the multidimensional nature of the school motivation construct among more than 7,000 Dutch secondary school students. The Inventory of School Motivation (McInerney and Ali, 2006) was used, which intends to measure four motivation dimensions (mastery, performance, social, and extrinsic motivation), each comprising of two first-order factors. One unidimensional model and three multidimensional models (4-factor, 8-factor, higher order) were fit to the data. Results of both approaches showed that the multidimensional models validly represented the school motivation among Dutch secondary school pupils, whereas model fit of the unidimensional model was poor. The differences in model fit between the three multidimensional models were small, although a different model was favoured by the two approaches. The need for improvement of some of the items and the need to increase measurement precision of several first-order factors are discussed.
LMSD: LIPID MAPS structure database
Sud, Manish; Fahy, Eoin; Cotter, Dawn; Brown, Alex; Dennis, Edward A.; Glass, Christopher K.; Merrill, Alfred H.; Murphy, Robert C.; Raetz, Christian R. H.; Russell, David W.; Subramaniam, Shankar
2007-01-01
The LIPID MAPS Structure Database (LMSD) is a relational database encompassing structures and annotations of biologically relevant lipids. Structures of lipids in the database come from four sources: (i) LIPID MAPS Consortium's core laboratories and partners; (ii) lipids identified by LIPID MAPS experiments; (iii) computationally generated structures for appropriate lipid classes; (iv) biologically relevant lipids manually curated from LIPID BANK, LIPIDAT and other public sources. All the lipid structures in LMSD are drawn in a consistent fashion. In addition to a classification-based retrieval of lipids, users can search LMSD using either text-based or structure-based search options. The text-based search implementation supports data retrieval by any combination of these data fields: LIPID MAPS ID, systematic or common name, mass, formula, category, main class, and subclass data fields. The structure-based search, in conjunction with optional data fields, provides the capability to perform a substructure search or exact match for the structure drawn by the user. Search results, in addition to structure and annotations, also include relevant links to external databases. The LMSD is publicly available at PMID:17098933
Graph Learning in Knowledge Bases
DOE Office of Scientific and Technical Information (OSTI.GOV)
Goldberg, Sean; Wang, Daisy Zhe
The amount of text data has been growing exponentially in recent years, giving rise to automatic information extraction methods that store text annotations in a database. The current state-of-theart structured prediction methods, however, are likely to contain errors and it’s important to be able to manage the overall uncertainty of the database. On the other hand, the advent of crowdsourcing has enabled humans to aid machine algorithms at scale. As part of this project we introduced pi-CASTLE , a system that optimizes and integrates human and machine computing as applied to a complex structured prediction problem involving conditional random fieldsmore » (CRFs). We proposed strategies grounded in information theory to select a token subset, formulate questions for the crowd to label, and integrate these labelings back into the database using a method of constrained inference. On both a text segmentation task over academic citations and a named entity recognition task over tweets we showed an order of magnitude improvement in accuracy gain over baseline methods.« less
Online Patent Searching: The Realities.
ERIC Educational Resources Information Center
Kaback, Stuart M.
1983-01-01
Considers patent subject searching capabilities of major online databases, noting patent claims, "deep-indexed" files, test searches, retrieval of related references, multi-database searching, improvements needed in indexing of chemical structures, full text searching, improvements needed in handling numerical data, and augmenting a…
Online Databases. Ingenta Grows in the U.S. Market.
ERIC Educational Resources Information Center
Tenopir, Carol
2002-01-01
Focuses on the growth of the United Kingdom company Ingenta in the United States. Highlights include a new partnership with Gale Group related to their InfoTrac databases; indexing and full text coverage; standards; and other Ingenta acquisitions. (LRW)
ERIC Educational Resources Information Center
Yao, Lihua
2014-01-01
The intent of this research was to find an item selection procedure in the multidimensional computer adaptive testing (CAT) framework that yielded higher precision for both the domain and composite abilities, had a higher usage of the item pool, and controlled the exposure rate. Five multidimensional CAT item selection procedures (minimum angle;…
ERIC Educational Resources Information Center
Yao, Shuqiao; Zou, Tao; Zhu, Xiongzhao; Abela, John R. Z.; Auerbach, Randy P.; Tong, Xi
2007-01-01
The objective of the current study was to develop a Chinese translation of the Multidimensional Anxiety Scale for Children (MASC) [March (1997) Multidimensional anxiety scale for children: Technical manual, Multi health systems, Toronto, ON], and to evaluate its reliability and validity. The original version of the MASC was translated into Chinese…
ERIC Educational Resources Information Center
Karatas, Zeynep; Tagay, Ozlem
2012-01-01
The purpose of this study is to determine whether there is a relationship between self-esteem, locus of control and multidimensional perfectionism, and the extent to which the variables of self-esteem, locus of control and multidimensional perfectionism contribute to the prediction of subjective well-being. The study was carried out with 318 final…
ERIC Educational Resources Information Center
Andreev, Valentin I.
2014-01-01
The main aim of this research is to disclose the essence of students' multi-dimensional thinking, also to reveal the rating of factors which stimulate the raising of effectiveness of self-development of students' multi-dimensional thinking in terms of subject-oriented teaching. Subject-oriented learning is characterized as a type of learning where…
ERIC Educational Resources Information Center
Li, Ying; Jiao, Hong; Lissitz, Robert W.
2012-01-01
This study investigated the application of multidimensional item response theory (IRT) models to validate test structure and dimensionality. Multiple content areas or domains within a single subject often exist in large-scale achievement tests. Such areas or domains may cause multidimensionality or local item dependence, which both violate the…
A Multidimensional Ideal Point Item Response Theory Model for Binary Data
ERIC Educational Resources Information Center
Maydeu-Olivares, Albert; Hernandez, Adolfo; McDonald, Roderick P.
2006-01-01
We introduce a multidimensional item response theory (IRT) model for binary data based on a proximity response mechanism. Under the model, a respondent at the mode of the item response function (IRF) endorses the item with probability one. The mode of the IRF is the ideal point, or in the multidimensional case, an ideal hyperplane. The model…
ERIC Educational Resources Information Center
Rinn, Anne N.; Jamieson, Kelly M.; Gross, Candace M.; McQueen, Kand S.
2009-01-01
This study examines the effects of social comparison, gender, and grade level on gifted adolescents' multidimensional self-concept. Participants include 248 gifted adolescents who had completed the sixth through tenth grade during the previous academic year. Multidimensional self-concept was measured using the Self Description Questionnaire II…
Some applications of the multi-dimensional fractional order for the Riemann-Liouville derivative
NASA Astrophysics Data System (ADS)
Ahmood, Wasan Ajeel; Kiliçman, Adem
2017-01-01
In this paper, the aim of this work is to study theorem for the one-dimensional space-time fractional deriative, generalize some function for the one-dimensional fractional by table represents the fractional Laplace transforms of some elementary functions to be valid for the multi-dimensional fractional Laplace transform and give the definition of the multi-dimensional fractional Laplace transform. This study includes that, dedicate the one-dimensional fractional Laplace transform for functions of only one independent variable and develop of the one-dimensional fractional Laplace transform to multi-dimensional fractional Laplace transform based on the modified Riemann-Liouville derivative.
Pohlheim, Hartmut
2006-01-01
Multidimensional scaling as a technique for the presentation of high-dimensional data with standard visualization techniques is presented. The technique used is often known as Sammon mapping. We explain the mathematical foundations of multidimensional scaling and its robust calculation. We also demonstrate the use of this technique in the area of evolutionary algorithms. First, we present the visualization of the path through the search space of the best individuals during an optimization run. We then apply multidimensional scaling to the comparison of multiple runs regarding the variables of individuals and multi-criteria objective values (path through the solution space).
Best Design for Multidimensional Computerized Adaptive Testing With the Bifactor Model
Seo, Dong Gi; Weiss, David J.
2015-01-01
Most computerized adaptive tests (CATs) have been studied using the framework of unidimensional item response theory. However, many psychological variables are multidimensional and might benefit from using a multidimensional approach to CATs. This study investigated the accuracy, fidelity, and efficiency of a fully multidimensional CAT algorithm (MCAT) with a bifactor model using simulated data. Four item selection methods in MCAT were examined for three bifactor pattern designs using two multidimensional item response theory models. To compare MCAT item selection and estimation methods, a fixed test length was used. The Ds-optimality item selection improved θ estimates with respect to a general factor, and either D- or A-optimality improved estimates of the group factors in three bifactor pattern designs under two multidimensional item response theory models. The MCAT model without a guessing parameter functioned better than the MCAT model with a guessing parameter. The MAP (maximum a posteriori) estimation method provided more accurate θ estimates than the EAP (expected a posteriori) method under most conditions, and MAP showed lower observed standard errors than EAP under most conditions, except for a general factor condition using Ds-optimality item selection. PMID:29795848
A Multidimensional Study of Vocal Function Following Radiation Therapy for Laryngeal Cancers.
Angadi, Vrushali; Dressler, Emily; Stemple, Joseph
2017-06-01
Radiation therapy (XRT) has proven to be an effective curative modality in the treatment of laryngeal cancers. However, XRT also has deleterious effects on vocal function. To demonstrate the multidimensional nature of deficits in vocal function as a result of radiation therapy for laryngeal cancer. Cohort study. Vocal function parameters were chosen from the 5 domains of voice assessment to complete a multidimensional assessment battery. Adults irradiated (XRT group) for laryngeal cancers were compared to a control group of individuals with no history of head and neck cancers or radiation therapy. The control group was matched in age, sex, and pack years of smoking. Eighteen participants were recruited for the study. The XRT group demonstrated significantly worse clinical values as compared to the control group across select parameters in the each of the 5 domains of voice assessment. Radiation therapy for laryngeal cancers results in multidimensional deficits in vocal function. Notably, these deficits persist long term. In the present study sample, multidimensional deficits were persistent 2 to 7 years following completion of XRT. The observed multidimensional persistent vocal difficulties highlight the importance of vocal rehabilitation in the irradiated larynx cancer population.
PaperBLAST: Text Mining Papers for Information about Homologs
Price, Morgan N.; Arkin, Adam P.
2017-08-15
Large-scale genome sequencing has identified millions of protein-coding genes whose function is unknown. Many of these proteins are similar to characterized proteins from other organisms, but much of this information is missing from annotation databases and is hidden in the scientific literature. To make this information accessible, PaperBLAST uses EuropePMC to search the full text of scientific articles for references to genes. PaperBLAST also takes advantage of curated resources (Swiss-Prot, GeneRIF, and EcoCyc) that link protein sequences to scientific articles. PaperBLAST’s database includes over 700,000 scientific articles that mention over 400,000 different proteins. Given a protein of interest, PaperBLAST quicklymore » finds similar proteins that are discussed in the literature and presents snippets of text from relevant articles or from the curators. With the recent explosion of genome sequencing data, there are now millions of uncharacterized proteins. If a scientist becomes interested in one of these proteins, it can be very difficult to find information as to its likely function. Often a protein whose sequence is similar, and which is likely to have a similar function, has been studied already, but this information is not available in any database. To help find articles about similar proteins, PaperBLAST searches the full text of scientific articles for protein identifiers or gene identifiers, and it links these articles to protein sequences. Then, given a protein of interest, it can quickly find similar proteins in its database by using standard software (BLAST), and it can show snippets of text from relevant papers. We hope that PaperBLAST will make it easier for biologists to predict proteins’ functions.« less
PaperBLAST: Text Mining Papers for Information about Homologs
DOE Office of Scientific and Technical Information (OSTI.GOV)
Price, Morgan N.; Arkin, Adam P.
Large-scale genome sequencing has identified millions of protein-coding genes whose function is unknown. Many of these proteins are similar to characterized proteins from other organisms, but much of this information is missing from annotation databases and is hidden in the scientific literature. To make this information accessible, PaperBLAST uses EuropePMC to search the full text of scientific articles for references to genes. PaperBLAST also takes advantage of curated resources (Swiss-Prot, GeneRIF, and EcoCyc) that link protein sequences to scientific articles. PaperBLAST’s database includes over 700,000 scientific articles that mention over 400,000 different proteins. Given a protein of interest, PaperBLAST quicklymore » finds similar proteins that are discussed in the literature and presents snippets of text from relevant articles or from the curators. With the recent explosion of genome sequencing data, there are now millions of uncharacterized proteins. If a scientist becomes interested in one of these proteins, it can be very difficult to find information as to its likely function. Often a protein whose sequence is similar, and which is likely to have a similar function, has been studied already, but this information is not available in any database. To help find articles about similar proteins, PaperBLAST searches the full text of scientific articles for protein identifiers or gene identifiers, and it links these articles to protein sequences. Then, given a protein of interest, it can quickly find similar proteins in its database by using standard software (BLAST), and it can show snippets of text from relevant papers. We hope that PaperBLAST will make it easier for biologists to predict proteins’ functions.« less
PaperBLAST: Text Mining Papers for Information about Homologs
Arkin, Adam P.
2017-01-01
ABSTRACT Large-scale genome sequencing has identified millions of protein-coding genes whose function is unknown. Many of these proteins are similar to characterized proteins from other organisms, but much of this information is missing from annotation databases and is hidden in the scientific literature. To make this information accessible, PaperBLAST uses EuropePMC to search the full text of scientific articles for references to genes. PaperBLAST also takes advantage of curated resources (Swiss-Prot, GeneRIF, and EcoCyc) that link protein sequences to scientific articles. PaperBLAST’s database includes over 700,000 scientific articles that mention over 400,000 different proteins. Given a protein of interest, PaperBLAST quickly finds similar proteins that are discussed in the literature and presents snippets of text from relevant articles or from the curators. PaperBLAST is available at http://papers.genomics.lbl.gov/. IMPORTANCE With the recent explosion of genome sequencing data, there are now millions of uncharacterized proteins. If a scientist becomes interested in one of these proteins, it can be very difficult to find information as to its likely function. Often a protein whose sequence is similar, and which is likely to have a similar function, has been studied already, but this information is not available in any database. To help find articles about similar proteins, PaperBLAST searches the full text of scientific articles for protein identifiers or gene identifiers, and it links these articles to protein sequences. Then, given a protein of interest, it can quickly find similar proteins in its database by using standard software (BLAST), and it can show snippets of text from relevant papers. We hope that PaperBLAST will make it easier for biologists to predict proteins’ functions. PMID:28845458
PaperBLAST: Text Mining Papers for Information about Homologs.
Price, Morgan N; Arkin, Adam P
2017-01-01
Large-scale genome sequencing has identified millions of protein-coding genes whose function is unknown. Many of these proteins are similar to characterized proteins from other organisms, but much of this information is missing from annotation databases and is hidden in the scientific literature. To make this information accessible, PaperBLAST uses EuropePMC to search the full text of scientific articles for references to genes. PaperBLAST also takes advantage of curated resources (Swiss-Prot, GeneRIF, and EcoCyc) that link protein sequences to scientific articles. PaperBLAST's database includes over 700,000 scientific articles that mention over 400,000 different proteins. Given a protein of interest, PaperBLAST quickly finds similar proteins that are discussed in the literature and presents snippets of text from relevant articles or from the curators. PaperBLAST is available at http://papers.genomics.lbl.gov/. IMPORTANCE With the recent explosion of genome sequencing data, there are now millions of uncharacterized proteins. If a scientist becomes interested in one of these proteins, it can be very difficult to find information as to its likely function. Often a protein whose sequence is similar, and which is likely to have a similar function, has been studied already, but this information is not available in any database. To help find articles about similar proteins, PaperBLAST searches the full text of scientific articles for protein identifiers or gene identifiers, and it links these articles to protein sequences. Then, given a protein of interest, it can quickly find similar proteins in its database by using standard software (BLAST), and it can show snippets of text from relevant papers. We hope that PaperBLAST will make it easier for biologists to predict proteins' functions.
Varni, James W; Limbers, Christine A; Bryant, William P; Wilson, Don P
2010-01-01
The PedsQL (Pediatric Quality of Life Inventory) is a modular instrument designed to measure health-related quality of life (HRQOL) and disease-specific symptoms in children and adolescents. The PedsQL Multidimensional Fatigue Scale was designed as a child self-report and parent proxy-report generic symptom-specific instrument to measure fatigue in pediatric patients. The objective of the present study was to determine the feasibility, reliability, and validity of the PedsQL Multidimensional Fatigue Scale in pediatric obesity. The 18-item PedsQL Multidimensional Fatigue Scale (General Fatigue, Sleep/Rest Fatigue, and Cognitive Fatigue domains) and the PedsQL 4.0 Generic Core Scales were completed by 41 pediatric patients with a physician-diagnosis of obesity and 43 parents from a hospital-based Pediatric Endocrinology Clinic. The PedsQL Multidimensional Fatigue Scale evidenced minimal missing responses (1.6%, child report; 0.5%, parent report), achieved excellent reliability for the Total Fatigue Scale Score (alpha = 0.90 child report, 0.90 parent report), distinguished between pediatric patients with obesity and healthy children, and was significantly correlated with the PedsQL 4.0 Generic Core Scales supporting construct validity. Pediatric patients with obesity experienced fatigue comparable with pediatric patients receiving cancer treatment, demonstrating the relative severity of their fatigue symptoms. The results demonstrate the measurement properties of the PedsQL Multidimensional Fatigue Scale in pediatric obesity. The findings suggest that the PedsQL Multidimensional Fatigue Scale may be utilized in the standardized evaluation of fatigue in pediatric patients with obesity.
Multidimensional Poverty and Child Survival in India
Mohanty, Sanjay K.
2011-01-01
Background Though the concept of multidimensional poverty has been acknowledged cutting across the disciplines (among economists, public health professionals, development thinkers, social scientists, policy makers and international organizations) and included in the development agenda, its measurement and application are still limited. Objectives and Methodology Using unit data from the National Family and Health Survey 3, India, this paper measures poverty in multidimensional space and examine the linkages of multidimensional poverty with child survival. The multidimensional poverty is measured in the dimension of knowledge, health and wealth and the child survival is measured with respect to infant mortality and under-five mortality. Descriptive statistics, principal component analyses and the life table methods are used in the analyses. Results The estimates of multidimensional poverty are robust and the inter-state differentials are large. While infant mortality rate and under-five mortality rate are disproportionately higher among the abject poor compared to the non-poor, there are no significant differences in child survival among educationally, economically and health poor at the national level. State pattern in child survival among the education, economical and health poor are mixed. Conclusion Use of multidimensional poverty measures help to identify abject poor who are unlikely to come out of poverty trap. The child survival is significantly lower among abject poor compared to moderate poor and non-poor. We urge to popularize the concept of multiple deprivations in research and program so as to reduce poverty and inequality in the population. PMID:22046384
ERIC Educational Resources Information Center
Benoit, Gerald
2002-01-01
Discusses data mining (DM) and knowledge discovery in databases (KDD), taking the view that KDD is the larger view of the entire process, with DM emphasizing the cleaning, warehousing, mining, and visualization of knowledge discovery in databases. Highlights include algorithms; users; the Internet; text mining; and information extraction.…
Multi-source and ontology-based retrieval engine for maize mutant phenotypes
Green, Jason M.; Harnsomburana, Jaturon; Schaeffer, Mary L.; Lawrence, Carolyn J.; Shyu, Chi-Ren
2011-01-01
Model Organism Databases, including the various plant genome databases, collect and enable access to massive amounts of heterogeneous information, including sequence data, gene product information, images of mutant phenotypes, etc, as well as textual descriptions of many of these entities. While a variety of basic browsing and search capabilities are available to allow researchers to query and peruse the names and attributes of phenotypic data, next-generation search mechanisms that allow querying and ranking of text descriptions are much less common. In addition, the plant community needs an innovative way to leverage the existing links in these databases to search groups of text descriptions simultaneously. Furthermore, though much time and effort have been afforded to the development of plant-related ontologies, the knowledge embedded in these ontologies remains largely unused in available plant search mechanisms. Addressing these issues, we have developed a unique search engine for mutant phenotypes from MaizeGDB. This advanced search mechanism integrates various text description sources in MaizeGDB to aid a user in retrieving desired mutant phenotype information. Currently, descriptions of mutant phenotypes, loci and gene products are utilized collectively for each search, though expansion of the search mechanism to include other sources is straightforward. The retrieval engine, to our knowledge, is the first engine to exploit the content and structure of available domain ontologies, currently the Plant and Gene Ontologies, to expand and enrich retrieval results in major plant genomic databases. Database URL: http:www.PhenomicsWorld.org/QBTA.php PMID:21558151
Conceptual biology, hypothesis discovery, and text mining: Swanson's legacy.
Bekhuis, Tanja
2006-04-03
Innovative biomedical librarians and information specialists who want to expand their roles as expert searchers need to know about profound changes in biology and parallel trends in text mining. In recent years, conceptual biology has emerged as a complement to empirical biology. This is partly in response to the availability of massive digital resources such as the network of databases for molecular biologists at the National Center for Biotechnology Information. Developments in text mining and hypothesis discovery systems based on the early work of Swanson, a mathematician and information scientist, are coincident with the emergence of conceptual biology. Very little has been written to introduce biomedical digital librarians to these new trends. In this paper, background for data and text mining, as well as for knowledge discovery in databases (KDD) and in text (KDT) is presented, then a brief review of Swanson's ideas, followed by a discussion of recent approaches to hypothesis discovery and testing. 'Testing' in the context of text mining involves partially automated methods for finding evidence in the literature to support hypothetical relationships. Concluding remarks follow regarding (a) the limits of current strategies for evaluation of hypothesis discovery systems and (b) the role of literature-based discovery in concert with empirical research. Report of an informatics-driven literature review for biomarkers of systemic lupus erythematosus is mentioned. Swanson's vision of the hidden value in the literature of science and, by extension, in biomedical digital databases, is still remarkably generative for information scientists, biologists, and physicians.
ERIC Educational Resources Information Center
Kirisci, Levent; Tarter, Ralph E.
2001-01-01
Designs and evaluates a multidimensional schema for the assessment of alcohol, tobacco and other drug use topology. Findings illustrate the value of multidimensional assessment for identifying youth at high risk for substance use disorder (SUD) as well as for elucidating the factors contributing to the transition to suprathreshold SUD. (Contains…
ERIC Educational Resources Information Center
Senarat, Somprasong; Tayraukham, Sombat; Piyapimonsit, Chatsiri; Tongkhambanjong, Sakesan
2013-01-01
The purpose of this research is to develop a multidimensional computerized adaptive test for diagnosing the cognitive process of grade 7 students in learning algebra by applying multidimensional item response theory. The research is divided into 4 steps: 1) the development of item bank of algebra, 2) the development of the multidimensional…
pGenN, a gene normalization tool for plant genes and proteins in scientific literature.
Ding, Ruoyao; Arighi, Cecilia N; Lee, Jung-Youn; Wu, Cathy H; Vijay-Shanker, K
2015-01-01
Automatically detecting gene/protein names in the literature and connecting them to databases records, also known as gene normalization, provides a means to structure the information buried in free-text literature. Gene normalization is critical for improving the coverage of annotation in the databases, and is an essential component of many text mining systems and database curation pipelines. In this manuscript, we describe a gene normalization system specifically tailored for plant species, called pGenN (pivot-based Gene Normalization). The system consists of three steps: dictionary-based gene mention detection, species assignment, and intra species normalization. We have developed new heuristics to improve each of these phases. We evaluated the performance of pGenN on an in-house expertly annotated corpus consisting of 104 plant relevant abstracts. Our system achieved an F-value of 88.9% (Precision 90.9% and Recall 87.2%) on this corpus, outperforming state-of-art systems presented in BioCreative III. We have processed over 440,000 plant-related Medline abstracts using pGenN. The gene normalization results are stored in a local database for direct query from the pGenN web interface (proteininformationresource.org/pgenn/). The annotated literature corpus is also publicly available through the PIR text mining portal (proteininformationresource.org/iprolink/).
Integrative neuroscience: the role of a standardized database.
Gordon, E; Cooper, N; Rennie, C; Hermens, D; Williams, L M
2005-04-01
Most brain related databases bring together specialized information, with a growing number that include neuroimaging measures. This article outlines the potential use and insights from the first entirely standardized and centralized database, which integrates information from neuroimaging measures (EEG, event related potential (ERP), structural/functional MRI), arousal (skin conductance responses (SCR)s, heart rate, respiration), neuropsychological and personality tests, genomics and demographics: The Brain Resource International Database. It comprises data from over 2000 "normative" subjects and a growing number of patients with neurological and psychiatric illnesses, acquired from over 50 laboratories (in the U.S.A, United Kingdom, Holland, South Africa, Israel and Australia), all with identical equipment and experimental procedures. Three primary goals of this database are to quantify individual differences in normative brain function, to compare an individual's performance to their database peers, and to provide a robust normative framework for clinical assessment and treatment prediction. We present three example demonstrations in relation to these goals. First, we show how consistent age differences may be quantified when large subject numbers are available, using EEG and ERP data from nearly 2000 stringently screened. normative subjects. Second, the use of a normalization technique provides a means to compare clinical subjects (50 ADHD subjects in this study) to the normative database with the effects of age and gender taken into account. Third, we show how a profile of EEG/ERP and autonomic measures potentially provides a means to predict treatment response in ADHD subjects. The example data consists of EEG under eyes open and eyes closed and ERP data for auditory oddball, working memory and Go-NoGo paradigms. Autonomic measures of skin conductance (tonic skin conductance level, SCL, and phasic skin conductance responses, SCRs) were acquired simultaneously with central EEG/ERP measures. The findings show that the power of large samples, tested using standardized protocols, allows for the quantification of individual differences that can subsequently be used to control such variation and to enhance the sensitivity and specificity of comparisons between normative and clinical groups. In terms of broader significance, the combination of size and multidimensional measures tapping the brain's core cognitive competencies, may provide a normative and evidence-based framework for individually-based assessments in "Personalized Medicine."
What is integrability of discrete variational systems?
Boll, Raphael; Petrera, Matteo; Suris, Yuri B
2014-02-08
We propose a notion of a pluri-Lagrangian problem, which should be understood as an analogue of multi-dimensional consistency for variational systems. This is a development along the line of research of discrete integrable Lagrangian systems initiated in 2009 by Lobb and Nijhoff, however, having its more remote roots in the theory of pluriharmonic functions, in the Z -invariant models of statistical mechanics and their quasiclassical limit, as well as in the theory of variational symmetries going back to Noether. A d -dimensional pluri-Lagrangian problem can be described as follows: given a d -form [Formula: see text] on an m -dimensional space (called multi-time, m > d ), whose coefficients depend on a sought-after function x of m independent variables (called field), find those fields x which deliver critical points to the action functionals [Formula: see text] for any d -dimensional manifold Σ in the multi-time. We derive the main building blocks of the multi-time Euler-Lagrange equations for a discrete pluri-Lagrangian problem with d =2, the so-called corner equations, and discuss the notion of consistency of the system of corner equations. We analyse the system of corner equations for a special class of three-point two-forms, corresponding to integrable quad-equations of the ABS list. This allows us to close a conceptual gap of the work by Lobb and Nijhoff by showing that the corresponding two-forms are closed not only on solutions of (non-variational) quad-equations, but also on general solutions of the corresponding corner equations. We also find an example of a pluri-Lagrangian system not coming from a multi-dimensionally consistent system of quad-equations.
DOE Research and Development Accomplishments Help
be used to search, locate, access, and electronically download full-text research and development (R Browse Downloading, Viewing, and/or Searching Full-text Documents/Pages Searching the Database Search Features Search allows you to search the OCRed full-text document and bibliographic information, the
The Effects of Elaboration on Self-Learning Procedures from Text.
ERIC Educational Resources Information Center
Yang, Fu-mei
This study investigated the effects of augmenting and deleting elaborations in an existing self-instructional text for a micro-computer database application, "Microsoft Works User's Manual." A total of 60 undergraduate students were randomly assigned to the original, elaborated, or unelaborated text versions. The elaborated version…
A Text Knowledge Base from the AI Handbook.
ERIC Educational Resources Information Center
Simmons, Robert F.
1987-01-01
Describes a prototype natural language text knowledge system (TKS) that was used to organize 50 pages of a handbook on artificial intelligence as an inferential knowledge base with natural language query and command capabilities. Representation of text, database navigation, query systems, discourse structuring, and future research needs are…
ERIC Educational Resources Information Center
Lamothe, Alain R.
2011-01-01
The purpose of this paper is to report the results of a quantitative analysis exploring the interaction and relationship between the online database and electronic journal collections at the J. N. Desmarais Library of Laurentian University. A very strong relationship exists between the number of searches and the size of the online database…
Worldwide Report, Telecommunications Policy, Research and Development
1985-12-31
3 Hong Kong Database Signs Contract With PRC (K. Gopinath; Hong Kong HONGKONG STANDARD, 18 Oct 85) .... 5 Briefs Hong Kong-London Data Link...KONG HONG KONG DATABASE SIGNS CONTRACT WITH PRC Hong Kong HONGKONG STANDARD in English Supplement 18 Oct 85 p 1 [Article by K. Gopinath] [Text...information group. The agreement between state-owned China Hua Yang Technology and Trade Corporation and DataBase Asia of Hong- kong, authorises
Approximation of epidemic models by diffusion processes and their statistical inference.
Guy, Romain; Larédo, Catherine; Vergu, Elisabeta
2015-02-01
Multidimensional continuous-time Markov jump processes [Formula: see text] on [Formula: see text] form a usual set-up for modeling [Formula: see text]-like epidemics. However, when facing incomplete epidemic data, inference based on [Formula: see text] is not easy to be achieved. Here, we start building a new framework for the estimation of key parameters of epidemic models based on statistics of diffusion processes approximating [Formula: see text]. First, previous results on the approximation of density-dependent [Formula: see text]-like models by diffusion processes with small diffusion coefficient [Formula: see text], where [Formula: see text] is the population size, are generalized to non-autonomous systems. Second, our previous inference results on discretely observed diffusion processes with small diffusion coefficient are extended to time-dependent diffusions. Consistent and asymptotically Gaussian estimates are obtained for a fixed number [Formula: see text] of observations, which corresponds to the epidemic context, and for [Formula: see text]. A correction term, which yields better estimates non asymptotically, is also included. Finally, performances and robustness of our estimators with respect to various parameters such as [Formula: see text] (the basic reproduction number), [Formula: see text], [Formula: see text] are investigated on simulations. Two models, [Formula: see text] and [Formula: see text], corresponding to single and recurrent outbreaks, respectively, are used to simulate data. The findings indicate that our estimators have good asymptotic properties and behave noticeably well for realistic numbers of observations and population sizes. This study lays the foundations of a generic inference method currently under extension to incompletely observed epidemic data. Indeed, contrary to the majority of current inference techniques for partially observed processes, which necessitates computer intensive simulations, our method being mostly an analytical approach requires only the classical optimization steps.
Holm, Sven; Russell, Greg; Nourrit, Vincent; McLoughlin, Niall
2017-01-01
A database of retinal fundus images, the DR HAGIS database, is presented. This database consists of 39 high-resolution color fundus images obtained from a diabetic retinopathy screening program in the UK. The NHS screening program uses service providers that employ different fundus and digital cameras. This results in a range of different image sizes and resolutions. Furthermore, patients enrolled in such programs often display other comorbidities in addition to diabetes. Therefore, in an effort to replicate the normal range of images examined by grading experts during screening, the DR HAGIS database consists of images of varying image sizes and resolutions and four comorbidity subgroups: collectively defined as the diabetic retinopathy, hypertension, age-related macular degeneration, and Glaucoma image set (DR HAGIS). For each image, the vasculature has been manually segmented to provide a realistic set of images on which to test automatic vessel extraction algorithms. Modified versions of two previously published vessel extraction algorithms were applied to this database to provide some baseline measurements. A method based purely on the intensity of images pixels resulted in a mean segmentation accuracy of 95.83% ([Formula: see text]), whereas an algorithm based on Gabor filters generated an accuracy of 95.71% ([Formula: see text]).
Multi-dimensional quantum state sharing based on quantum Fourier transform
NASA Astrophysics Data System (ADS)
Qin, Huawang; Tso, Raylin; Dai, Yuewei
2018-03-01
A scheme of multi-dimensional quantum state sharing is proposed. The dealer performs the quantum SUM gate and the quantum Fourier transform to encode a multi-dimensional quantum state into an entanglement state. Then the dealer distributes each participant a particle of the entanglement state, to share the quantum state among n participants. In the recovery, n-1 participants measure their particles and supply their measurement results; the last participant performs the unitary operation on his particle according to these measurement results and can reconstruct the initial quantum state. The proposed scheme has two merits: It can share the multi-dimensional quantum state and it does not need the entanglement measurement.
Some theorems and properties of multi-dimensional fractional Laplace transforms
NASA Astrophysics Data System (ADS)
Ahmood, Wasan Ajeel; Kiliçman, Adem
2016-06-01
The aim of this work is to study theorems and properties for the one-dimensional fractional Laplace transform, generalize some properties for the one-dimensional fractional Lapalce transform to be valid for the multi-dimensional fractional Lapalce transform and is to give the definition of the multi-dimensional fractional Lapalce transform. This study includes: dedicate the one-dimensional fractional Laplace transform for functions of only one independent variable with some of important theorems and properties and develop of some properties for the one-dimensional fractional Laplace transform to multi-dimensional fractional Laplace transform. Also, we obtain a fractional Laplace inversion theorem after a short survey on fractional analysis based on the modified Riemann-Liouville derivative.
SCRIPDB: a portal for easy access to syntheses, chemicals and reactions in patents
Heifets, Abraham; Jurisica, Igor
2012-01-01
The patent literature is a rich catalog of biologically relevant chemicals; many public and commercial molecular databases contain the structures disclosed in patent claims. However, patents are an equally rich source of metadata about bioactive molecules, including mechanism of action, disease class, homologous experimental series, structural alternatives, or the synthetic pathways used to produce molecules of interest. Unfortunately, this metadata is discarded when chemical structures are deposited separately in databases. SCRIPDB is a chemical structure database designed to make this metadata accessible. SCRIPDB provides the full original patent text, reactions and relationships described within any individual patent, in addition to the molecular files common to structural databases. We discuss how such information is valuable in medical text mining, chemical image analysis, reaction extraction and in silico pharmaceutical lead optimization. SCRIPDB may be searched by exact chemical structure, substructure or molecular similarity and the results may be restricted to patents describing synthetic routes. SCRIPDB is available at http://dcv.uhnres.utoronto.ca/SCRIPDB. PMID:22067445
Chen, Chou-Cheng; Ho, Chung-Liang
2014-01-01
While a huge amount of information about biological literature can be obtained by searching the PubMed database, reading through all the titles and abstracts resulting from such a search for useful information is inefficient. Text mining makes it possible to increase this efficiency. Some websites use text mining to gather information from the PubMed database; however, they are database-oriented, using pre-defined search keywords while lacking a query interface for user-defined search inputs. We present the PubMed Abstract Reading Helper (PubstractHelper) website which combines text mining and reading assistance for an efficient PubMed search. PubstractHelper can accept a maximum of ten groups of keywords, within each group containing up to ten keywords. The principle behind the text-mining function of PubstractHelper is that keywords contained in the same sentence are likely to be related. PubstractHelper highlights sentences with co-occurring keywords in different colors. The user can download the PMID and the abstracts with color markings to be reviewed later. The PubstractHelper website can help users to identify relevant publications based on the presence of related keywords, which should be a handy tool for their research. http://bio.yungyun.com.tw/ATM/PubstractHelper.aspx and http://holab.med.ncku.edu.tw/ATM/PubstractHelper.aspx.
The Necessity-Concerns-Framework: A Multidimensional Theory Benefits from Multidimensional Analysis
Phillips, L. Alison; Diefenbach, Michael; Kronish, Ian M.; Negron, Rennie M.; Horowitz, Carol R.
2014-01-01
Background Patients’ medication-related concerns and necessity-beliefs predict adherence. Evaluation of the potentially complex interplay of these two dimensions has been limited because of methods that reduce them to a single dimension (difference scores). Purpose We use polynomial regression to assess the multidimensional effect of stroke-event survivors’ medication-related concerns and necessity-beliefs on their adherence to stroke-prevention medication. Methods Survivors (n=600) rated their concerns, necessity-beliefs, and adherence to medication. Confirmatory and exploratory polynomial regression determined the best-fitting multidimensional model. Results As posited by the Necessity-Concerns Framework (NCF), the greatest and lowest adherence was reported by those with strong necessity-beliefs/weak concerns and strong concerns/weak necessity-beliefs, respectively. However, as could not be assessed using a difference-score model, patients with ambivalent beliefs were less adherent than those exhibiting indifference. Conclusions Polynomial regression allows for assessment of the multidimensional nature of the NCF. Clinicians/Researchers should be aware that concerns and necessity dimensions are not polar opposites. PMID:24500078
The necessity-concerns framework: a multidimensional theory benefits from multidimensional analysis.
Phillips, L Alison; Diefenbach, Michael A; Kronish, Ian M; Negron, Rennie M; Horowitz, Carol R
2014-08-01
Patients' medication-related concerns and necessity-beliefs predict adherence. Evaluation of the potentially complex interplay of these two dimensions has been limited because of methods that reduce them to a single dimension (difference scores). We use polynomial regression to assess the multidimensional effect of stroke-event survivors' medication-related concerns and necessity beliefs on their adherence to stroke-prevention medication. Survivors (n = 600) rated their concerns, necessity beliefs, and adherence to medication. Confirmatory and exploratory polynomial regression determined the best-fitting multidimensional model. As posited by the necessity-concerns framework (NCF), the greatest and lowest adherence was reported by those necessity weak concerns and strong concerns/weak Necessity-Beliefs, respectively. However, as could not be assessed using a difference-score model, patients with ambivalent beliefs were less adherent than those exhibiting indifference. Polynomial regression allows for assessment of the multidimensional nature of the NCF. Clinicians/Researchers should be aware that concerns and necessity dimensions are not polar opposites.
Weng, Chia-Ying; Lin, I-Mei; Jiang, Ding-Yu
2010-08-01
The purpose of this study was to examine the effects of gender on the relationship between multidimensional hostility and psychosomatic symptoms in Chinese culture. The participants in this study were 398 Chinese college students (40% female) recruited from Taiwan. Four dimensions of multidimensional hostility-hostility cognition, hostility affect, expressive hostility behavior, and suppressive hostility behavior-were measured by the Chinese Hostility Inventory. After controlling for the effects of depression and anxiety, the results of path analysis revealed that the multidimensional hostility predicted psychosomatic symptoms directly, and predicted psychosomatic symptoms indirectly through negative health behavior. Furthermore, gender moderated the relationships between multidimensional hostility and health outcomes. Expressive hostility exacerbated psychosomatic symptom in females but buffered it in males, while affective hostility exacerbated psychosomatic symptoms in males. Additionally, suppressive hostility behavior was correlated to psychosomatic symptoms indirectly through negative health behavior in females. Moreover, expressive hostility was correlated to psychosomatic symptoms indirectly through negative health behavior more in males than in females.
The ergonomic relevance of anthropometrical proportions. Part I: body length measures.
Küchmeister, Gerd; Behrenbruch, Kay; Ursula, Pieper; Leyk, Dieter
2009-06-01
Anthropometrical findings about proportions of the human body have a long tradition; most of them, however, are rather theoretical. To improve this situation a pragmatic and applicable definition of proportions, normal as well as deviating, is needed. The purpose of this paper is to set up three proportion types for body length measures: leggy-normal-stocky, and their rate in the population of adults in an industrial society (Germany). Based on queries of an actual anthropometrical database metrical limits for these three types are suggested regarding their influence on the most important body length measures in standing and sitting positions. The number of cases with normal proportions is about 60%, leggy and stocky types both reaching up to 20%. The metrical limits are proposed in order to account for differences between those proportion types which are relevant for ergonomics. When solving complex multidimensional design tasks (e.g. workplaces, vehicle interior design, personal equipment) users of anthropometrical data who do not have the opportunity to work with multivariate databases need supplementary information about the proportions of the human body. For this reason such supplementary information concerning proportions is being given here by physiological anthropologists. The application of the findings presented is exemplified in scenarios of ergonomic relevance.
On new physics searches with multidimensional differential shapes
NASA Astrophysics Data System (ADS)
Ferreira, Felipe; Fichet, Sylvain; Sanz, Veronica
2018-03-01
In the context of upcoming new physics searches at the LHC, we investigate the impact of multidimensional differential rates in typical LHC analyses. We discuss the properties of shape information, and argue that multidimensional rates bring limited information in the scope of a discovery, but can have a large impact on model discrimination. We also point out subtleties about systematic uncertainties cancellations and the Cauchy-Schwarz bound on interference terms.
ERIC Educational Resources Information Center
Tilga, Henri; Hein, Vello; Koka, Andre
2017-01-01
This research aimed to develop and validate an instrument to assess the students' perceptions of the teachers' autonomy-supportive behavior by the multi-dimensional scale (Multi-Dimensional Perceived Autonomy Support Scale for Physical Education). The participants were 1,476 students aged 12- to 15-years-old. In Study 1, a pool of 37 items was…
Multidimensional modulation for next-generation transmission systems
NASA Astrophysics Data System (ADS)
Millar, David S.; Koike-Akino, Toshiaki; Kojima, Keisuke; Parsons, Kieran
2017-01-01
Recent research in multidimensional modulation has shown great promise in long reach applications. In this work, we will investigate the origins of this gain, the different approaches to multidimensional constellation design, and different performance metrics for coded modulation. We will also discuss the reason that such coded modulation schemes seem to have limited application at shorter distances, and the potential for other coded modulation schemes in future transmission systems.
Multidimensional Data Modeling for Business Process Analysis
NASA Astrophysics Data System (ADS)
Mansmann, Svetlana; Neumuth, Thomas; Scholl, Marc H.
The emerging area of business process intelligence attempts to enhance the analytical capabilities of business process management systems by employing data warehousing and mining technologies. This paper presents an approach to re-engineering the business process modeling in conformity with the multidimensional data model. Since the business process and the multidimensional model are driven by rather different objectives and assumptions, there is no straightforward solution to converging these models.
Al-Gamal, Ekhlas; Long, Tony
2017-09-01
Fatigue is considered to be one of the most reported symptoms experienced by children with cancer. A major aim of this study was to develop an Arabic version of the Pediatric Quality of Life (PedsQL) Multidimensional Fatigue Scale (child report) and to test its psychometric proprieties for the assessment of fatigue in Arabic children with cancer. The PedsQL Multidimensional Fatigue Scale (Arabic version) and the PedsQL TM 4.0 Generic Core scale (existing Arabic version) were completed by 70 Jordanian children with cancer. Cronbach's alpha coefficients were found to be 0.90 for the total PedsQL Multidimensional Fatigue Scale (Arabic version), 0.94 for the general fatigue subscale, 0.67 for the sleep/rest fatigue subscale, and 0.87 for the cognitive fatigue subscale. The PedsQL Multidimensional Fatigue Scale scores correlated significantly with the PedsQL TM 4.0 Generic Core scale and demonstrated good construct validity. The results demonstrate excellent reliability and good validity of the PedsQL Multidimensional Fatigue Scale (Arabic version) for children with cancer. This is the first validated scale that assesses fatigue in Arabic children with cancer. The English scale has been used with several pediatric clinical populations, so this Arabic version may be equally useful beyond the field of cancer.
Sims, Mario; Wyatt, Sharon B.; Gutierrez, Mary Lou; Taylor, Herman A.; Williams, David R.
2009-01-01
Objective Assessing the discrimination-health disparities hypothesis requires psychometrically sound, multidimensional measures of discrimination. Among the available discrimination measures, few are multidimensional and none have adequate psychometric testing in a large, African American sample. We report the development and psychometric testing of the multidimensional Jackson Heart Study Discrimination (JHSDIS) Instrument. Methods A multidimensional measure assessing the occurrence, frequency, attribution, and coping responses to perceived everyday and lifetime discrimination; lifetime burden of discrimination; and effect of skin color was developed and tested in the 5302-member cohort of the Jackson Heart Study. Internal consistency was calculated by using Cronbach α. coefficient. Confirmatory factor analysis established the dimensions, and intercorrelation coefficients assessed the discriminant validity of the instrument. Setting Tri-county area of the Jackson, MS metropolitan statistical area. Results The JHSDIS was psychometrically sound (overall α=.78, .84 and .77, respectively, for the everyday and lifetime subscales). Confirmatory factor analysis yielded 11 factors, which confirmed the a priori dimensions represented. Conclusions The JHSDIS combined three scales into a single multidimensional instrument with good psychometric properties in a large sample of African Americans. This analysis lays the foundation for using this instrument in research that will examine the association between perceived discrimination and CVD among African Americans. PMID:19341164
Hutz, Janna E; Nelson, Thomas; Wu, Hua; McAllister, Gregory; Moutsatsos, Ioannis; Jaeger, Savina A; Bandyopadhyay, Somnath; Nigsch, Florian; Cornett, Ben; Jenkins, Jeremy L; Selinger, Douglas W
2013-04-01
Screens using high-throughput, information-rich technologies such as microarrays, high-content screening (HCS), and next-generation sequencing (NGS) have become increasingly widespread. Compared with single-readout assays, these methods produce a more comprehensive picture of the effects of screened treatments. However, interpreting such multidimensional readouts is challenging. Univariate statistics such as t-tests and Z-factors cannot easily be applied to multidimensional profiles, leaving no obvious way to answer common screening questions such as "Is treatment X active in this assay?" and "Is treatment X different from (or equivalent to) treatment Y?" We have developed a simple, straightforward metric, the multidimensional perturbation value (mp-value), which can be used to answer these questions. Here, we demonstrate application of the mp-value to three data sets: a multiplexed gene expression screen of compounds and genomic reagents, a microarray-based gene expression screen of compounds, and an HCS compound screen. In all data sets, active treatments were successfully identified using the mp-value, and simulations and follow-up analyses supported the mp-value's statistical and biological validity. We believe the mp-value represents a promising way to simplify the analysis of multidimensional data while taking full advantage of its richness.
NASA TLX: software for assessing subjective mental workload.
Cao, Alex; Chintamani, Keshav K; Pandya, Abhilash K; Ellis, R Darin
2009-02-01
The NASA Task Load Index (TLX) is a popular technique for measuring subjective mental workload. It relies on a multidimensional construct to derive an overall workload score based on a weighted average of ratings on six subscales: mental demand, physical demand, temporal demand, performance, effort, and frustration level. A program for implementing a computerized version of the NASA TLX is described. The software version assists in simplifying collection, postprocessing, and storage of raw data. The program collects raw data from the subject and calculates the weighted (or unweighted) workload score, which is output to a text file. The program can also be tailored to a specific experiment using a simple input text file, if desired. The program was designed in Visual Studio 2005 and is capable of running on a Pocket PC with Windows CE or on a PC with Windows 2000 or higher. The NASA TLX program is available for free download.
Varni, James W; Limbers, Christine A; Bryant, William P; Wilson, Don P
2009-08-01
The Pediatric Quality of Life Inventory (PedsQL, Mapi Research Trust, Lyon, France; www.pedsql.org) is a modular instrument designed to measure health-related quality of life and disease-specific symptoms in children and adolescents. The PedsQL Multidimensional Fatigue Scale was designed as a child self-report and parent proxy-report generic symptom-specific instrument to measure fatigue in pediatric patients. The objective of the present study was to determine the feasibility, reliability, and validity of the PedsQL Multidimensional Fatigue Scale in type 1 diabetes. The 18-item PedsQL Multidimensional Fatigue Scale (General Fatigue, Sleep/Rest Fatigue, and Cognitive Fatigue domains) and the PedsQL 4.0 Generic Core Scales were administered to 83 pediatric patients with type 1 diabetes and 84 parents. The PedsQL Multidimensional Fatigue Scale evidenced minimal missing responses (0.3% child report and 0.3% parent report), achieved excellent reliability for the Total Fatigue Scale score (alpha= 0.92 child report, 0.94 parent report), distinguished between pediatric patients with diabetes and healthy children, and was significantly correlated with the PedsQL 4.0 Generic Core Scales supporting construct validity. Pediatric patients with diabetes experienced fatigue that was comparable to pediatric patients with cancer on treatment, demonstrating the relative severity of their fatigue symptoms. The results demonstrate the measurement properties of the PedsQL Multidimensional Fatigue Scale in type 1 diabetes. The findings suggest that the PedsQL Multidimensional Fatigue Scale may be utilized in the standardized evaluation of fatigue in pediatric patients with type 1 diabetes.
Varni, James W; Limbers, Christine A
2008-02-01
The PedsQL (Pediatric Quality of Life Inventory) is a modular instrument designed to measure health-related quality of life (HRQOL) and disease-specific symptoms in children and adolescents ages 2-18. The PedsQL Multidimensional Fatigue Scale was designed as a generic symptom-specific instrument to measure fatigue in pediatric patients ages 2-18. Since a sizeable number of pediatric patients prefer to remain with their pediatric providers after age 18, the objective of the present study was to determine the feasibility, reliability, and validity of the PedsQL Multidimensional Fatigue Scale in young adults. The 18-item PedsQL Multidimensional Fatigue Scale (General Fatigue, Sleep/Rest Fatigue, and Cognitive Fatigue domains), the PedsQL 4.0 Generic Core Scales Young Adult Version, and the SF-8 Health Survey were completed by 423 university students ages 18-25. The PedsQL Multidimensional Fatigue Scale evidenced minimal missing responses, achieved excellent reliability for the Total Scale Score (alpha = 0.90), distinguished between healthy young adults and young adults with chronic health conditions, was significantly correlated with the relevant PedsQL 4.0 Generic Core Scales and the SF-8 standardized scores, and demonstrated a factor-derived structure largely consistent with the a priori conceptual model. The results demonstrate the measurement properties of the PedsQL Multidimensional Fatigue Scale in a convenience sample of young adult university students. The findings suggest that the PedsQL Multidimensional Fatigue Scale may be utilized in the evaluation of fatigue for a broad age range.
Fatigue and multidimensional disease severity in chronic obstructive pulmonary disease.
Inal-Ince, Deniz; Savci, Sema; Saglam, Melda; Calik, Ebru; Arikan, Hulya; Bosnak-Guclu, Meral; Vardar-Yagli, Naciye; Coplu, Lutfi
2010-06-30
Fatigue is associated with longitudinal ratings of health in patients with chronic obstructive pulmonary disease (COPD). Although the degree of airflow obstruction is often used to grade disease severity in patients with COPD, multidimensional grading systems have recently been developed. The aim of this study was to investigate the relationship between perceived and actual fatigue level and multidimensional disease severity in patients with COPD. Twenty-two patients with COPD (aged 52-74 years) took part in the study. Multidimensional disease severity was measured using the SAFE and BODE indices. Perceived fatigue was assessed using the Fatigue Severity Scale (FSS) and the Fatigue Impact Scale (FIS). Peripheral muscle endurance was evaluated using the number of sit-ups, squats, and modified push-ups that each patient could do. Thirteen patients (59%) had severe fatigue, and their St George's Respiratory Questionnaire scores were significantly higher (p < 0.05). The SAFE index score was significantly correlated with the number of sit-ups, number of squats, FSS score and FIS score (p < 0.05). The BODE index was significantly associated with the numbers of sit-ups, squats and modified push-ups, and with the FSS and FIS scores (p < 0.05). Peripheral muscle endurance and fatigue perception in patients with COPD was related to multidimensional disease severity measured with both the SAFE and BODE indices. Improvements in perceived and actual fatigue levels may positively affect multidimensional disease severity and health status in COPD patients. Further research is needed to investigate the effects of fatigue perception and exercise training on patients with different stages of multidimensional COPD severity.
Austvoll-Dahlgren, Astrid; Guttersrud, Øystein; Nsangi, Allen; Semakula, Daniel; Oxman, Andrew D
2017-01-01
Background The Claim Evaluation Tools database contains multiple-choice items for measuring people’s ability to apply the key concepts they need to know to be able to assess treatment claims. We assessed items from the database using Rasch analysis to develop an outcome measure to be used in two randomised trials in Uganda. Rasch analysis is a form of psychometric testing relying on Item Response Theory. It is a dynamic way of developing outcome measures that are valid and reliable. Objectives To assess the validity, reliability and responsiveness of 88 items addressing 22 key concepts using Rasch analysis. Participants We administrated four sets of multiple-choice items in English to 1114 people in Uganda and Norway, of which 685 were children and 429 were adults (including 171 health professionals). We scored all items dichotomously. We explored summary and individual fit statistics using the RUMM2030 analysis package. We used SPSS to perform distractor analysis. Results Most items conformed well to the Rasch model, but some items needed revision. Overall, the four item sets had satisfactory reliability. We did not identify significant response dependence between any pairs of items and, overall, the magnitude of multidimensionality in the data was acceptable. The items had a high level of difficulty. Conclusion Most of the items conformed well to the Rasch model’s expectations. Following revision of some items, we concluded that most of the items were suitable for use in an outcome measure for evaluating the ability of children or adults to assess treatment claims. PMID:28550019
Martyna, Agnieszka; Michalska, Aleksandra; Zadora, Grzegorz
2015-05-01
The problem of interpretation of common provenance of the samples within the infrared spectra database of polypropylene samples from car body parts and plastic containers as well as Raman spectra databases of blue solid and metallic automotive paints was under investigation. The research involved statistical tools such as likelihood ratio (LR) approach for expressing the evidential value of observed similarities and differences in the recorded spectra. Since the LR models can be easily proposed for databases described by a few variables, research focused on the problem of spectra dimensionality reduction characterised by more than a thousand variables. The objective of the studies was to combine the chemometric tools easily dealing with multidimensionality with an LR approach. The final variables used for LR models' construction were derived from the discrete wavelet transform (DWT) as a data dimensionality reduction technique supported by methods for variance analysis and corresponded with chemical information, i.e. typical absorption bands for polypropylene and peaks associated with pigments present in the car paints. Univariate and multivariate LR models were proposed, aiming at obtaining more information about the chemical structure of the samples. Their performance was controlled by estimating the levels of false positive and false negative answers and using the empirical cross entropy approach. The results for most of the LR models were satisfactory and enabled solving the stated comparison problems. The results prove that the variables generated from DWT preserve signal characteristic, being a sparse representation of the original signal by keeping its shape and relevant chemical information.
Estimating Missing Features to Improve Multimedia Information Retrieval
DOE Office of Scientific and Technical Information (OSTI.GOV)
Bagherjeiran, A; Love, N S; Kamath, C
Retrieval in a multimedia database usually involves combining information from different modalities of data, such as text and images. However, all modalities of the data may not be available to form the query. The retrieval results from such a partial query are often less than satisfactory. In this paper, we present an approach to complete a partial query by estimating the missing features in the query. Our experiments with a database of images and their associated captions show that, with an initial text-only query, our completion method has similar performance to a full query with both image and text features.more » In addition, when we use relevance feedback, our approach outperforms the results obtained using a full query.« less
Puig-Barrachina, Vanessa; Vanroelen, Christophe; Vives, Alejandra; Martínez, José Miguel; Muntaner, Carles; Levecque, Katia; Benach, Joan; Louckx, Fred
2014-01-01
Precarious employment is becoming an increasingly important social determinant of health inequalities among workers. The way in which contemporary employment arrangements and their health consequences are addressed in empirical research is mostly based on the contract-related or employment instability dimension. A broader conceptual approach including various important characteristics of the degrading of employment conditions and relations is needed. The general objective of this paper is to empirically test a new multidimensional construct for measuring precarious employment in an existing database. Special focus is on the social distribution of precarious employment. A subsample of 21,415 participants in the EU-27 from the Fourth European Working Conditions Survey-2005 was analysed. A cross-sectional study of the social distribution of precarious employment was conducted through the analysis of proportional differences according to gender, social class and credentials for the European Union as a whole and within each country. The 8 dimensions of the employment precariousness construct were represented by 11 indicators. In general, women, workers without supervisory authority, those with fewer credentials, and those living in Eastern and Southern European countries suffer the highest levels of precarious employment. Exceptionally, men, workers with supervisory authority and those with the highest credentials suffer the highest levels of long working hours, schedule unpredictability and uncompensated flexible working times. This article offers the first validation for an innovative multidimensional conceptualisation of employment precariousness applied to the analysis of existing survey data, showing the unequal distribution of precarious employment across the European labour force. This set of indicators can be useful for monitoring precarious employment.
Video-Assisted Thoracic Surgical Lobectomy for Lung Cancer: Description of a Learning Curve.
Yao, Fei; Wang, Jian; Yao, Ju; Hang, Fangrong; Cao, Shiqi; Cao, Yongke
2017-07-01
Video-assisted thoracic surgical (VATS) lobectomy is gaining popularity in the treatment of lung cancer. The aim of this study is to investigate the learning curve of VATS lobectomy by using multidimensional methods and to compare the learning curve groups with respect to perioperative clinical outcomes. We retrospectively reviewed a prospective database to identify 67 consecutive patients who underwent VATS lobectomy for lung cancer by a single surgeon. The learning curve was analyzed by using moving average and the cumulative sum (CUSUM) method. With the moving average and CUSUM analyses for the operation time, patients were stratified into two groups, with chronological order defining early and late experiences. Perioperative clinical outcomes were compared between the two learning curve groups. According to the moving average method, the peak point for operation time occurred at the 26th case. The CUSUM method also showed the operation time peak point at the 26th case. When results were compared between early- and late-experience periods, the operation time, duration of chest drainage, and postoperative hospital stay were significantly longer in the early-experience group (cases 1 to 26). The intraoperative estimated blood loss was significantly less in the late-experience group (cases 27 to 67). CUSUM charts showed a decreasing duration of chest drainage after the 36th case and shortening postoperative hospital stay after the 37th case. Multidimensional statistical analyses suggested that the learning curve for VATS lobectomy for lung cancer required ∼26 cases. Favorable intraoperative and postoperative care parameters for VATS lobectomy were observed in the late-experience group.
Kerr, David C R; DeGarmo, David S; Leve, Leslie D; Chamberlain, Patricia
2014-08-01
Multidimensional Treatment Foster Care (MTFC) has been found to reduce delinquency among girls in juvenile justice through 2-year follow-up. Given that such girls are at elevated risk for suicide and depression into adulthood, we tested MTFC effects on long-term trajectories of suicidal ideation and depressive symptoms. Girls (N = 166; mean [SD] age = 15.3 [1.2] years; 68% White) with a recent criminal referral who were mandated to out-of-home care were enrolled in 2 sequential cohorts. Girls were randomized to receive MTFC (n = 81) or group care (GC) treatment as usual (TAU; n = 85); the second MTFC cohort also received modules targeting substance use and risky sexual behavior. Depressive symptoms and suicidal ideation were assessed repeatedly through early adulthood (mean [SD] follow-up = 8.8 [2.9] years). Suicide attempt history was assessed in early adulthood. Girls assigned to MTFC showed significantly greater decreases in depressive symptoms across the long-term follow-up than GC girls (π = -.86, p < .05). Decreases in suicidal ideation rates were slightly stronger in MTFC than in GC as indicated by a marginal main effect (odds ratio [OR] = .92, p < .10) and a significant interaction that favored MTFC in the second cohort relative to the first (OR = .88, p < .01). There were no significant MTFC effects on suicide attempt. MTFC decreased depressive symptoms and suicidal thinking beyond the decreases attributable to time and TAU. Thus, MTFC has further impact on girls' lives than originally anticipated. PsycINFO Database Record (c) 2014 APA, all rights reserved.
Hiott, Deanna B; Phillips, Shannon; Amella, Elaine
2017-07-31
Adolescent risk-taking behavior choices can affect future health outcomes. The purpose of this integrative literature review is to evaluate adolescent risk screening instruments available to primary care providers in the United States using the Donabedian Framework of structure, process, and outcome. To examine the literature concerning multidimensional adolescent risk screening instruments available in the United States for use in the primary care setting, library searches, ancestry searches, and Internet searches were conducted. Library searches included a systematic search of the Cumulative Index to Nursing and Allied Health Literature (CINAHL), Academic Search Premier, Health Source Nursing Academic Ed, Medline, PsycINFO, the Psychology and Behavioral Sciences Collection, and PubMed databases with CINAHL headings using the following Boolean search terms: "primary care" and screening and pediatric. Criteria for inclusion consisted of studies conducted in the United States that involved broad multidimensional adolescent risk screening instruments for use in the pediatric primary care setting. Instruments that focused solely on one unhealthy behavior were excluded, as were developmental screens and screens not validated or designed for all ages of adolescents. In all 25 manuscripts reviewed, 16 screens met the inclusion criteria and were included in the study. These 16 screens were examined for factors associated with the Donabedian structure-process-outcome model. This review revealed that many screens contain structural issues related to cost and length that inhibit provider implementation in the primary care setting. Process limitations regarding the report method and administration format were also identified. The Pediatric Symptom Checklist was identified as a free, short tool that is valid and reliable.
Konrath, Sara; Falk, Emily; Fuhrel-Forbis, Andrea; Liu, Mary; Swain, James; Tolman, Richard; Cunningham, Rebecca; Walton, Maureen
2015-01-01
To what extent can simple mental exercises cause shifts in empathic habits? Can we use mobile technology to make people more empathic? It may depend on how empathy is measured. Scholars have identified a number of different facets and correlates of empathy. This study is among the first to take a comprehensive, multidimensional approach to empathy to determine how empathy training could affect these different facets and correlates. In doing so, we can learn more about empathy and its multifaceted nature. Participants (N = 90) were randomly assigned to receive either an empathy-building text message program (Text to Connect) or one of two control conditions (active versus passive). Respondents completed measures of dispositional empathy (i.e. self-perceptions of being an empathic person), affective empathy (i.e. motivations to help, immediate feelings of empathic concern), and prosocial behavior (i.e. self-reports and observer-reports) at baseline, and then again after the 14 day intervention period. We found that empathy-building messages increased affective indicators of empathy and prosocial behaviors, but actually decreased self-perceptions of empathy, relative to control messages. Although the brief text messaging intervention did not consistently impact empathy-related personality traits, it holds promise for the use of mobile technology for changing empathic motivations and behaviors.
Konrath, Sara; Falk, Emily; Fuhrel-Forbis, Andrea; Liu, Mary; Swain, James; Tolman, Richard; Cunningham, Rebecca; Walton, Maureen
2015-01-01
To what extent can simple mental exercises cause shifts in empathic habits? Can we use mobile technology to make people more empathic? It may depend on how empathy is measured. Scholars have identified a number of different facets and correlates of empathy. This study is among the first to take a comprehensive, multidimensional approach to empathy to determine how empathy training could affect these different facets and correlates. In doing so, we can learn more about empathy and its multifaceted nature. Participants (N = 90) were randomly assigned to receive either an empathy-building text message program (Text to Connect) or one of two control conditions (active versus passive). Respondents completed measures of dispositional empathy (i.e. self-perceptions of being an empathic person), affective empathy (i.e. motivations to help, immediate feelings of empathic concern), and prosocial behavior (i.e. self-reports and observer-reports) at baseline, and then again after the 14 day intervention period. We found that empathy-building messages increased affective indicators of empathy and prosocial behaviors, but actually decreased self-perceptions of empathy, relative to control messages. Although the brief text messaging intervention did not consistently impact empathy-related personality traits, it holds promise for the use of mobile technology for changing empathic motivations and behaviors. PMID:26356504
NASA Astrophysics Data System (ADS)
Appel, Marius; Lahn, Florian; Pebesma, Edzer; Buytaert, Wouter; Moulds, Simon
2016-04-01
Today's amount of freely available data requires scientists to spend large parts of their work on data management. This is especially true in environmental sciences when working with large remote sensing datasets, such as obtained from earth-observation satellites like the Sentinel fleet. Many frameworks like SpatialHadoop or Apache Spark address the scalability but target programmers rather than data analysts, and are not dedicated to imagery or array data. In this work, we use the open-source data management and analytics system SciDB to bring large earth-observation datasets closer to analysts. Its underlying data representation as multidimensional arrays fits naturally to earth-observation datasets, distributes storage and computational load over multiple instances by multidimensional chunking, and also enables efficient time-series based analyses, which is usually difficult using file- or tile-based approaches. Existing interfaces to R and Python furthermore allow for scalable analytics with relatively little learning effort. However, interfacing SciDB and file-based earth-observation datasets that come as tiled temporal snapshots requires a lot of manual bookkeeping during ingestion, and SciDB natively only supports loading data from CSV-like and custom binary formatted files, which currently limits its practical use in earth-observation analytics. To make it easier to work with large multi-temporal datasets in SciDB, we developed software tools that enrich SciDB with earth observation metadata and allow working with commonly used file formats: (i) the SciDB extension library scidb4geo simplifies working with spatiotemporal arrays by adding relevant metadata to the database and (ii) the Geospatial Data Abstraction Library (GDAL) driver implementation scidb4gdal allows to ingest and export remote sensing imagery from and to a large number of file formats. Using added metadata on temporal resolution and coverage, the GDAL driver supports time-based ingestion of imagery to existing multi-temporal SciDB arrays. While our SciDB plugin works directly in the database, the GDAL driver has been specifically developed using a minimum amount of external dependencies (i.e. CURL). Source code for both tools is available from github [1]. We present these tools in a case-study that demonstrates the ingestion of multi-temporal tiled earth-observation data to SciDB, followed by a time-series analysis using R and SciDBR. Through the exclusive use of open-source software, our approach supports reproducibility in scalable large-scale earth-observation analytics. In the future, these tools can be used in an automated way to let scientists only work on ready-to-use SciDB arrays to significantly reduce the data management workload for domain scientists. [1] https://github.com/mappl/scidb4geo} and \\url{https://github.com/mappl/scidb4gdal
Both texting and eating are associated with impaired simulated driving performance.
Alosco, Michael L; Spitznagel, Mary Beth; Fischer, Kimberly Hall; Miller, Lindsay A; Pillai, Vivek; Hughes, Joel; Gunstad, John
2012-09-01
Distracted driving is a known contributor to traffic accidents, and many states have banned texting while driving. However, little is known about the potential accident risk of other common activities while driving, such as eating. The objective of the current study was to examine the adverse impact of eating/drinking behavior relative to texting and nondistracted behaviors on a simulated driving task. A total of 186 participants were recruited from undergraduate psychology courses over 2 semesters at Kent State University. We utilized the Kent Multidimensional Assessment Driving Simulation (K-MADS) to compare simulated driving performance among participants randomly assigned to texting (N = 45), eating (N = 45), and control (N = 96) conditions. Multivariate analyses of variance (MANOVA) were conducted to examine between-group differences on simulated driving indices. MANOVA analyses indicated that groups differed in simulated driving performance, F(14, 366) = 7.70, P < .001. Both texting and eating produced impaired driving performance relative to controls, though these behaviors had approximately equal effect. Specifically, both texting and eating groups had more collisions, pedestrian strikes, and center line crossings than controls. In addition, the texting group had more road edge excursions than either eating or control participants and the eating group missed more stop signs than controls. These findings suggest that both texting and eating are associated with poorer simulated driving performance. Future work is needed to determine whether these findings generalize to real-world driving and the development of strategies to reduce distracted driving.
Integrating query of relational and textual data in clinical databases: a case study.
Fisk, John M; Mutalik, Pradeep; Levin, Forrest W; Erdos, Joseph; Taylor, Caroline; Nadkarni, Prakash
2003-01-01
The authors designed and implemented a clinical data mart composed of an integrated information retrieval (IR) and relational database management system (RDBMS). Using commodity software, which supports interactive, attribute-centric text and relational searches, the mart houses 2.8 million documents that span a five-year period and supports basic IR features such as Boolean searches, stemming, and proximity and fuzzy searching. Results are relevance-ranked using either "total documents per patient" or "report type weighting." Non-curated medical text has a significant degree of malformation with respect to spelling and punctuation, which creates difficulties for text indexing and searching. Presently, the IR facilities of RDBMS packages lack the features necessary to handle such malformed text adequately. A robust IR+RDBMS system can be developed, but it requires integrating RDBMSs with third-party IR software. RDBMS vendors need to make their IR offerings more accessible to non-programmers.
System, method and apparatus for generating phrases from a database
NASA Technical Reports Server (NTRS)
McGreevy, Michael W. (Inventor)
2004-01-01
A phrase generation is a method of generating sequences of terms, such as phrases, that may occur within a database of subsets containing sequences of terms, such as text. A database is provided and a relational model of the database is created. A query is then input. The query includes a term or a sequence of terms or multiple individual terms or multiple sequences of terms or combinations thereof. Next, several sequences of terms that are contextually related to the query are assembled from contextual relations in the model of the database. The sequences of terms are then sorted and output. Phrase generation can also be an iterative process used to produce sequences of terms from a relational model of a database.
Liu, Leigh Anne; Chua, Chei Hwee; Stahl, Günter K
2010-05-01
In an increasingly globalized workplace, the ability to communicate effectively across cultures is critical. We propose that the quality of communication experienced by individuals plays a significant role in the outcomes of intercultural interactions, such as cross-border negotiations. In 4 studies, we developed and validated a multidimensional conceptualization of quality of communication experience (QCE) and examined its consequences in intracultural versus intercultural business negotiations. We proposed and found 3 dimensions of QCE-namely, Clarity, Responsiveness, and Comfort. Findings from intercultural and same-cultural negotiations supported the hypotheses that QCE is lower in intercultural negotiation than in intracultural negotiation and that a higher degree of QCE leads to better negotiation outcomes. Moreover, we found evidence that the beneficial effects of higher QCE on negotiation outcomes are more pronounced in intercultural than in intracultural negotiations. We propose an agenda for future research and identify implications for practice. PsycINFO Database Record (c) 2010 APA, all rights reserved.
Data analysis of the COMPTEL instrument on the NASA gamma ray observatory
NASA Technical Reports Server (NTRS)
Diehl, R.; Bennett, K.; Collmar, W.; Connors, A.; Denherder, J. W.; Hermsen, W.; Lichti, G. G.; Lockwood, J. A.; Macri, J.; Mcconnell, M.
1992-01-01
The Compton imaging telescope (COMPTEL) on the Gamma Ray Observatory (GRO) is a wide field of view instrument. The coincidence measurement technique in two scintillation detector layers requires specific analysis methods. Straightforward event projection into the sky is impossible. Therefore, detector events are analyzed in a multi-dimensional dataspace using a gamma ray sky hypothesis convolved with the point spread function of the instrument in this dataspace. Background suppression and analysis techniques have important implications on the gamma ray source results for this background limited telescope. The COMPTEL collaboration applies a software system of analysis utilities, organized around a database management system. The use of this system for the assistance of guest investigators at the various collaboration sites and external sites is foreseen and allows different detail levels of cooperation with the COMPTEL institutes, dependent on the type of data to be studied.
Dissociative complexity: Antecedents and clinical correlates of a new construct.
Briere, John; Dietrich, Anne; Semple, Randye J
2016-09-01
To the extent that dissociation is a multidimensional phenomenon, and given a growing literature on complex posttraumatic outcomes, we hypothesized a new construct: dissociative complexity (DC). DC is defined as the tendency to simultaneously endorse multiple, relatively independent dissociative dimensions into their clinical ranges, and therefore represents the overall breadth or complexity of an individual's dissociative response. DC was evaluated in general population and prison participants using the Multiscale Dissociation Inventory (MDI; Briere, 2002). DC was higher among prisoners and women, and, as hypothesized, was associated with cumulative trauma and serious comorbidities (suicidality and substance abuse), even when controlling for generally elevated dissociation. DC appears to be a meaningful clinical construct that is phenomenologically and empirically different from a unidimensional index of dissociative severity. DC may serve as a clinical marker for multiple trauma exposures, complex dissociative outcomes, and risk of problematic comorbidities. (PsycINFO Database Record (c) 2016 APA, all rights reserved).
The BioIntelligence Framework: a new computational platform for biomedical knowledge computing.
Farley, Toni; Kiefer, Jeff; Lee, Preston; Von Hoff, Daniel; Trent, Jeffrey M; Colbourn, Charles; Mousses, Spyro
2013-01-01
Breakthroughs in molecular profiling technologies are enabling a new data-intensive approach to biomedical research, with the potential to revolutionize how we study, manage, and treat complex diseases. The next great challenge for clinical applications of these innovations will be to create scalable computational solutions for intelligently linking complex biomedical patient data to clinically actionable knowledge. Traditional database management systems (DBMS) are not well suited to representing complex syntactic and semantic relationships in unstructured biomedical information, introducing barriers to realizing such solutions. We propose a scalable computational framework for addressing this need, which leverages a hypergraph-based data model and query language that may be better suited for representing complex multi-lateral, multi-scalar, and multi-dimensional relationships. We also discuss how this framework can be used to create rapid learning knowledge base systems to intelligently capture and relate complex patient data to biomedical knowledge in order to automate the recovery of clinically actionable information.
Predicting adverse hemodynamic events in critically ill patients.
Yoon, Joo H; Pinsky, Michael R
2018-06-01
The art of predicting future hemodynamic instability in the critically ill has rapidly become a science with the advent of advanced analytical processed based on computer-driven machine learning techniques. How these methods have progressed beyond severity scoring systems to interface with decision-support is summarized. Data mining of large multidimensional clinical time-series databases using a variety of machine learning tools has led to our ability to identify alert artifact and filter it from bedside alarms, display real-time risk stratification at the bedside to aid in clinical decision-making and predict the subsequent development of cardiorespiratory insufficiency hours before these events occur. This fast evolving filed is primarily limited by linkage of high-quality granular to physiologic rationale across heterogeneous clinical care domains. Using advanced analytic tools to glean knowledge from clinical data streams is rapidly becoming a reality whose clinical impact potential is great.
Development of Web-Based Menu Planning Support System and its Solution Using Genetic Algorithm
NASA Astrophysics Data System (ADS)
Kashima, Tomoko; Matsumoto, Shimpei; Ishii, Hiroaki
2009-10-01
Recently lifestyle-related diseases have become an object of public concern, while at the same time people are being more health conscious. As an essential factor for causing the lifestyle-related diseases, we assume that the knowledge circulation on dietary habits is still insufficient. This paper focuses on everyday meals close to our life and proposes a well-balanced menu planning system as a preventive measure of lifestyle-related diseases. The system is developed by using a Web-based frontend and it provides multi-user services and menu information sharing capabilities like social networking services (SNS). The system is implemented on a Web server running Apache (HTTP server software), MySQL (database management system), and PHP (scripting language for dynamic Web pages). For the menu planning, a genetic algorithm is applied by understanding this problem as multidimensional 0-1 integer programming.
NASA Astrophysics Data System (ADS)
Nora, R.; Field, J. E.; Peterson, J. Luc; Spears, B.; Kruse, M.; Humbird, K.; Gaffney, J.; Springer, P. T.; Brandon, S.; Langer, S.
2017-10-01
We present an experimentally corroborated hydrodynamic extrapolation of several recent BigFoot implosions on the National Ignition Facility. An estimate on the value and error of the hydrodynamic scale necessary for ignition (for each individual BigFoot implosion) is found by hydrodynamically scaling a distribution of multi-dimensional HYDRA simulations whose outputs correspond to their experimental observables. The 11-parameter database of simulations, which include arbitrary drive asymmetries, dopant fractions, hydrodynamic scaling parameters, and surface perturbations due to surrogate tent and fill-tube engineering features, was computed on the TRINITY supercomputer at Los Alamos National Laboratory. This simple extrapolation is the first step in providing a rigorous calibration of our workflow to provide an accurate estimate of the efficacy of achieving ignition on the National Ignition Facility. This work performed under the auspices of the U.S. Department of Energy by Lawrence Livermore National Laboratory under Contract DE-AC52-07NA27344.
ArrayNinja: An Open Source Platform for Unified Planning and Analysis of Microarray Experiments.
Dickson, B M; Cornett, E M; Ramjan, Z; Rothbart, S B
2016-01-01
Microarray-based proteomic platforms have emerged as valuable tools for studying various aspects of protein function, particularly in the field of chromatin biochemistry. Microarray technology itself is largely unrestricted in regard to printable material and platform design, and efficient multidimensional optimization of assay parameters requires fluidity in the design and analysis of custom print layouts. This motivates the need for streamlined software infrastructure that facilitates the combined planning and analysis of custom microarray experiments. To this end, we have developed ArrayNinja as a portable, open source, and interactive application that unifies the planning and visualization of microarray experiments and provides maximum flexibility to end users. Array experiments can be planned, stored to a private database, and merged with the imaged results for a level of data interaction and centralization that is not currently attainable with available microarray informatics tools. © 2016 Elsevier Inc. All rights reserved.
[Towards understanding human ecology in nursing practice: a concept analysis].
Huynh, Truc; Alderson, Marie
2010-06-01
Human ecology is an umbrella concept encompassing several social, physical, and cultural elements existing in the individual's external environment. The pragmatic utility method was used to analyze the "human ecology" concept in order to ascertain the conceptual fit with nursing epistemology and to promote its use by nurses in clinical practice. Relevant articles for the review were retrieved from the MEDLINE, CINAHL, PsycINFO, and CSA databases using the terms "human ecology," "environment," "nursing," and "ecology." Data analysis revealed that human ecology is perceived as a theoretical perspective designating a complex, multilayered, and multidimensional system, one that comprises individuals and their reciprocal interactions with their global environments and the subsequent impact of these interactions upon their health. Human ecology preconditions include the individuals, their environments, and their transactions. Attributes of this concept encompass the characteristics of an open system (e.g., interdependence, reciprocal).
Relations between inductive reasoning and deductive reasoning.
Heit, Evan; Rotello, Caren M
2010-05-01
One of the most important open questions in reasoning research is how inductive reasoning and deductive reasoning are related. In an effort to address this question, we applied methods and concepts from memory research. We used 2 experiments to examine the effects of logical validity and premise-conclusion similarity on evaluation of arguments. Experiment 1 showed 2 dissociations: For a common set of arguments, deduction judgments were more affected by validity, and induction judgments were more affected by similarity. Moreover, Experiment 2 showed that fast deduction judgments were like induction judgments-in terms of being more influenced by similarity and less influenced by validity, compared with slow deduction judgments. These novel results pose challenges for a 1-process account of reasoning and are interpreted in terms of a 2-process account of reasoning, which was implemented as a multidimensional signal detection model and applied to receiver operating characteristic data. PsycINFO Database Record (c) 2010 APA, all rights reserved.
Central Schemes for Multi-Dimensional Hamilton-Jacobi Equations
NASA Technical Reports Server (NTRS)
Bryson, Steve; Levy, Doron; Biegel, Bryan (Technical Monitor)
2002-01-01
We present new, efficient central schemes for multi-dimensional Hamilton-Jacobi equations. These non-oscillatory, non-staggered schemes are first- and second-order accurate and are designed to scale well with an increasing dimension. Efficiency is obtained by carefully choosing the location of the evolution points and by using a one-dimensional projection step. First-and second-order accuracy is verified for a variety of multi-dimensional, convex and non-convex problems.
Full-Text Searching on Major Supermarket Systems: Dialog, Data-Star, and Nexis.
ERIC Educational Resources Information Center
Tenopir, Carol; Berglund, Sharon
1993-01-01
Examines the similarities, differences, and full-text features of the three most-used online systems for full-text searching in general libraries: DIALOG, Data-Star, and NEXIS. Overlapping databases, unique sources, search features, proximity operators, set building, language enhancement and word equivalencies, and display features are discussed.…
Clinical records anonymisation and text extraction (CRATE): an open-source software system.
Cardinal, Rudolf N
2017-04-26
Electronic medical records contain information of value for research, but contain identifiable and often highly sensitive confidential information. Patient-identifiable information cannot in general be shared outside clinical care teams without explicit consent, but anonymisation/de-identification allows research uses of clinical data without explicit consent. This article presents CRATE (Clinical Records Anonymisation and Text Extraction), an open-source software system with separable functions: (1) it anonymises or de-identifies arbitrary relational databases, with sensitivity and precision similar to previous comparable systems; (2) it uses public secure cryptographic methods to map patient identifiers to research identifiers (pseudonyms); (3) it connects relational databases to external tools for natural language processing; (4) it provides a web front end for research and administrative functions; and (5) it supports a specific model through which patients may consent to be contacted about research. Creation and management of a research database from sensitive clinical records with secure pseudonym generation, full-text indexing, and a consent-to-contact process is possible and practical using entirely free and open-source software.
The Desired Concept Maps and Goal Setting for Assessing Professionalism in Medicine.
Guraya, Salman Y; Guraya, Shaista S; Mahabbat, Nehal Anam; Fallatah, Khulood Yahya; Al-Ahmadi, Bashaer Ahmad; Alalawi, Hadeel Hadi
2016-05-01
Due to the multi-dimensional characteristics of professionalism, no single assessment modality has shown to reliably assess professionalism. This review aims to describe some of the popular assessment tools that are being used to assess professionalism with a view to formulate a framework of assessment of professionalism in medicine. In December 2015, the online research databases of MEDLINE, the Educational Resources Information Center (ERIC), Elton Bryson Stephens Company (EBSCO), SCOPUS, OVID and PsychINFO were searched for full-text English language articles published during 2000 to 2015. MeSH terms "professionalism" AND "duty" AND "assessment" OR "professionalism behavioural" AND "professionalism-cognitive" were used. The research articles that assessed professionalism across medical fields along with other areas of competencies were included. A final list of 35 articles were selected for this review. Several assessment tools are available for assessing professionalism that includes, but not limited to, mini clinical evaluation exercise, standardised direct observation of procedural skills, professionalism mini-evaluation exercise, multi-source feedback and 360 degree evaluation, and case based discussions. Because professionalism is a complex construct, it is less likely that a single assessment strategy will adequately measure it. Since every single assessment tool has its own weaknesses, triangulation involving multiple tools can compensate the shortcomings associated with any single approach. Assessment of professionalism necessitates a combination of modalities at individual, interpersonal, societal, and institutional levels and should be accompanied by feedback and motivational reflection that will, in turn, lead to behaviour and identity formation. The assessment of professionalism in medicine should meet the criteria of validity, reliability, feasibility and acceptability. Educators are urged to enhance the depth and quality of assessment instruments in the existing medical curricula for ensuring validity and reliability of assessment tools for professionalism.
Wright, Alexis A; Hegedus, Eric J; Tarara, Daniel T; Ray, Samantha C; Dischiavi, Steven L
2018-02-01
To produce a best evidence synthesis of exercise prescription used when treating shoulder pathology in the overhead athlete. A systematic review of exercises used in overhead athletes including case studies and clinical commentaries. MEDLINE, PubMed, SPORTDiscus and CINAHL from database inception through July 8, 2016. We examined data from randomised controlled trials and prospective cohort (level I-IV evidence) studies that addressed exercise intervention in the rehabilitation of the overhead athlete with shoulder pathology. Case studies and clinical commentaries (level V evidence) were examined to account for expert opinion-based research. Data were combined using best evidence synthesis and graded (A-F) recommendations (Centre for Evidence-Based Medicine). There were 33 unique exercises in six level I-IV studies that met our inclusion criteria. Most exercises were single-plane, upper extremity exercises performed below 90 o of elevation. There were 102 unique exercises in 33 level V studies that met our inclusion criteria. These exercises emphasised plyometrics, kinetic chain and sport-specific training. Overall, evidence for exercise interventions in overhead athletes with shoulder pathology is dominated by expert opinion (grade D). There is great variability between exercise approaches suggested by experts and those investigated in research studies and the overall level of evidence is low. The strongest available evidence (level B) supports the use of single-plane, open chain upper extremity exercises performed below 90° of elevation and closed chain upper extremity exercises. Clinical expert pieces support a more advanced, global treatment approach consistent with the complex, multidimensional nature of sport. © Article author(s) (or their employer(s) unless otherwise stated in the text of the article) 2018. All rights reserved. No commercial use is permitted unless otherwise expressly granted.
The Desired Concept Maps and Goal Setting for Assessing Professionalism in Medicine
Guraya, Shaista S.; Mahabbat, Nehal Anam; Fallatah, Khulood Yahya; Al-Ahmadi, Bashaer Ahmad; Alalawi, Hadeel Hadi
2016-01-01
Due to the multi-dimensional characteristics of professionalism, no single assessment modality has shown to reliably assess professionalism. This review aims to describe some of the popular assessment tools that are being used to assess professionalism with a view to formulate a framework of assessment of professionalism in medicine. In December 2015, the online research databases of MEDLINE, the Educational Resources Information Center (ERIC), Elton Bryson Stephens Company (EBSCO), SCOPUS, OVID and PsychINFO were searched for full-text English language articles published during 2000 to 2015. MeSH terms “professionalism” AND “duty” AND “assessment” OR “professionalism behavioural” AND “professionalism–cognitive” were used. The research articles that assessed professionalism across medical fields along with other areas of competencies were included. A final list of 35 articles were selected for this review. Several assessment tools are available for assessing professionalism that includes, but not limited to, mini clinical evaluation exercise, standardised direct observation of procedural skills, professionalism mini-evaluation exercise, multi-source feedback and 360 degree evaluation, and case based discussions. Because professionalism is a complex construct, it is less likely that a single assessment strategy will adequately measure it. Since every single assessment tool has its own weaknesses, triangulation involving multiple tools can compensate the shortcomings associated with any single approach. Assessment of professionalism necessitates a combination of modalities at individual, interpersonal, societal, and institutional levels and should be accompanied by feedback and motivational reflection that will, in turn, lead to behaviour and identity formation. The assessment of professionalism in medicine should meet the criteria of validity, reliability, feasibility and acceptability. Educators are urged to enhance the depth and quality of assessment instruments in the existing medical curricula for ensuring validity and reliability of assessment tools for professionalism. PMID:27437247
UKPMC: a full text article resource for the life sciences.
McEntyre, Johanna R; Ananiadou, Sophia; Andrews, Stephen; Black, William J; Boulderstone, Richard; Buttery, Paula; Chaplin, David; Chevuru, Sandeepreddy; Cobley, Norman; Coleman, Lee-Ann; Davey, Paul; Gupta, Bharti; Haji-Gholam, Lesley; Hawkins, Craig; Horne, Alan; Hubbard, Simon J; Kim, Jee-Hyub; Lewin, Ian; Lyte, Vic; MacIntyre, Ross; Mansoor, Sami; Mason, Linda; McNaught, John; Newbold, Elizabeth; Nobata, Chikashi; Ong, Ernest; Pillai, Sharmila; Rebholz-Schuhmann, Dietrich; Rosie, Heather; Rowbotham, Rob; Rupp, C J; Stoehr, Peter; Vaughan, Philip
2011-01-01
UK PubMed Central (UKPMC) is a full-text article database that extends the functionality of the original PubMed Central (PMC) repository. The UKPMC project was launched as the first 'mirror' site to PMC, which in analogy to the International Nucleotide Sequence Database Collaboration, aims to provide international preservation of the open and free-access biomedical literature. UKPMC (http://ukpmc.ac.uk) has undergone considerable development since its inception in 2007 and now includes both a UKPMC and PubMed search, as well as access to other records such as Agricola, Patents and recent biomedical theses. UKPMC also differs from PubMed/PMC in that the full text and abstract information can be searched in an integrated manner from one input box. Furthermore, UKPMC contains 'Cited By' information as an alternative way to navigate the literature and has incorporated text-mining approaches to semantically enrich content and integrate it with related database resources. Finally, UKPMC also offers added-value services (UKPMC+) that enable grantees to deposit manuscripts, link papers to grants, publish online portfolios and view citation information on their papers. Here we describe UKPMC and clarify the relationship between PMC and UKPMC, providing historical context and future directions, 10 years on from when PMC was first launched.
UKPMC: a full text article resource for the life sciences
McEntyre, Johanna R.; Ananiadou, Sophia; Andrews, Stephen; Black, William J.; Boulderstone, Richard; Buttery, Paula; Chaplin, David; Chevuru, Sandeepreddy; Cobley, Norman; Coleman, Lee-Ann; Davey, Paul; Gupta, Bharti; Haji-Gholam, Lesley; Hawkins, Craig; Horne, Alan; Hubbard, Simon J.; Kim, Jee-Hyub; Lewin, Ian; Lyte, Vic; MacIntyre, Ross; Mansoor, Sami; Mason, Linda; McNaught, John; Newbold, Elizabeth; Nobata, Chikashi; Ong, Ernest; Pillai, Sharmila; Rebholz-Schuhmann, Dietrich; Rosie, Heather; Rowbotham, Rob; Rupp, C. J.; Stoehr, Peter; Vaughan, Philip
2011-01-01
UK PubMed Central (UKPMC) is a full-text article database that extends the functionality of the original PubMed Central (PMC) repository. The UKPMC project was launched as the first ‘mirror’ site to PMC, which in analogy to the International Nucleotide Sequence Database Collaboration, aims to provide international preservation of the open and free-access biomedical literature. UKPMC (http://ukpmc.ac.uk) has undergone considerable development since its inception in 2007 and now includes both a UKPMC and PubMed search, as well as access to other records such as Agricola, Patents and recent biomedical theses. UKPMC also differs from PubMed/PMC in that the full text and abstract information can be searched in an integrated manner from one input box. Furthermore, UKPMC contains ‘Cited By’ information as an alternative way to navigate the literature and has incorporated text-mining approaches to semantically enrich content and integrate it with related database resources. Finally, UKPMC also offers added-value services (UKPMC+) that enable grantees to deposit manuscripts, link papers to grants, publish online portfolios and view citation information on their papers. Here we describe UKPMC and clarify the relationship between PMC and UKPMC, providing historical context and future directions, 10 years on from when PMC was first launched. PMID:21062818
Kathol, Nicole; Sgoutas-Emch, Sandra
2017-04-01
Developing interventions to address the problem of college drinking requires the identification of contributing factors to drinking behavior. It is believed that religion and spirituality (R/S) play a role, but the mechanism is unclear. Using a multi-dimensional R/S measure, an alcohol behavior inventory, and a religious affiliation proscription question, this study was designed to dive deeper into this connection. This study found that religious singing/chanting and reading sacred text were the best predictors of lower alcohol consumption. Furthermore, participants who perceive their religious tradition to be proscriptive reported less alcohol consumption and higher religious/spiritual profiles.
NASA scientific and technical information for the 1990s
NASA Technical Reports Server (NTRS)
Cotter, Gladys A.
1990-01-01
Projections for NASA scientific and technical information (STI) in the 1990s are outlined. NASA STI for the 1990s will maintain a quality bibliographic and full-text database, emphasizing electronic input and products supplemented by networked access to a wide variety of sources, particularly numeric databases.
The Missing Link: Context Loss in Online Databases
ERIC Educational Resources Information Center
Mi, Jia; Nesta, Frederick
2005-01-01
Full-text databases do not allow for the complexity of the interaction of the human eye and brain with printed matter. As a result, both content and context may be lost. The authors propose additional indexing fields that would maintain the content and context of print in electronic formats.
Redefining Information Access to Serials Information.
ERIC Educational Resources Information Center
Chen, Ching-chih
1992-01-01
Describes full-text document delivery services that have been introduced in conjunction with available databases in response to economic and technological changes affecting libraries: (1) CARL System's UnCover database and UnCover2 service; (2) Research Libraries Group's CitaDel delivery service; and (3) Faxon Research Service's Faxon Finder and…
ERIC Educational Resources Information Center
Miller-Whitehead, Marie
Keyword and text string searches of online library catalogs often provide different results according to library and database used and depending upon how books and journals are indexed. For this reason, online databases such as ERIC often provide tutorials and recommendations for searching their site, such as how to use Boolean search strategies.…
NASA Technical Reports Server (NTRS)
Bohnhoff-Hlavacek, Gail
1992-01-01
One of the objectives of the team supporting the LDEF Systems and Materials Special Investigative Groups is to develop databases of experimental findings. These databases identify the hardware flown, summarize results and conclusions, and provide a system for acknowledging investigators, tracing sources of data, and future design suggestions. To date, databases covering the optical experiments, and thermal control materials (chromic acid anodized aluminum, silverized Teflon blankets, and paints) have been developed at Boeing. We used the Filemaker Pro software, the database manager for the Macintosh computer produced by the Claris Corporation. It is a flat, text-retrievable database that provides access to the data via an intuitive user interface, without tedious programming. Though this software is available only for the Macintosh computer at this time, copies of the databases can be saved to a format that is readable on a personal computer as well. Further, the data can be exported to more powerful relational databases, capabilities, and use of the LDEF databases and describe how to get copies of the database for your own research.
Thakkar, Jay; Redfern, Julie; Khan, Ehsan; Atkins, Emily; Ha, Jeffrey; Vo, Kha; Thiagalingam, Aravinda; Chow, Clara K
2018-05-23
The 'Tobacco, Exercise and Diet Messages' (TEXT ME) study was a 6-month, single-centre randomised clinical trial (RCT) that found a text message support program improved levels of cardiovascular risk factors in patients with coronary heart disease (CHD). The current analyses examined whether receipt of text messages influenced participants' engagement with conventional healthcare resources. The TEXT ME study database (N=710) was linked with routinely collected health department databases. Number of doctor consultations, investigations and cardiac medication prescriptions in the two study groups were compared. The most frequently accessed health service was consultations with a General Practitioner (mean 7.1, s.d. 5.4). The numbers of medical consultations, biochemical tests or cardiac-specific investigations were similar between the study groups. There was at least one prescription registered for statin, ACEI/ARBs and β-blockers in 79, 66 and 50% of patients respectively, with similar refill rates in both the study groups. The study identified TEXT ME text messaging program did not increase use of Medicare Benefits Schedule (MBS) and Pharmaceutical Benefits Scheme (PBS) captured healthcare services. The observed benefits of TEXT ME reflect direct effects of intervention independent of conventional healthcare resource engagement.
Névéol, Aurélie; Wilbur, W John; Lu, Zhiyong
2012-01-01
High-throughput experiments and bioinformatics techniques are creating an exploding volume of data that are becoming overwhelming to keep track of for biologists and researchers who need to access, analyze and process existing data. Much of the available data are being deposited in specialized databases, such as the Gene Expression Omnibus (GEO) for microarrays or the Protein Data Bank (PDB) for protein structures and coordinates. Data sets are also being described by their authors in publications archived in literature databases such as MEDLINE and PubMed Central. Currently, the curation of links between biological databases and the literature mainly relies on manual labour, which makes it a time-consuming and daunting task. Herein, we analysed the current state of link curation between GEO, PDB and MEDLINE. We found that the link curation is heterogeneous depending on the sources and databases involved, and that overlap between sources is low, <50% for PDB and GEO. Furthermore, we showed that text-mining tools can automatically provide valuable evidence to help curators broaden the scope of articles and database entries that they review. As a result, we made recommendations to improve the coverage of curated links, as well as the consistency of information available from different databases while maintaining high-quality curation. Database URLs: http://www.ncbi.nlm.nih.gov/PubMed, http://www.ncbi.nlm.nih.gov/geo/, http://www.rcsb.org/pdb/
Measuring use patterns of online journals and databases
De Groote, Sandra L.; Dorsch, Josephine L.
2003-01-01
Purpose: This research sought to determine use of online biomedical journals and databases and to assess current user characteristics associated with the use of online resources in an academic health sciences center. Setting: The Library of the Health Sciences–Peoria is a regional site of the University of Illinois at Chicago (UIC) Library with 350 print journals, more than 4,000 online journals, and multiple online databases. Methodology: A survey was designed to assess online journal use, print journal use, database use, computer literacy levels, and other library user characteristics. A survey was sent through campus mail to all (471) UIC Peoria faculty, residents, and students. Results: Forty-one percent (188) of the surveys were returned. Ninety-eight percent of the students, faculty, and residents reported having convenient access to a computer connected to the Internet. While 53% of the users indicated they searched MEDLINE at least once a week, other databases showed much lower usage. Overall, 71% of respondents indicated a preference for online over print journals when possible. Conclusions: Users prefer online resources to print, and many choose to access these online resources remotely. Convenience and full-text availability appear to play roles in selecting online resources. The findings of this study suggest that databases without links to full text and online journal collections without links from bibliographic databases will have lower use. These findings have implications for collection development, promotion of library resources, and end-user training. PMID:12883574
Névéol, Aurélie; Wilbur, W. John; Lu, Zhiyong
2012-01-01
High-throughput experiments and bioinformatics techniques are creating an exploding volume of data that are becoming overwhelming to keep track of for biologists and researchers who need to access, analyze and process existing data. Much of the available data are being deposited in specialized databases, such as the Gene Expression Omnibus (GEO) for microarrays or the Protein Data Bank (PDB) for protein structures and coordinates. Data sets are also being described by their authors in publications archived in literature databases such as MEDLINE and PubMed Central. Currently, the curation of links between biological databases and the literature mainly relies on manual labour, which makes it a time-consuming and daunting task. Herein, we analysed the current state of link curation between GEO, PDB and MEDLINE. We found that the link curation is heterogeneous depending on the sources and databases involved, and that overlap between sources is low, <50% for PDB and GEO. Furthermore, we showed that text-mining tools can automatically provide valuable evidence to help curators broaden the scope of articles and database entries that they review. As a result, we made recommendations to improve the coverage of curated links, as well as the consistency of information available from different databases while maintaining high-quality curation. Database URLs: http://www.ncbi.nlm.nih.gov/PubMed, http://www.ncbi.nlm.nih.gov/geo/, http://www.rcsb.org/pdb/ PMID:22685160
Magnetic Helicity of Alfven Simple Waves
NASA Technical Reports Server (NTRS)
Webb, Gary M.; Hu, Q.; Dasgupta, B.; Zank, G. P.; Roberts, D.
2010-01-01
The magnetic helicity of fully nonlinear, multi-dimensional Alfven simple waves are investigated, by using relative helicity formulae and also by using an approach involving poloidal and toroidal decomposition of the magnetic field and magnetic vector potential. Different methods to calculate the magnetic vector potential are used, including the homotopy and Biot-Savart formulas. Two basic Alfven modes are identified: (a) the plane 1D Alfven simple wave given in standard texts, in which the Alfven wave propagates along the z-axis, with wave phase varphi=k_0(z-lambda t), where k_0 is the wave number and lambda is the group velocity of the wave, and (b)\\ the generalized Barnes (1976) simple Alfven wave in which the wave normal {bf n} moves in a circle in the xy-plane perpendicular to the mean field, which is directed along the z-axis. The plane Alfven wave (a) is analogous to the slab Alfven mode and the generalized Barnes solution (b) is analogous to the 2D mode in Alfvenic, incompressible turbulence. The helicity characteristics of these two basic Alfven modes are distinct. The helicity characteristics of more general multi-dimensional simple Alfven waves are also investigated. Applications to nonlinear Aifvenic fluctuations and structures observed in the solar wind are discussed.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Abbott, Jennifer; Sandberg, Tami
The Wind-Wildlife Impacts Literature Database (WILD), formerly known as the Avian Literature Database, was created in 1997. The goal of the database was to begin tracking the research that detailed the potential impact of wind energy development on birds. The Avian Literature Database was originally housed on a proprietary platform called Livelink ECM from Open- Text and maintained by in-house technical staff. The initial set of records was added by library staff. A vital part of the newly launched Drupal-based WILD database is the Bibliography module. Many of the resources included in the database have digital object identifiers (DOI). Themore » bibliographic information for any item that has a DOI can be imported into the database using this module. This greatly reduces the amount of manual data entry required to add records to the database. The content available in WILD is international in scope, which can be easily discerned by looking at the tags available in the browse menu.« less
US Army Research Laboratory Joint Interagency Field Experimentation 15-2 Final Report
2015-12-01
February 2015, at Alameda Island, California. Advanced text analytics capabilities were demonstrated in a logically coherent workflow pipeline that... text processing capabilities allowed the targeted use of a persistent imagery sensor for rapid detection of mission- critical events. The creation of...a very large text database from open source data provides a relevant and unclassified foundation for continued development of text -processing
Text-mining and information-retrieval services for molecular biology
Krallinger, Martin; Valencia, Alfonso
2005-01-01
Text-mining in molecular biology - defined as the automatic extraction of information about genes, proteins and their functional relationships from text documents - has emerged as a hybrid discipline on the edges of the fields of information science, bioinformatics and computational linguistics. A range of text-mining applications have been developed recently that will improve access to knowledge for biologists and database annotators. PMID:15998455
Encoding in the visual word form area: an fMRI adaptation study of words versus handwriting.
Barton, Jason J S; Fox, Christopher J; Sekunova, Alla; Iaria, Giuseppe
2010-08-01
Written texts are not just words but complex multidimensional stimuli, including aspects such as case, font, and handwriting style, for example. Neuropsychological reports suggest that left fusiform lesions can impair the reading of text for word (lexical) content, being associated with alexia, whereas right-sided lesions may impair handwriting recognition. We used fMRI adaptation in 13 healthy participants to determine if repetition-suppression occurred for words but not handwriting in the left visual word form area (VWFA) and the reverse in the right fusiform gyrus. Contrary to these expectations, we found adaptation for handwriting but not for words in both the left VWFA and the right VWFA homologue. A trend to adaptation for words but not handwriting was seen only in the left middle temporal gyrus. An analysis of anterior and posterior subdivisions of the left VWFA also failed to show any adaptation for words. We conclude that the right and the left fusiform gyri show similar patterns of adaptation for handwriting, consistent with a predominantly perceptual contribution to text processing.
Duncan, Morvwen; Moschopoulou, Elisavet; Herrington, Eldrid; Deane, Jennifer; Roylance, Rebecca; Jones, Louise; Bourke, Liam; Morgan, Adrienne; Chalder, Trudie; Thaha, Mohamed A; Taylor, Stephanie C; Korszun, Ania; White, Peter D; Bhui, Kamaldeep
2017-11-28
Over two million people in the UK are living with and beyond cancer. A third report diminished quality of life. A review of published systematic reviews to identify effective non-pharmacological interventions to improve the quality of life of cancer survivors. Databases searched until May 2017 included PubMed, Cochrane Central, EMBASE, MEDLINE, Web of Science, the Cumulative Index to Nursing and Allied Health Literature, and PsycINFO. Published systematic reviews of randomised trials of non-pharmacological interventions for people living with and beyond cancer were included; included reviews targeted patients aged over 18. All participants had already received a cancer diagnosis. Interventions located in any healthcare setting, home or online were included. Reviews of alternative therapies or those non-English reports were excluded. Two researchers independently assessed titles, abstracts and the full text of papers, and independently extracted the data. The primary outcome of interest was any measure of global (overall) quality of life. Quality assessment assessing methdological quality of systematic reviews (AMSTAR) and narrative synthesis, evaluating effectiveness of non-pharmacological interventions and their components. Of 14 430 unique titles, 21 were included in the review of reviews. There was little overlap in the primary papers across these reviews. Thirteen reviews covered mixed tumour groups, seven focused on breast cancer and one focused on prostate cancer. Face-to-face interventions were often combined with online, telephone and paper-based reading materials. Interventions included physical, psychological or behavioural, multidimensional rehabilitation and online approaches. Yoga specifically, physical exercise more generally, cognitive behavioural therapy (CBT) and mindfulness-based stress reduction (MBSR) programmes showed benefit in terms of quality of life. Exercise-based interventions were effective in the short (less than 3-8 months) and long term. CBT and MBSR also showed benefits, especially in the short term. The evidence for multidisciplinary, online and educational interventions was equivocal. © Article author(s) (or their employer(s) unless otherwise stated in the text of the article) 2017. All rights reserved. No commercial use is permitted unless otherwise expressly granted.
Varni, James W; Burwinkle, Tasha M; Katz, Ernest R; Meeske, Kathy; Dickinson, Paige
2002-04-01
The Pediatric Quality of Life Inventory (PedsQL) is a modular instrument designed to measure health-related quality of life (HRQOL) in children and adolescents ages 2-18 years. The PedsQL 4.0 Generic Core Scales are multidimensional child self-report and parent proxy-report scales developed as the generic core measure to be integrated with the PedsQL disease specific modules. The PedsQL Multidimensional Fatigue Scale was designed to measure fatigue in pediatric patients. The PedsQL 3.0 Cancer Module was designed to measure pediatric cancer specific HRQOL. The PedsQL Generic Core Scales, Multidimensional Fatigue Scale, and Cancer Module were administered to 339 families (220 child self-reports; 337 parent proxy-reports). Internal consistency reliability for the PedsQL Generic Core Total Scale Score (alpha = 0.88 child, 0.93 parent report), Multidimensional Fatigue Total Scale Score (alpha = 0.89 child, 0.92 parent report) and most Cancer Module Scales (average alpha = 0.72 child, 0.87 parent report) demonstrated reliability acceptable for group comparisons. Validity was demonstrated using the known-groups method. The PedsQL distinguished between healthy children and children with cancer as a group, and among children on-treatment versus off-treatment. The validity of the PedsQL Multidimensional Fatigue Scale was further demonstrated through hypothesized intercorrelations with dimensions of generic and cancer specific HRQOL. The results demonstrate the reliability and validity of the PedsQL Generic Core Scales, Multidimensional Fatigue Scale, and Cancer Module in pediatric cancer. The PedsQL may be utilized as an outcome measure in clinical trials, research, and clinical practice. Copyright 2002 American Cancer Society.
The PedsQL Multidimensional Fatigue Scale in pediatric rheumatology: reliability and validity.
Varni, James W; Burwinkle, Tasha M; Szer, Ilona S
2004-12-01
. The PedsQL (Pediatric Quality of Life Inventory) is a modular instrument designed to measure health related quality of life (HRQOL) in children and adolescents ages 2-18 years. The recently developed 18-item PedsQL Multidimensional Fatigue Scale was designed to measure fatigue in pediatric patients and comprises the General Fatigue Scale (6 items), Sleep/Rest Fatigue Scale (6 items), and Cognitive Fatigue Scale (6 items). The PedsQL 4.0 Generic Core Scales were developed as the generic core measure to be integrated with the PedsQL Disease-Specific Modules. The PedsQL 3.0 Rheumatology Module was designed to measure pediatric rheumatology-specific HRQOL. Methods. The PedsQL Multidimensional Fatigue Scale, Generic Core Scales, and Rheumatology Module were administered to 163 children and 154 parents (183 families accrued overall) recruited from a pediatric rheumatology clinic. Results. Internal consistency reliability for the PedsQL Multidimensional Fatigue Scale Total Score (a = 0.95 child, 0.95 parent report), General Fatigue Scale (a = 0.93 child, 0.92 parent), Sleep/Rest Fatigue Scale (a = 0.88 child, 0.90 parent), and Cognitive Fatigue Scale (a = 0.93 child, 0.96 parent) were excellent for group and individual comparisons. The validity of the PedsQL Multidimensional Fatigue Scale was confirmed through hypothesized intercorrelations with dimensions of generic and rheumatology-specific HRQOL. The PedsQL Multidimensional Fatigue Scale distinguished between healthy children and children with rheumatic diseases as a group, and was associated with greater disease severity. Children with fibromyalgia manifested greater fatigue than children with other rheumatic diseases. The results confirm the initial reliability and validity of the PedsQL Multidimensional Fatigue Scale in pediatric rheumatology.
NASA Astrophysics Data System (ADS)
He, Jiayi; Shang, Pengjian; Xiong, Hui
2018-06-01
Stocks, as the concrete manifestation of financial time series with plenty of potential information, are often used in the study of financial time series. In this paper, we utilize the stock data to recognize their patterns through out the dissimilarity matrix based on modified cross-sample entropy, then three-dimensional perceptual maps of the results are provided through multidimensional scaling method. Two modified multidimensional scaling methods are proposed in this paper, that is, multidimensional scaling based on Kronecker-delta cross-sample entropy (MDS-KCSE) and multidimensional scaling based on permutation cross-sample entropy (MDS-PCSE). These two methods use Kronecker-delta based cross-sample entropy and permutation based cross-sample entropy to replace the distance or dissimilarity measurement in classical multidimensional scaling (MDS). Multidimensional scaling based on Chebyshev distance (MDSC) is employed to provide a reference for comparisons. Our analysis reveals a clear clustering both in synthetic data and 18 indices from diverse stock markets. It implies that time series generated by the same model are easier to have similar irregularity than others, and the difference in the stock index, which is caused by the country or region and the different financial policies, can reflect the irregularity in the data. In the synthetic data experiments, not only the time series generated by different models can be distinguished, the one generated under different parameters of the same model can also be detected. In the financial data experiment, the stock indices are clearly divided into five groups. Through analysis, we find that they correspond to five regions, respectively, that is, Europe, North America, South America, Asian-Pacific (with the exception of mainland China), mainland China and Russia. The results also demonstrate that MDS-KCSE and MDS-PCSE provide more effective divisions in experiments than MDSC.
Multidimensional Risk Analysis: MRISK
NASA Technical Reports Server (NTRS)
McCollum, Raymond; Brown, Douglas; O'Shea, Sarah Beth; Reith, William; Rabulan, Jennifer; Melrose, Graeme
2015-01-01
Multidimensional Risk (MRISK) calculates the combined multidimensional score using Mahalanobis distance. MRISK accounts for covariance between consequence dimensions, which de-conflicts the interdependencies of consequence dimensions, providing a clearer depiction of risks. Additionally, in the event the dimensions are not correlated, Mahalanobis distance reduces to Euclidean distance normalized by the variance and, therefore, represents the most flexible and optimal method to combine dimensions. MRISK is currently being used in NASA's Environmentally Responsible Aviation (ERA) project o assess risk and prioritize scarce resources.
Callander, Emily J; Schofield, Deborah J
2017-02-01
Self-efficacy has numerous benefits for active and healthy aging, including giving the people the ability to make positive changes to their living standards and lifestyles. The present study aims to determine whether falling into multidimensional poverty lowers self-efficacy. Longitudinal analysis of waves 7-11 (2007-2011) of the nationally representative Household, Income and Labor Dynamics in Australia survey using linear regression models. The analysis focused on the Australian population aged 65 years and older. The Freedom Poverty Measure was used to identify those in multidimensional poverty. Those who fell into multidimensional poverty for 3 or 4 years between 2007 and 2011 had their self-efficacy scores decline by an average of 27 points (SD 21.2). Those who fell into poverty had significantly lower self-efficacy scores in 2011 - up to 57% lower (-66.6%, -45.7% P < 0.0001) after being in multidimensional poverty for 3 or 4 years between 2007 and 2011 than those who were never in poverty. Falling into multidimensional poverty lowers the self-efficacy scores of older people. In order to improve the chances of older people making long-term changes to improve their living standards, feelings of self-efficacy should first be assessed and improved. Geriatr Gerontol Int 2017; 17: 308-314. © 2015 Japan Geriatrics Society.
Signature Verification Based on Handwritten Text Recognition
NASA Astrophysics Data System (ADS)
Viriri, Serestina; Tapamo, Jules-R.
Signatures continue to be an important biometric trait because it remains widely used primarily for authenticating the identity of human beings. This paper presents an efficient text-based directional signature recognition algorithm which verifies signatures, even when they are composed of special unconstrained cursive characters which are superimposed and embellished. This algorithm extends the character-based signature verification technique. The experiments carried out on the GPDS signature database and an additional database created from signatures captured using the ePadInk tablet, show that the approach is effective and efficient, with a positive verification rate of 94.95%.
pGenN, a Gene Normalization Tool for Plant Genes and Proteins in Scientific Literature
Ding, Ruoyao; Arighi, Cecilia N.; Lee, Jung-Youn; Wu, Cathy H.; Vijay-Shanker, K.
2015-01-01
Background Automatically detecting gene/protein names in the literature and connecting them to databases records, also known as gene normalization, provides a means to structure the information buried in free-text literature. Gene normalization is critical for improving the coverage of annotation in the databases, and is an essential component of many text mining systems and database curation pipelines. Methods In this manuscript, we describe a gene normalization system specifically tailored for plant species, called pGenN (pivot-based Gene Normalization). The system consists of three steps: dictionary-based gene mention detection, species assignment, and intra species normalization. We have developed new heuristics to improve each of these phases. Results We evaluated the performance of pGenN on an in-house expertly annotated corpus consisting of 104 plant relevant abstracts. Our system achieved an F-value of 88.9% (Precision 90.9% and Recall 87.2%) on this corpus, outperforming state-of-art systems presented in BioCreative III. We have processed over 440,000 plant-related Medline abstracts using pGenN. The gene normalization results are stored in a local database for direct query from the pGenN web interface (proteininformationresource.org/pgenn/). The annotated literature corpus is also publicly available through the PIR text mining portal (proteininformationresource.org/iprolink/). PMID:26258475
NASA Astrophysics Data System (ADS)
Liu, Maoyuan; Jacob, Aurélie; Schmetterer, Clemens; Masset, Patrick J.; Hennet, Louis; Fischer, Henry E.; Kozaily, Jad; Jahn, Sandro; Gray-Weale, Angus
2016-04-01
Calcium aluminosilicate \\text{CaO}-\\text{A}{{\\text{l}}2}{{\\text{O}}3}-\\text{Si}{{\\text{O}}2} (CAS) melts with compositions {{≤ft(\\text{CaO}-\\text{Si}{{\\text{O}}2}\\right)}x}{{≤ft(\\text{A}{{\\text{l}}2}{{\\text{O}}3}\\right)}1-x} for x < 0.5 and {{≤ft(\\text{A}{{\\text{l}}2}{{\\text{O}}3}\\right)}x}{{≤ft(\\text{Si}{{\\text{O}}2}\\right)}1-x} for x≥slant 0.5 are studied using neutron diffraction with aerodynamic levitation and density functional theory molecular dynamics modelling. Simulated structure factors are found to be in good agreement with experimental structure factors. Local atomic structures from simulations reveal the role of calcium cations as a network modifier, and aluminium cations as a non-tetrahedral network former. Distributions of tetrahedral order show that an increasing concentration of the network former Al increases entropy, while an increasing concentration of the network modifier Ca decreases entropy. This trend is opposite to the conventional understanding that increasing amounts of network former should increase order in the network liquid, and so decrease entropy. The two-body correlation entropy S 2 is found to not correlate with the excess entropy values obtained from thermochemical databases, while entropies including higher-order correlations such as tetrahedral order, O-M-O or M-O-M bond angles and Q N environments show a clear linear correlation between computed entropy and database excess entropy. The possible relationship between atomic structures and excess entropy is discussed.
Further Validation of the Multidimensional Fatigue Symptom Inventory-Short Form
Stein, Kevin D.; Jacobsen, Paul B.; Blanchard, Chris M.; Thors, Christina
2008-01-01
A growing body of evidence is documenting the multidimensional nature of cancer-related fatigue. Although several multidimensional measures of fatigue have been developed, further validation of these scales is needed. To this end, the current study sought to evaluate the factorial and construct validity of the 30-item Multidimensional Fatigue Symptom Inventory-Short Form (MFSI-SF). A heterogeneous sample of 304 cancer patients (mean age 55 years) completed the MFSI-SF, along with several other measures of psychosocial functioning including the MOS-SF-36 and Fatigue Symptom Inventory, following the fourth cycle of chemotherapy treatment. The results of a confirmatory factor analysis indicated the 5-factor model provided a good fit to the data as evidenced by commonly used goodness of fit indices (CFI 0.90 and IFI 0.90). Additional evidence for the validity of the MFSI-SF was provided via correlations with other relevant instruments (range −0.21 to 0.82). In sum, the current study provides support for the MFSI-SF as a valuable tool for the multidimensional assessment of cancer-related fatigue. PMID:14711465
Description of 'REQUEST-KYUSHYU' for KYUKEICHO regional data base
NASA Astrophysics Data System (ADS)
Takimoto, Shin'ichi
Kyushu Economic Research Association (a foundational juridical person) initiated the regional database services, ' REQUEST-Kyushu ' recently. It is the full scale databases compiled based on the information and know-hows which the Association has accumulated over forty years. It covers the regional information database for journal and newspaper articles, and statistical information database for economic statistics. As to the former database it is searched on a personal computer and then a search result (original text) is sent through a facsimile. As to the latter, it is also searched on a personal computer where the data is processed, edited or downloaded. This paper describes characteristics, content and the system outline of 'REQUEST-Kyushu'.
Ambiguity and variability of database and software names in bioinformatics.
Duck, Geraint; Kovacevic, Aleksandar; Robertson, David L; Stevens, Robert; Nenadic, Goran
2015-01-01
There are numerous options available to achieve various tasks in bioinformatics, but until recently, there were no tools that could systematically identify mentions of databases and tools within the literature. In this paper we explore the variability and ambiguity of database and software name mentions and compare dictionary and machine learning approaches to their identification. Through the development and analysis of a corpus of 60 full-text documents manually annotated at the mention level, we report high variability and ambiguity in database and software mentions. On a test set of 25 full-text documents, a baseline dictionary look-up achieved an F-score of 46 %, highlighting not only variability and ambiguity but also the extensive number of new resources introduced. A machine learning approach achieved an F-score of 63 % (with precision of 74 %) and 70 % (with precision of 83 %) for strict and lenient matching respectively. We characterise the issues with various mention types and propose potential ways of capturing additional database and software mentions in the literature. Our analyses show that identification of mentions of databases and tools is a challenging task that cannot be achieved by relying on current manually-curated resource repositories. Although machine learning shows improvement and promise (primarily in precision), more contextual information needs to be taken into account to achieve a good degree of accuracy.
NASA Astrophysics Data System (ADS)
Aksenov, A. G.; Chechetkin, V. M.
2018-04-01
Most of the energy released in the gravitational collapse of the cores of massive stars is carried away by neutrinos. Neutrinos play a pivotal role in explaining core-collape supernovae. Currently, mathematical models of the gravitational collapse are based on multi-dimensional gas dynamics and thermonuclear reactions, while neutrino transport is considered in a simplified way. Multidimensional gas dynamics is used with neutrino transport in the flux-limited diffusion approximation to study the role of multi-dimensional effects. The possibility of large-scale convection is discussed, which is interesting both for explaining SN II and for setting up observations to register possible high-energy (≳10MeV) neutrinos from the supernova. A new multi-dimensional, multi-temperature gas dynamics method with neutrino transport is presented.
Wallot, Sebastian; Roepstorff, Andreas; Mønster, Dan
2016-01-01
We introduce Multidimensional Recurrence Quantification Analysis (MdRQA) as a tool to analyze multidimensional time-series data. We show how MdRQA can be used to capture the dynamics of high-dimensional signals, and how MdRQA can be used to assess coupling between two or more variables. In particular, we describe applications of the method in research on joint and collective action, as it provides a coherent analysis framework to systematically investigate dynamics at different group levels—from individual dynamics, to dyadic dynamics, up to global group-level of arbitrary size. The Appendix in Supplementary Material contains a software implementation in MATLAB to calculate MdRQA measures. PMID:27920748
NASA Astrophysics Data System (ADS)
Huang, Wen-Min; Mou, Chung-Yu; Chang, Cheng-Hung
2010-02-01
While the scattering phase for several one-dimensional potentials can be exactly derived, less is known in multi-dimensional quantum systems. This work provides a method to extend the one-dimensional phase knowledge to multi-dimensional quantization rules. The extension is illustrated in the example of Bogomolny's transfer operator method applied in two quantum wells bounded by step potentials of different heights. This generalized semiclassical method accurately determines the energy spectrum of the systems, which indicates the substantial role of the proposed phase correction. Theoretically, the result can be extended to other semiclassical methods, such as Gutzwiller trace formula, dynamical zeta functions, and semiclassical Landauer-Büttiker formula. In practice, this recipe enhances the applicability of semiclassical methods to multi-dimensional quantum systems bounded by general soft potentials.
Wallot, Sebastian; Roepstorff, Andreas; Mønster, Dan
2016-01-01
We introduce Multidimensional Recurrence Quantification Analysis (MdRQA) as a tool to analyze multidimensional time-series data. We show how MdRQA can be used to capture the dynamics of high-dimensional signals, and how MdRQA can be used to assess coupling between two or more variables. In particular, we describe applications of the method in research on joint and collective action, as it provides a coherent analysis framework to systematically investigate dynamics at different group levels-from individual dynamics, to dyadic dynamics, up to global group-level of arbitrary size. The Appendix in Supplementary Material contains a software implementation in MATLAB to calculate MdRQA measures.
Multi-dimensional simulations of core-collapse supernova explosions with CHIMERA
NASA Astrophysics Data System (ADS)
Messer, O. E. B.; Harris, J. A.; Hix, W. R.; Lentz, E. J.; Bruenn, S. W.; Mezzacappa, A.
2018-04-01
Unraveling the core-collapse supernova (CCSN) mechanism is a problem that remains essentially unsolved despite more than four decades of effort. Spherically symmetric models with otherwise high physical fidelity generally fail to produce explosions, and it is widely accepted that CCSNe are inherently multi-dimensional. Progress in realistic modeling has occurred recently through the availability of petascale platforms and the increasing sophistication of supernova codes. We will discuss our most recent work on understanding neutrino-driven CCSN explosions employing multi-dimensional neutrino-radiation hydrodynamics simulations with the Chimera code. We discuss the inputs and resulting outputs from these simulations, the role of neutrino radiation transport, and the importance of multi-dimensional fluid flows in shaping the explosions. We also highlight the production of 48Ca in long-running Chimera simulations.
Evaluation and comparison of bioinformatic tools for the enrichment analysis of metabolomics data.
Marco-Ramell, Anna; Palau-Rodriguez, Magali; Alay, Ania; Tulipani, Sara; Urpi-Sarda, Mireia; Sanchez-Pla, Alex; Andres-Lacueva, Cristina
2018-01-02
Bioinformatic tools for the enrichment of 'omics' datasets facilitate interpretation and understanding of data. To date few are suitable for metabolomics datasets. The main objective of this work is to give a critical overview, for the first time, of the performance of these tools. To that aim, datasets from metabolomic repositories were selected and enriched data were created. Both types of data were analysed with these tools and outputs were thoroughly examined. An exploratory multivariate analysis of the most used tools for the enrichment of metabolite sets, based on a non-metric multidimensional scaling (NMDS) of Jaccard's distances, was performed and mirrored their diversity. Codes (identifiers) of the metabolites of the datasets were searched in different metabolite databases (HMDB, KEGG, PubChem, ChEBI, BioCyc/HumanCyc, LipidMAPS, ChemSpider, METLIN and Recon2). The databases that presented more identifiers of the metabolites of the dataset were PubChem, followed by METLIN and ChEBI. However, these databases had duplicated entries and might present false positives. The performance of over-representation analysis (ORA) tools, including BioCyc/HumanCyc, ConsensusPathDB, IMPaLA, MBRole, MetaboAnalyst, Metabox, MetExplore, MPEA, PathVisio and Reactome and the mapping tool KEGGREST, was examined. Results were mostly consistent among tools and between real and enriched data despite the variability of the tools. Nevertheless, a few controversial results such as differences in the total number of metabolites were also found. Disease-based enrichment analyses were also assessed, but they were not found to be accurate probably due to the fact that metabolite disease sets are not up-to-date and the difficulty of predicting diseases from a list of metabolites. We have extensively reviewed the state-of-the-art of the available range of tools for metabolomic datasets, the completeness of metabolite databases, the performance of ORA methods and disease-based analyses. Despite the variability of the tools, they provided consistent results independent of their analytic approach. However, more work on the completeness of metabolite and pathway databases is required, which strongly affects the accuracy of enrichment analyses. Improvements will be translated into more accurate and global insights of the metabolome.
Multidimensional Programming Methods for Energy Facility Siting: Alternative Approaches
NASA Technical Reports Server (NTRS)
Solomon, B. D.; Haynes, K. E.
1982-01-01
The use of multidimensional optimization methods in solving power plant siting problems, which are characterized by several conflicting, noncommensurable objectives is addressed. After a discussion of data requirements and exclusionary site screening methods for bounding the decision space, classes of multiobjective and goal programming models are discussed in the context of finite site selection. Advantages and limitations of these approaches are highlighted and the linkage of multidimensional methods with the subjective, behavioral components of the power plant siting process is emphasized.
Optimal multi-dimensional poverty lines: The state of poverty in Iraq
NASA Astrophysics Data System (ADS)
Ameen, Jamal R. M.
2017-09-01
Poverty estimation based on calories intake is unrealistic. The established concept of multidimensional poverty has methodological weaknesses in the treatment of different dimensions and there is disagreement in methods of combining them into a single poverty line. This paper introduces a methodology to estimate optimal multidimensional poverty lines and uses the Iraqi household socio-economic survey data of 2012 to demonstrate the idea. The optimal poverty line for Iraq is found to be 170.5 Thousand Iraqi Dinars (TID).
Spectral signature verification using statistical analysis and text mining
NASA Astrophysics Data System (ADS)
DeCoster, Mallory E.; Firpi, Alexe H.; Jacobs, Samantha K.; Cone, Shelli R.; Tzeng, Nigel H.; Rodriguez, Benjamin M.
2016-05-01
In the spectral science community, numerous spectral signatures are stored in databases representative of many sample materials collected from a variety of spectrometers and spectroscopists. Due to the variety and variability of the spectra that comprise many spectral databases, it is necessary to establish a metric for validating the quality of spectral signatures. This has been an area of great discussion and debate in the spectral science community. This paper discusses a method that independently validates two different aspects of a spectral signature to arrive at a final qualitative assessment; the textual meta-data and numerical spectral data. Results associated with the spectral data stored in the Signature Database1 (SigDB) are proposed. The numerical data comprising a sample material's spectrum is validated based on statistical properties derived from an ideal population set. The quality of the test spectrum is ranked based on a spectral angle mapper (SAM) comparison to the mean spectrum derived from the population set. Additionally, the contextual data of a test spectrum is qualitatively analyzed using lexical analysis text mining. This technique analyzes to understand the syntax of the meta-data to provide local learning patterns and trends within the spectral data, indicative of the test spectrum's quality. Text mining applications have successfully been implemented for security2 (text encryption/decryption), biomedical3 , and marketing4 applications. The text mining lexical analysis algorithm is trained on the meta-data patterns of a subset of high and low quality spectra, in order to have a model to apply to the entire SigDB data set. The statistical and textual methods combine to assess the quality of a test spectrum existing in a database without the need of an expert user. This method has been compared to other validation methods accepted by the spectral science community, and has provided promising results when a baseline spectral signature is present for comparison. The spectral validation method proposed is described from a practical application and analytical perspective.
Medical Language Processing for Knowledge Representation and Retrievals
Lyman, Margaret; Sager, Naomi; Chi, Emile C.; Tick, Leo J.; Nhan, Ngo Thanh; Su, Yun; Borst, Francois; Scherrer, Jean-Raoul
1989-01-01
The Linguistic String Project-Medical Language Processor, a system for computer analysis of narrative patient documents in English, is being adapted for French Lettres de Sortie. The system converts the free-text input to a semantic representation which is then mapped into a relational database. Retrievals of clinical data from the database are described.
NASA Technical Reports Server (NTRS)
Grams, R. R.
1982-01-01
A system designed to access a large range of available medical textbook information in an online interactive fashion is described. A high level query type database manager, INQUIRE, is used. Operating instructions, system flow diagrams, database descriptions, text generation, and error messages are discussed. User information is provided.
Fast neutron mutants database and web displays at SoyBase
USDA-ARS?s Scientific Manuscript database
SoyBase, the USDA-ARS soybean genetics and genomics database, has been expanded to include data for the fast neutron mutants produced by Bolon, Vance, et al. In addition to the expected text and sequence homology searches and visualization of the indels in the context of the genome sequence viewer, ...
Extending the Online Public Access Catalog into the Microcomputer Environment.
ERIC Educational Resources Information Center
Sutton, Brett
1990-01-01
Describes PCBIS, a database program for MS-DOS microcomputers that features a utility for automatically converting online public access catalog search results stored as text files into structured database files that can be searched, sorted, edited, and printed. Topics covered include the general features of the program, record structure, record…
Semi-Automated Annotation of Biobank Data Using Standard Medical Terminologies in a Graph Database.
Hofer, Philipp; Neururer, Sabrina; Goebel, Georg
2016-01-01
Data describing biobank resources frequently contains unstructured free-text information or insufficient coding standards. (Bio-) medical ontologies like Orphanet Rare Diseases Ontology (ORDO) or the Human Disease Ontology (DOID) provide a high number of concepts, synonyms and entity relationship properties. Such standard terminologies increase quality and granularity of input data by adding comprehensive semantic background knowledge from validated entity relationships. Moreover, cross-references between terminology concepts facilitate data integration across databases using different coding standards. In order to encourage the use of standard terminologies, our aim is to identify and link relevant concepts with free-text diagnosis inputs within a biobank registry. Relevant concepts are selected automatically by lexical matching and SPARQL queries against a RDF triplestore. To ensure correctness of annotations, proposed concepts have to be confirmed by medical data administration experts before they are entered into the registry database. Relevant (bio-) medical terminologies describing diseases and phenotypes were identified and stored in a graph database which was tied to a local biobank registry. Concept recommendations during data input trigger a structured description of medical data and facilitate data linkage between heterogeneous systems.
Development of Human Face Literature Database Using Text Mining Approach: Phase I.
Kaur, Paramjit; Krishan, Kewal; Sharma, Suresh K
2018-06-01
The face is an important part of the human body by which an individual communicates in the society. Its importance can be highlighted by the fact that a person deprived of face cannot sustain in the living world. The amount of experiments being performed and the number of research papers being published under the domain of human face have surged in the past few decades. Several scientific disciplines, which are conducting research on human face include: Medical Science, Anthropology, Information Technology (Biometrics, Robotics, and Artificial Intelligence, etc.), Psychology, Forensic Science, Neuroscience, etc. This alarms the need of collecting and managing the data concerning human face so that the public and free access of it can be provided to the scientific community. This can be attained by developing databases and tools on human face using bioinformatics approach. The current research emphasizes on creating a database concerning literature data of human face. The database can be accessed on the basis of specific keywords, journal name, date of publication, author's name, etc. The collected research papers will be stored in the form of a database. Hence, the database will be beneficial to the research community as the comprehensive information dedicated to the human face could be found at one place. The information related to facial morphologic features, facial disorders, facial asymmetry, facial abnormalities, and many other parameters can be extracted from this database. The front end has been developed using Hyper Text Mark-up Language and Cascading Style Sheets. The back end has been developed using hypertext preprocessor (PHP). The JAVA Script has used as scripting language. MySQL (Structured Query Language) is used for database development as it is most widely used Relational Database Management System. XAMPP (X (cross platform), Apache, MySQL, PHP, Perl) open source web application software has been used as the server.The database is still under the developmental phase and discusses the initial steps of its creation. The current paper throws light on the work done till date.
Valkhoff, Vera E; Coloma, Preciosa M; Masclee, Gwen M C; Gini, Rosa; Innocenti, Francesco; Lapi, Francesco; Molokhia, Mariam; Mosseveld, Mees; Nielsson, Malene Schou; Schuemie, Martijn; Thiessard, Frantz; van der Lei, Johan; Sturkenboom, Miriam C J M; Trifirò, Gianluca
2014-08-01
To evaluate the accuracy of disease codes and free text in identifying upper gastrointestinal bleeding (UGIB) from electronic health-care records (EHRs). We conducted a validation study in four European electronic health-care record (EHR) databases such as Integrated Primary Care Information (IPCI), Health Search/CSD Patient Database (HSD), ARS, and Aarhus, in which we identified UGIB cases using free text or disease codes: (1) International Classification of Disease (ICD)-9 (HSD, ARS); (2) ICD-10 (Aarhus); and (3) International Classification of Primary Care (ICPC) (IPCI). From each database, we randomly selected and manually reviewed 200 cases to calculate positive predictive values (PPVs). We employed different case definitions to assess the effect of outcome misclassification on estimation of risk of drug-related UGIB. PPV was 22% [95% confidence interval (CI): 16, 28] and 21% (95% CI: 16, 28) in IPCI for free text and ICPC codes, respectively. PPV was 91% (95% CI: 86, 95) for ICD-9 codes and 47% (95% CI: 35, 59) for free text in HSD. PPV for ICD-9 codes in ARS was 72% (95% CI: 65, 78) and 77% (95% CI: 69, 83) for ICD-10 codes (Aarhus). More specific definitions did not have significant impact on risk estimation of drug-related UGIB, except for wider CIs. ICD-9-CM and ICD-10 disease codes have good PPV in identifying UGIB from EHR; less granular terminology (ICPC) may require additional strategies. Use of more specific UGIB definitions affects precision, but not magnitude, of risk estimates. Copyright © 2014 Elsevier Inc. All rights reserved.
Thakar, Sambhaji B; Ghorpade, Pradnya N; Kale, Manisha V; Sonawane, Kailas D
2015-01-01
Fern plants are known for their ethnomedicinal applications. Huge amount of fern medicinal plants information is scattered in the form of text. Hence, database development would be an appropriate endeavor to cope with the situation. So by looking at the importance of medicinally useful fern plants, we developed a web based database which contains information about several group of ferns, their medicinal uses, chemical constituents as well as protein/enzyme sequences isolated from different fern plants. Fern ethnomedicinal plant database is an all-embracing, content management web-based database system, used to retrieve collection of factual knowledge related to the ethnomedicinal fern species. Most of the protein/enzyme sequences have been extracted from NCBI Protein sequence database. The fern species, family name, identification, taxonomy ID from NCBI, geographical occurrence, trial for, plant parts used, ethnomedicinal importance, morphological characteristics, collected from various scientific literatures and journals available in the text form. NCBI's BLAST, InterPro, phylogeny, Clustal W web source has also been provided for the future comparative studies. So users can get information related to fern plants and their medicinal applications at one place. This Fern ethnomedicinal plant database includes information of 100 fern medicinal species. This web based database would be an advantageous to derive information specifically for computational drug discovery, botanists or botanical interested persons, pharmacologists, researchers, biochemists, plant biotechnologists, ayurvedic practitioners, doctors/pharmacists, traditional medicinal users, farmers, agricultural students and teachers from universities as well as colleges and finally fern plant lovers. This effort would be useful to provide essential knowledge for the users about the adventitious applications for drug discovery, applications, conservation of fern species around the world and finally to create social awareness.
Architecture for biomedical multimedia information delivery on the World Wide Web
NASA Astrophysics Data System (ADS)
Long, L. Rodney; Goh, Gin-Hua; Neve, Leif; Thoma, George R.
1997-10-01
Research engineers at the National Library of Medicine are building a prototype system for the delivery of multimedia biomedical information on the World Wide Web. This paper discuses the architecture and design considerations for the system, which will be used initially to make images and text from the third National Health and Nutrition Examination Survey (NHANES) publicly available. We categorized our analysis as follows: (1) fundamental software tools: we analyzed trade-offs among use of conventional HTML/CGI, X Window Broadway, and Java; (2) image delivery: we examined the use of unconventional TCP transmission methods; (3) database manager and database design: we discuss the capabilities and planned use of the Informix object-relational database manager and the planned schema for the HNANES database; (4) storage requirements for our Sun server; (5) user interface considerations; (6) the compatibility of the system with other standard research and analysis tools; (7) image display: we discuss considerations for consistent image display for end users. Finally, we discuss the scalability of the system in terms of incorporating larger or more databases of similar data, and the extendibility of the system for supporting content-based retrieval of biomedical images. The system prototype is called the Web-based Medical Information Retrieval System. An early version was built as a Java applet and tested on Unix, PC, and Macintosh platforms. This prototype used the MiniSQL database manager to do text queries on a small database of records of participants in the second NHANES survey. The full records and associated x-ray images were retrievable and displayable on a standard Web browser. A second version has now been built, also a Java applet, using the MySQL database manager.
GRBase, a new gene regulation data base available by anonymous ftp.
Collier, B; Danielsen, M
1994-01-01
The Gene Regulation Database (GRBase) is a compendium of information on the structure and function of proteins involved in the control of gene expression in eukaryotes. These proteins include transcription factors, proteins involved in signal transduction, and receptors. The database can be obtained by FTP in Filemaker Pro, text, and postscript formats. The database will be expanded in the coming year to include reviews on families of proteins involved in gene regulation and to allow online searching. PMID:7937071
Staradmin -- Starlink User Database Maintainer
NASA Astrophysics Data System (ADS)
Fish, Adrian
The subject of this SSN is a utility called STARADMIN. This utility allows the system administrator to build and maintain a Starlink User Database (UDB). The principal source of information for each user is a text file, named after their username. The content of each file is a list consisting of one keyword followed by the relevant user data per line. These user database files reside in a single directory. The STARADMIN program is used to manipulate these user data files and automatically generate user summary lists.
Creating a VAPEPS database: A VAPEPS tutorial
NASA Technical Reports Server (NTRS)
Graves, George
1989-01-01
A procedural method is outlined for creating a Vibroacoustic Payload Environment Prediction System (VAPEPS) Database. The method of presentation employs flowcharts of sequential VAPEPS Commands used to create a VAPEPS Database. The commands are accompanied by explanatory text to the right of the command in order to minimize the need for repetitive reference to the VAPEPS user's manual. The method is demonstrated by examples of varying complexity. It is assumed that the reader has acquired a basic knowledge of the VAPEPS software program.
Selecting Full-Text Undergraduate Periodicals Databases.
ERIC Educational Resources Information Center
Still, Julie M.; Kassabian, Vibiana
1999-01-01
Examines how libraries and librarians can compare full-text general periodical indices, using ProQuest Direct, Periodical Abstracts (via Ovid), and EBSCOhost as examples. Explores breadth and depth of coverage; manipulation of results (email/download/print); ease of use (searching); and indexing quirks. (AEF)
Multi-dimensional simulations of core-collapse supernova explosions with CHIMERA
DOE Office of Scientific and Technical Information (OSTI.GOV)
Messer, Bronson; Harris, James Austin; Hix, William Raphael
Unraveling the core-collapse supernova (CCSN) mechanism is a problem that remains essentially unsolved despite more than four decades of effort. Spherically symmetric models with otherwise high physical fidelity generally fail to produce explosions, and it is widely accepted that CCSNe are inherently multi-dimensional. Progress in realistic modeling has occurred recently through the availability of petascale platforms and the increasing sophistication of supernova codes. We will discuss our most recent work on understanding neutrino-driven CCSN explosions employing multi-dimensional neutrino-radiation hydrodynamics simulations with the Chimera code. We discuss the inputs and resulting outputs from these simulations, the role of neutrino radiation transport,more » and the importance of multi-dimensional fluid flows in shaping the explosions. We also highlight the production of 48Ca in long-running Chimera simulations.« less
Almeida, Fernando R.; Brayner, Angelo; Rodrigues, Joel J. P. C.; Maia, Jose E. Bessa
2017-01-01
An efficient strategy for reducing message transmission in a wireless sensor network (WSN) is to group sensors by means of an abstraction denoted cluster. The key idea behind the cluster formation process is to identify a set of sensors whose sensed values present some data correlation. Nowadays, sensors are able to simultaneously sense multiple different physical phenomena, yielding in this way multidimensional data. This paper presents three methods for clustering sensors in WSNs whose sensors collect multidimensional data. The proposed approaches implement the concept of multidimensional behavioral clustering. To show the benefits introduced by the proposed methods, a prototype has been implemented and experiments have been carried out on real data. The results prove that the proposed methods decrease the amount of data flowing in the network and present low root-mean-square error (RMSE). PMID:28590450
The Role of a Multidimensional Concept of Trust in the Performance of Global Virtual Teams
NASA Technical Reports Server (NTRS)
Bodensteiner, Nan Muir; Stecklein, Jonette M.
2002-01-01
This paper focuses on the concept of trust as an important ingredient of effective global virtual team performance. Definitions of trust and virtual teams are presented. The concept of trust is developed from its unilateral application (trust, absence of trust) to a multidimensional concept including cognitive and affective components. The special challenges of a virtual team are then discussed with particular emphasis on how a multidimensional concept of trust impacts these challenges. Propositions suggesting the multidimensional concept of trust moderates the negative impacts of distance, cross cultural and organizational differences, the effects of electronically mediated communication, reluctance to share information and a lack of hi story/future on the performance of virtual teams are stated. The paper concludes with recommendations and a set of techniques to build both cognitive and affective trust in virtual teams.
Almeida, Fernando R; Brayner, Angelo; Rodrigues, Joel J P C; Maia, Jose E Bessa
2017-06-07
An efficient strategy for reducing message transmission in a wireless sensor network (WSN) is to group sensors by means of an abstraction denoted cluster. The key idea behind the cluster formation process is to identify a set of sensors whose sensed values present some data correlation. Nowadays, sensors are able to simultaneously sense multiple different physical phenomena, yielding in this way multidimensional data. This paper presents three methods for clustering sensors in WSNs whose sensors collect multidimensional data. The proposed approaches implement the concept of multidimensional behavioral clustering . To show the benefits introduced by the proposed methods, a prototype has been implemented and experiments have been carried out on real data. The results prove that the proposed methods decrease the amount of data flowing in the network and present low root-mean-square error (RMSE).
Multidimensional Profiling of Task Stress States for Human Factors: A Brief Review.
Matthews, Gerald
2016-09-01
This article advocates multidimensional assessment of task stress in human factors and reviews the use of the Dundee Stress State Questionnaire (DSSQ) for evaluation of systems and operators. Contemporary stress research has progressed from an exclusive focus on environmental stressors to transactional perspectives on the stress process. Performance impacts of stress reflect the operator's dynamic attempts to understand and cope with task demands. Multidimensional stress assessments are necessary to gauge the different forms of system-operator interaction. This review discusses the theoretical and practical use of the DSSQ in evaluating multidimensional patterns of stress response. It presents psychometric evidence for the multidimensional perspective and illustrative profiles of subjective state response to task stressors and environments. Evidence is also presented on stress state correlations with related variables, including personality, stress process measures, psychophysiological response, and objective task performance. Evidence supports the validity of the DSSQ as a task stress measure. Studies of various simulated environments show that different tasks elicit different profiles of stress state response. Operator characteristics such as resilience predict individual differences in state response to stressors. Structural equation modeling may be used to understand performance impacts of stress states. Multidimensional assessment affords insight into the stress process in a variety of human factors contexts. Integrating subjective and psychophysiological assessment is a priority for future research. Stress state measurement contributes to evaluating system design, countermeasures to stress and fatigue, and performance vulnerabilities. It may also support personnel selection and diagnostic monitoring of operators. © 2016, Human Factors and Ergonomics Society.
Inferring Higher Functional Information for RIKEN Mouse Full-Length cDNA Clones With FACTS
Nagashima, Takeshi; Silva, Diego G.; Petrovsky, Nikolai; Socha, Luis A.; Suzuki, Harukazu; Saito, Rintaro; Kasukawa, Takeya; Kurochkin, Igor V.; Konagaya, Akihiko; Schönbach, Christian
2003-01-01
FACTS (Functional Association/Annotation of cDNA Clones from Text/Sequence Sources) is a semiautomated knowledge discovery and annotation system that integrates molecular function information derived from sequence analysis results (sequence inferred) with functional information extracted from text. Text-inferred information was extracted from keyword-based retrievals of MEDLINE abstracts and by matching of gene or protein names to OMIM, BIND, and DIP database entries. Using FACTS, we found that 47.5% of the 60,770 RIKEN mouse cDNA FANTOM2 clone annotations were informative for text searches. MEDLINE queries yielded molecular interaction-containing sentences for 23.1% of the clones. When disease MeSH and GO terms were matched with retrieved abstracts, 22.7% of clones were associated with potential diseases, and 32.5% with GO identifiers. A significant number (23.5%) of disease MeSH-associated clones were also found to have a hereditary disease association (OMIM Morbidmap). Inferred neoplastic and nervous system disease represented 49.6% and 36.0% of disease MeSH-associated clones, respectively. A comparison of sequence-based GO assignments with informative text-based GO assignments revealed that for 78.2% of clones, identical GO assignments were provided for that clone by either method, whereas for 21.8% of clones, the assignments differed. In contrast, for OMIM assignments, only 28.5% of clones had identical sequence-based and text-based OMIM assignments. Sequence, sentence, and term-based functional associations are included in the FACTS database (http://facts.gsc.riken.go.jp/), which permits results to be annotated and explored through web-accessible keyword and sequence search interfaces. The FACTS database will be a critical tool for investigating the functional complexity of the mouse transcriptome, cDNA-inferred interactome (molecular interactions), and pathome (pathologies). PMID:12819151
NASA Technical Reports Server (NTRS)
Ambur, Manjula Y.; Adams, David L.; Trinidad, P. Paul
1997-01-01
NASA Langley Technical Library has been involved in developing systems for full-text information delivery of NACA/NASA technical reports since 1991. This paper will describe the two prototypes it has developed and the present production system configuration. The prototype systems are a NACA CD-ROM of thirty-three classic paper NACA reports and a network-based Full-text Electronic Reports Documents System (FEDS) constructed from both paper and electronic formats of NACA and NASA reports. The production system is the DigiDoc System (DIGItal Documents) presently being developed based on the experiences gained from the two prototypes. DigiDoc configuration integrates the on-line catalog database World Wide Web interface and PDF technology to provide a powerful and flexible search and retrieval system. It describes in detail significant achievements and lessons learned in terms of data conversion, storage technologies, full-text searching and retrieval, and image databases. The conclusions from the experiences of digitization and full- text access and future plans for DigiDoc system implementation are discussed.
"Staying safe" - a narrative review of falls prevention in people with Parkinson's - "PDSAFE".
Hulbert, Sophia; Rochester, Lynn; Nieuwboer, Alice; Goodwin, Vicki; Fitton, Carolyn; Chivers-Seymour, Kim; Ashburn, Ann
2018-05-18
Parkinson's disease demonstrates a spectrum of motor and non-motor symptoms. Falling is common and disabling. Current medical management shows minimal impact to reduce falls or fall-related risk factors, such as deficits in gait, strength, and postural instability. Despite evidence supporting rehabilitation in reducing fall risk factors, the most appropriate intervention to reduce overall fall rate remains inconclusive. This article aims to 1) synthesise current evidence and conceptual models of falls rehabilitation in Parkinson's in a narrative review; and based on this evidence, 2) introduce the treatment protocol used in the falls prevention and multi-centre clinical trial "PDSAFE". Search of four bibliographic databases using the terms "Parkinson*" and "Fall*" combined with each of the following; "Rehab*, Balanc*, Strength*, Strateg*and Exercis*" and a framework for narrative review was followed. A total of 3557 papers were identified, 416 were selected for review. The majority report the impact of rehabilitation on isolated fall risk factors. Twelve directly measure the impact on overall fall rate. Results were used to construct a narrative review with conceptual discussion based on the "International Classification of Functioning", leading to presentation of the "PDSAFE" intervention protocol. Evidence suggests training single, fall risk factors may not affect overall fall rate. Combining with behavioural and strategy training in a functional, personalised multi-dimensional model, addressing all components of the "International Classification of Functioning" is likely to provide a greater influence on falls reduction. "PDSAFE" is a multi-dimensional, physiotherapist delivered, individually tailored, progressive, home-based programme. It is designed with a strong evidence-based approach and illustrates a model for the clinical delivery of the conceptual theory discussed. Implications for Rehabilitation Parkinson's disease demonstrates a spectrum of motor and non-motor symptoms, where falling is common and disabling. Current medical and surgical management have minimal impact on falls, rehabilitation of falls risk factors has strong evidence but the most appropriate intervention to reduce overall fall rate remains inconclusive. Addressing all components of the International Classification of Function in a multifactorial model when designing falls rehabilitation interventions may be more effective at reducing fall rates in people with Parkinson's than treating isolated risk factors. The clinical model for falls rehabilitation in people with Parkinson's should be multi-dimensional.
Automatic reconstruction of a bacterial regulatory network using Natural Language Processing
Rodríguez-Penagos, Carlos; Salgado, Heladia; Martínez-Flores, Irma; Collado-Vides, Julio
2007-01-01
Background Manual curation of biological databases, an expensive and labor-intensive process, is essential for high quality integrated data. In this paper we report the implementation of a state-of-the-art Natural Language Processing system that creates computer-readable networks of regulatory interactions directly from different collections of abstracts and full-text papers. Our major aim is to understand how automatic annotation using Text-Mining techniques can complement manual curation of biological databases. We implemented a rule-based system to generate networks from different sets of documents dealing with regulation in Escherichia coli K-12. Results Performance evaluation is based on the most comprehensive transcriptional regulation database for any organism, the manually-curated RegulonDB, 45% of which we were able to recreate automatically. From our automated analysis we were also able to find some new interactions from papers not already curated, or that were missed in the manual filtering and review of the literature. We also put forward a novel Regulatory Interaction Markup Language better suited than SBML for simultaneously representing data of interest for biologists and text miners. Conclusion Manual curation of the output of automatic processing of text is a good way to complement a more detailed review of the literature, either for validating the results of what has been already annotated, or for discovering facts and information that might have been overlooked at the triage or curation stages. PMID:17683642
Text Mining to Support Gene Ontology Curation and Vice Versa.
Ruch, Patrick
2017-01-01
In this chapter, we explain how text mining can support the curation of molecular biology databases dealing with protein functions. We also show how curated data can play a disruptive role in the developments of text mining methods. We review a decade of efforts to improve the automatic assignment of Gene Ontology (GO) descriptors, the reference ontology for the characterization of genes and gene products. To illustrate the high potential of this approach, we compare the performances of an automatic text categorizer and show a large improvement of +225 % in both precision and recall on benchmarked data. We argue that automatic text categorization functions can ultimately be embedded into a Question-Answering (QA) system to answer questions related to protein functions. Because GO descriptors can be relatively long and specific, traditional QA systems cannot answer such questions. A new type of QA system, so-called Deep QA which uses machine learning methods trained with curated contents, is thus emerging. Finally, future advances of text mining instruments are directly dependent on the availability of high-quality annotated contents at every curation step. Databases workflows must start recording explicitly all the data they curate and ideally also some of the data they do not curate.
Bhanot, Gyan V [Princeton, NJ; Chen, Dong [Croton-On-Hudson, NY; Gara, Alan G [Mount Kisco, NY; Giampapa, Mark E [Irvington, NY; Heidelberger, Philip [Cortlandt Manor, NY; Steinmacher-Burow, Burkhard D [Mount Kisco, NY; Vranas, Pavlos M [Bedford Hills, NY
2012-01-10
The present in invention is directed to a method, system and program storage device for efficiently implementing a multidimensional Fast Fourier Transform (FFT) of a multidimensional array comprising a plurality of elements initially distributed in a multi-node computer system comprising a plurality of nodes in communication over a network, comprising: distributing the plurality of elements of the array in a first dimension across the plurality of nodes of the computer system over the network to facilitate a first one-dimensional FFT; performing the first one-dimensional FFT on the elements of the array distributed at each node in the first dimension; re-distributing the one-dimensional FFT-transformed elements at each node in a second dimension via "all-to-all" distribution in random order across other nodes of the computer system over the network; and performing a second one-dimensional FFT on elements of the array re-distributed at each node in the second dimension, wherein the random order facilitates efficient utilization of the network thereby efficiently implementing the multidimensional FFT. The "all-to-all" re-distribution of array elements is further efficiently implemented in applications other than the multidimensional FFT on the distributed-memory parallel supercomputer.
Bhanot, Gyan V [Princeton, NJ; Chen, Dong [Croton-On-Hudson, NY; Gara, Alan G [Mount Kisco, NY; Giampapa, Mark E [Irvington, NY; Heidelberger, Philip [Cortlandt Manor, NY; Steinmacher-Burow, Burkhard D [Mount Kisco, NY; Vranas, Pavlos M [Bedford Hills, NY
2008-01-01
The present in invention is directed to a method, system and program storage device for efficiently implementing a multidimensional Fast Fourier Transform (FFT) of a multidimensional array comprising a plurality of elements initially distributed in a multi-node computer system comprising a plurality of nodes in communication over a network, comprising: distributing the plurality of elements of the array in a first dimension across the plurality of nodes of the computer system over the network to facilitate a first one-dimensional FFT; performing the first one-dimensional FFT on the elements of the array distributed at each node in the first dimension; re-distributing the one-dimensional FFT-transformed elements at each node in a second dimension via "all-to-all" distribution in random order across other nodes of the computer system over the network; and performing a second one-dimensional FFT on elements of the array re-distributed at each node in the second dimension, wherein the random order facilitates efficient utilization of the network thereby efficiently implementing the multidimensional FFT. The "all-to-all" re-distribution of array elements is further efficiently implemented in applications other than the multidimensional FFT on the distributed-memory parallel supercomputer.
Image matrix processor for fast multi-dimensional computations
Roberson, George P.; Skeate, Michael F.
1996-01-01
An apparatus for multi-dimensional computation which comprises a computation engine, including a plurality of processing modules. The processing modules are configured in parallel and compute respective contributions to a computed multi-dimensional image of respective two dimensional data sets. A high-speed, parallel access storage system is provided which stores the multi-dimensional data sets, and a switching circuit routes the data among the processing modules in the computation engine and the storage system. A data acquisition port receives the two dimensional data sets representing projections through an image, for reconstruction algorithms such as encountered in computerized tomography. The processing modules include a programmable local host, by which they may be configured to execute a plurality of different types of multi-dimensional algorithms. The processing modules thus include an image manipulation processor, which includes a source cache, a target cache, a coefficient table, and control software for executing image transformation routines using data in the source cache and the coefficient table and loading resulting data in the target cache. The local host processor operates to load the source cache with a two dimensional data set, loads the coefficient table, and transfers resulting data out of the target cache to the storage system, or to another destination.