NASA Astrophysics Data System (ADS)
Chen, Zhu-an; Zhang, Li-ting; Liu, Lu
2009-10-01
ESRI's GIS components MapObjects are applied in many cadastral information system because of its miniaturization and flexibility. Some cadastral information was saved in cadastral database directly by MapObjects's Shape file format in this cadastral information system. However, MapObjects didn't provide the function of building attribute field for map layer's attribute data file in cadastral database and user cann't save the result of analysis. This present paper designed and realized the function of building attribute field in MapObjects based on the method of Jackson's system development.
Development of a Life History Database for Upper Mississippi River Fishes
2007-05-01
prevailing ecological and river theories with existing empirical data, investigating anthropogenic controls on functional attributes of ecosystems...2001; 2005a). database closely reflect the ecological attributes Finally, the life history database will allow the of UMRS fish species. These...34 Functional Feeding Guilds attribute class provide information on reproductive capacity, timing and mode for UMRS fish species. Our first example used the
WebEAV: automatic metadata-driven generation of web interfaces to entity-attribute-value databases.
Nadkarni, P M; Brandt, C M; Marenco, L
2000-01-01
The task of creating and maintaining a front end to a large institutional entity-attribute-value (EAV) database can be cumbersome when using traditional client-server technology. Switching to Web technology as a delivery vehicle solves some of these problems but introduces others. In particular, Web development environments tend to be primitive, and many features that client-server developers take for granted are missing. WebEAV is a generic framework for Web development that is intended to streamline the process of Web application development for databases having a significant EAV component. It also addresses some challenging user interface issues that arise when any complex system is created. The authors describe the architecture of WebEAV and provide an overview of its features with suitable examples.
Angermeier, Paul L.; Frimpong, Emmanuel A.
2009-01-01
The need for integrated and widely accessible sources of species traits data to facilitate studies of ecology, conservation, and management has motivated development of traits databases for various taxa. In spite of the increasing number of traits-based analyses of freshwater fishes in the United States, no consolidated database of traits of this group exists publicly, and much useful information on these species is documented only in obscure sources. The largely inaccessible and unconsolidated traits information makes large-scale analysis involving many fishes and/or traits particularly challenging. FishTraits is a database of >100 traits for 809 (731 native and 78 exotic) fish species found in freshwaters of the conterminous United States, including 37 native families and 145 native genera. The database contains information on four major categories of traits: (1) trophic ecology, (2) body size and reproductive ecology (life history), (3) habitat associations, and (4) salinity and temperature tolerances. Information on geographic distribution and conservation status is also included. Together, we refer to the traits, distribution, and conservation status information as attributes. Descriptions of attributes are available here. Many sources were consulted to compile attributes, including state and regional species accounts and other databases.
Research on keyword retrieval method of HBase database based on index structure
NASA Astrophysics Data System (ADS)
Gong, Pijin; Lv, Congmin; Gong, Yongsheng; Ma, Haozhi; Sun, Yang; Wang, Lu
2017-10-01
With the rapid development of manned spaceflight engineering, the scientific experimental data in space application system is increasing rapidly. How to efficiently query the specific data in the mass data volume has become a problem. In this paper, a method of retrieving the object data based on the object attribute as the keyword is proposed. The HBase database is used to store the object data and object attributes, and the secondary index is constructed. The research shows that this method is a good way to retrieve specified data based on object attributes.
GrainGenes: Changing Times, Changing Databases, Digital Evolution.
USDA-ARS?s Scientific Manuscript database
The GrainGenes database is one of few agricultural databases that had an early start on the Internet and that has changed with the times. Initial goals were to collect a wide range of data relating to the developing maps and attributes of small grains crops, and to make them easily accessible. The ...
Wang, Lizhu; Riseng, Catherine M.; Mason, Lacey; Werhrly, Kevin; Rutherford, Edward; McKenna, James E.; Castiglione, Chris; Johnson, Lucinda B.; Infante, Dana M.; Sowa, Scott P.; Robertson, Mike; Schaeffer, Jeff; Khoury, Mary; Gaiot, John; Hollenhurst, Tom; Brooks, Colin N.; Coscarelli, Mark
2015-01-01
Managing the world's largest and most complex freshwater ecosystem, the Laurentian Great Lakes, requires a spatially hierarchical basin-wide database of ecological and socioeconomic information that is comparable across the region. To meet such a need, we developed a spatial classification framework and database — Great Lakes Aquatic Habitat Framework (GLAHF). GLAHF consists of catchments, coastal terrestrial, coastal margin, nearshore, and offshore zones that encompass the entire Great Lakes Basin. The catchments captured in the database as river pour points or coastline segments are attributed with data known to influence physicochemical and biological characteristics of the lakes from the catchments. The coastal terrestrial zone consists of 30-m grid cells attributed with data from the terrestrial region that has direct connection with the lakes. The coastal margin and nearshore zones consist of 30-m grid cells attributed with data describing the coastline conditions, coastal human disturbances, and moderately to highly variable physicochemical and biological characteristics. The offshore zone consists of 1.8-km grid cells attributed with data that are spatially less variable compared with the other aquatic zones. These spatial classification zones and their associated data are nested within lake sub-basins and political boundaries and allow the synthesis of information from grid cells to classification zones, within and among political boundaries, lake sub-basins, Great Lakes, or within the entire Great Lakes Basin. This spatially structured database could help the development of basin-wide management plans, prioritize locations for funding and specific management actions, track protection and restoration progress, and conduct research for science-based decision making.
GraQL: A Query Language for High-Performance Attributed Graph Databases
DOE Office of Scientific and Technical Information (OSTI.GOV)
Chavarría-Miranda, Daniel; Castellana, Vito G.; Morari, Alessandro
Graph databases have gained increasing interest in the last few years due to the emergence of data sources which are not easily analyzable in traditional relational models or for which a graph data model is the natural representation. In order to understand the design and implementation choices for an attributed graph database backend and query language, we have started to design our infrastructure for attributed graph databases. In this paper, we describe the design considerations of our in-memory attributed graph database system with a particular focus on the data definition and query language components.
AN ASSESSMENT OF GROUND TRUTH VARIABILITY USING A "VIRTUAL FIELD REFERENCE DATABASE"
A "Virtual Field Reference Database (VFRDB)" was developed from field measurment data that included location and time, physical attributes, flora inventory, and digital imagery (camera) documentation foy 1,01I sites in the Neuse River basin, North Carolina. The sampling f...
World-wide precision airports for SVS
NASA Astrophysics Data System (ADS)
Schiefele, Jens; Lugsch, Bill; Launer, Marc; Baca, Diana
2004-08-01
Future cockpit and aviation applications require high quality airport databases. Accuracy, resolution, integrity, completeness, traceability, and timeliness [1] are key requirements. For most aviation applications, attributed vector databases are needed. The geometry is based on points, lines, and closed polygons. To document the needs for aviation industry RTCA and EUROCAE developed in a joint committee, the DO-272/ED-99 document. It states industry needs for data features, attributes, coding, and capture rules for Airport Mapping Databases (AMDB). This paper describes the technical approach Jeppesen has taken to generate a world-wide set of three-hundred AMDB airports. All AMDB airports are DO-200A/ED-76 [1] and DO-272/ED-99 [2] compliant. Jeppesen airports have a 5m (CE90) accuracy and an 10-3 integrity. World-wide all AMDB data is delivered in WGS84 coordinates. Jeppesen continually updates the databases.
Development of water environment information management and water pollution accident response system
NASA Astrophysics Data System (ADS)
Zhang, J.; Ruan, H.
2009-12-01
In recent years, many water pollution accidents occurred with the rapid economical development. In this study, water environment information management and water pollution accident response system are developed based on geographic information system (GIS) techniques. The system integrated spatial database, attribute database, hydraulic model, and water quality model under a user-friendly interface in a GIS environment. System ran in both Client/Server (C/S) and Browser/Server (B/S) platform which focused on model and inquiry respectively. System provided spatial and attribute data inquiry, water quality evaluation, statics, water pollution accident response case management (opening reservoir etc) and 2D and 3D visualization function, and gave assistant information to make decision on water pollution accident response. Polluted plume in Huaihe River were selected to simulate the transport of pollutes.
External Data and Attribute Hyperlink Programs for Promis*e(Registered Trademark)
NASA Technical Reports Server (NTRS)
Derengowski, Rich; Gruel, Andrew
2001-01-01
External Data and Attribute Hyperlink are computer programs that can be added to Promis*e(trademark) which is a commercial software system that automates routine tasks in the design (including drawing schematic diagrams) of electrical control systems. The programs were developed under the Stennis Space Center's (SSC) Dual Use Technology Development Program to provide capabilities for SSC's BMCS configuration management system which uses Promis*e(trademark). The External Data program enables the storage and management of information in an external database linked to a drawing. Changes can be made either in the database or on the drawing. Information that originates outside Promis*e(trademark) can be stored in custom fields that can be added to the database. Although this information is not available in Promis*e(trademark) printed drawings, it can be associated with symbols in the drawings, and can be retrieved through the drawings when the software is running. The Attribute Hyperlink program enables the addition of hyperlink information as attributes of symbols. This program enables the formation of a direct hyperlink between a schematic diagram and an Internet site or a file on a compact disk, on the user's hard drive, or on another computer on a network to which the user's computer is connected. The user can then obtain information directly related to the part (e.g., maintenance, or troubleshooting information) associated with the hyperlink.
Personal Database Management System I TRIAS
NASA Astrophysics Data System (ADS)
Yamamoto, Yoneo; Kashihara, Akihiro; Kawagishi, Keisuke
The current paper provides TRIAS (TRIple Associative System) which is a database management system for a personal use. In order to implement TRIAS, we have developed an associative database, whose format is (e,a,v) : e for entity, a for attribute, v for value. ML-TREE is used to construct (e,a,v). ML-TREE is a reversion of B+-tree that is multiway valanced tree. The paper focuses mainly on the usage of associative database, demonstrating how to use basic commands, primary functions and applcations.
Technologies and standards in the information systems of the soil-geographic database of Russia
NASA Astrophysics Data System (ADS)
Golozubov, O. M.; Rozhkov, V. A.; Alyabina, I. O.; Ivanov, A. V.; Kolesnikova, V. M.; Shoba, S. A.
2015-01-01
The achievements, problems, and challenges of the modern stage of the development of the Soil-Geographic Database of Russia (SGDBR) and the history of this project are outlined. The structure of the information system of the SGDBR as an internet-based resource to collect data on soil profiles and to integrate the geographic and attribute databases on the same platform is described. The pilot project in Rostov oblast illustrates the inclusion of regional information in the SGDBR and its application for solving practical problems. For the first time in Russia, the GeoRSS standard based on the structured hypertext representation of the geographic and attribute information has been applied in the state system for the agromonitoring of agricultural lands in Rostov oblast and information exchange through the internet.
Zhang, Jie; Wang, Yuping; Feng, Junhong
2013-01-01
In association rule mining, evaluating an association rule needs to repeatedly scan database to compare the whole database with the antecedent, consequent of a rule and the whole rule. In order to decrease the number of comparisons and time consuming, we present an attribute index strategy. It only needs to scan database once to create the attribute index of each attribute. Then all metrics values to evaluate an association rule do not need to scan database any further, but acquire data only by means of the attribute indices. The paper visualizes association rule mining as a multiobjective problem rather than a single objective one. In order to make the acquired solutions scatter uniformly toward the Pareto frontier in the objective space, elitism policy and uniform design are introduced. The paper presents the algorithm of attribute index and uniform design based multiobjective association rule mining with evolutionary algorithm, abbreviated as IUARMMEA. It does not require the user-specified minimum support and minimum confidence anymore, but uses a simple attribute index. It uses a well-designed real encoding so as to extend its application scope. Experiments performed on several databases demonstrate that the proposed algorithm has excellent performance, and it can significantly reduce the number of comparisons and time consumption.
Wang, Yuping; Feng, Junhong
2013-01-01
In association rule mining, evaluating an association rule needs to repeatedly scan database to compare the whole database with the antecedent, consequent of a rule and the whole rule. In order to decrease the number of comparisons and time consuming, we present an attribute index strategy. It only needs to scan database once to create the attribute index of each attribute. Then all metrics values to evaluate an association rule do not need to scan database any further, but acquire data only by means of the attribute indices. The paper visualizes association rule mining as a multiobjective problem rather than a single objective one. In order to make the acquired solutions scatter uniformly toward the Pareto frontier in the objective space, elitism policy and uniform design are introduced. The paper presents the algorithm of attribute index and uniform design based multiobjective association rule mining with evolutionary algorithm, abbreviated as IUARMMEA. It does not require the user-specified minimum support and minimum confidence anymore, but uses a simple attribute index. It uses a well-designed real encoding so as to extend its application scope. Experiments performed on several databases demonstrate that the proposed algorithm has excellent performance, and it can significantly reduce the number of comparisons and time consumption. PMID:23766683
Chen, Yi- Ping Phoebe; Hanan, Jim
2002-01-01
Models of plant architecture allow us to explore how genotype environment interactions effect the development of plant phenotypes. Such models generate masses of data organised in complex hierarchies. This paper presents a generic system for creating and automatically populating a relational database from data generated by the widely used L-system approach to modelling plant morphogenesis. Techniques from compiler technology are applied to generate attributes (new fields) in the database, to simplify query development for the recursively-structured branching relationship. Use of biological terminology in an interactive query builder contributes towards making the system biologist-friendly.
An online database for informing ecological network models: http://kelpforest.ucsc.edu.
Beas-Luna, Rodrigo; Novak, Mark; Carr, Mark H; Tinker, Martin T; Black, August; Caselle, Jennifer E; Hoban, Michael; Malone, Dan; Iles, Alison
2014-01-01
Ecological network models and analyses are recognized as valuable tools for understanding the dynamics and resiliency of ecosystems, and for informing ecosystem-based approaches to management. However, few databases exist that can provide the life history, demographic and species interaction information necessary to parameterize ecological network models. Faced with the difficulty of synthesizing the information required to construct models for kelp forest ecosystems along the West Coast of North America, we developed an online database (http://kelpforest.ucsc.edu/) to facilitate the collation and dissemination of such information. Many of the database's attributes are novel yet the structure is applicable and adaptable to other ecosystem modeling efforts. Information for each taxonomic unit includes stage-specific life history, demography, and body-size allometries. Species interactions include trophic, competitive, facilitative, and parasitic forms. Each data entry is temporally and spatially explicit. The online data entry interface allows researchers anywhere to contribute and access information. Quality control is facilitated by attributing each entry to unique contributor identities and source citations. The database has proven useful as an archive of species and ecosystem-specific information in the development of several ecological network models, for informing management actions, and for education purposes (e.g., undergraduate and graduate training). To facilitate adaptation of the database by other researches for other ecosystems, the code and technical details on how to customize this database and apply it to other ecosystems are freely available and located at the following link (https://github.com/kelpforest-cameo/databaseui).
An Online Database for Informing Ecological Network Models: http://kelpforest.ucsc.edu
Beas-Luna, Rodrigo; Novak, Mark; Carr, Mark H.; Tinker, Martin T.; Black, August; Caselle, Jennifer E.; Hoban, Michael; Malone, Dan; Iles, Alison
2014-01-01
Ecological network models and analyses are recognized as valuable tools for understanding the dynamics and resiliency of ecosystems, and for informing ecosystem-based approaches to management. However, few databases exist that can provide the life history, demographic and species interaction information necessary to parameterize ecological network models. Faced with the difficulty of synthesizing the information required to construct models for kelp forest ecosystems along the West Coast of North America, we developed an online database (http://kelpforest.ucsc.edu/) to facilitate the collation and dissemination of such information. Many of the database's attributes are novel yet the structure is applicable and adaptable to other ecosystem modeling efforts. Information for each taxonomic unit includes stage-specific life history, demography, and body-size allometries. Species interactions include trophic, competitive, facilitative, and parasitic forms. Each data entry is temporally and spatially explicit. The online data entry interface allows researchers anywhere to contribute and access information. Quality control is facilitated by attributing each entry to unique contributor identities and source citations. The database has proven useful as an archive of species and ecosystem-specific information in the development of several ecological network models, for informing management actions, and for education purposes (e.g., undergraduate and graduate training). To facilitate adaptation of the database by other researches for other ecosystems, the code and technical details on how to customize this database and apply it to other ecosystems are freely available and located at the following link (https://github.com/kelpforest-cameo/databaseui). PMID:25343723
An online database for informing ecological network models: http://kelpforest.ucsc.edu
Beas-Luna, Rodrigo; Tinker, M. Tim; Novak, Mark; Carr, Mark H.; Black, August; Caselle, Jennifer E.; Hoban, Michael; Malone, Dan; Iles, Alison C.
2014-01-01
Ecological network models and analyses are recognized as valuable tools for understanding the dynamics and resiliency of ecosystems, and for informing ecosystem-based approaches to management. However, few databases exist that can provide the life history, demographic and species interaction information necessary to parameterize ecological network models. Faced with the difficulty of synthesizing the information required to construct models for kelp forest ecosystems along the West Coast of North America, we developed an online database (http://kelpforest.ucsc.edu/) to facilitate the collation and dissemination of such information. Many of the database's attributes are novel yet the structure is applicable and adaptable to other ecosystem modeling efforts. Information for each taxonomic unit includes stage-specific life history, demography, and body-size allometries. Species interactions include trophic, competitive, facilitative, and parasitic forms. Each data entry is temporally and spatially explicit. The online data entry interface allows researchers anywhere to contribute and access information. Quality control is facilitated by attributing each entry to unique contributor identities and source citations. The database has proven useful as an archive of species and ecosystem-specific information in the development of several ecological network models, for informing management actions, and for education purposes (e.g., undergraduate and graduate training). To facilitate adaptation of the database by other researches for other ecosystems, the code and technical details on how to customize this database and apply it to other ecosystems are freely available and located at the following link (https://github.com/kelpforest-cameo/databaseui).
Security of statistical data bases: invasion of privacy through attribute correlational modeling
DOE Office of Scientific and Technical Information (OSTI.GOV)
Palley, M.A.
This study develops, defines, and applies a statistical technique for the compromise of confidential information in a statistical data base. Attribute Correlational Modeling (ACM) recognizes that the information contained in a statistical data base represents real world statistical phenomena. As such, ACM assumes correlational behavior among the database attributes. ACM proceeds to compromise confidential information through creation of a regression model, where the confidential attribute is treated as the dependent variable. The typical statistical data base may preclude the direct application of regression. In this scenario, the research introduces the notion of a synthetic data base, created through legitimate queriesmore » of the actual data base, and through proportional random variation of responses to these queries. The synthetic data base is constructed to resemble the actual data base as closely as possible in a statistical sense. ACM then applies regression analysis to the synthetic data base, and utilizes the derived model to estimate confidential information in the actual database.« less
The risk of paradoxical embolism (RoPE) study: initial description of the completed database.
Thaler, David E; Di Angelantonio, Emanuele; Di Tullio, Marco R; Donovan, Jennifer S; Griffith, John; Homma, Shunichi; Jaigobin, Cheryl; Mas, Jean-Louis; Mattle, Heinrich P; Michel, Patrik; Mono, Marie-Luise; Nedeltchev, Krassen; Papetti, Federica; Ruthazer, Robin; Serena, Joaquín; Weimar, Christian; Elkind, Mitchell S V; Kent, David M
2013-12-01
Detecting a benefit from closure of patent foramen ovale in patients with cryptogenic stroke is hampered by low rates of stroke recurrence and uncertainty about the causal role of patent foramen ovale in the index event. A method to predict patent foramen ovale-attributable recurrence risk is needed. However, individual databases generally have too few stroke recurrences to support risk modeling. Prior studies of this population have been limited by low statistical power for examining factors related to recurrence. The aim of this study was to develop a database to support modeling of patent foramen ovale-attributable recurrence risk by combining extant data sets. We identified investigators with extant databases including subjects with cryptogenic stroke investigated for patent foramen ovale, determined the availability and characteristics of data in each database, collaboratively specified the variables to be included in the Risk of Paradoxical Embolism database, harmonized the variables across databases, and collected new primary data when necessary and feasible. The Risk of Paradoxical Embolism database has individual clinical, radiologic, and echocardiographic data from 12 component databases, including subjects with cryptogenic stroke both with (n = 1925) and without (n = 1749) patent foramen ovale. In the patent foramen ovale subjects, a total of 381 outcomes (stroke, transient ischemic attack, death) occurred (median follow-up 2·2 years). While there were substantial variations in data collection between studies, there was sufficient overlap to define a common set of variables suitable for risk modeling. While individual studies are inadequate for modeling patent foramen ovale-attributable recurrence risk, collaboration between investigators has yielded a database with sufficient power to identify those patients at highest risk for a patent foramen ovale-related stroke recurrence who may have the greatest potential benefit from patent foramen ovale closure. © 2012 The Authors. International Journal of Stroke © 2012 World Stroke Organization.
Generation And Understanding Of Natural Language Using Information In A Frame Structure
NASA Astrophysics Data System (ADS)
Perkins, Walton A.
1989-03-01
Many expert systems and relational database systems store factual information in the form of attributes values of objects. Problems arise in transforming from that attribute (frame) database representation into English surface structure and in transforming the English surface structure into a representation that references information in the frame database. In this paper we consider mainly the generation process, as it is this area in which we have made the most significant progress. In its interaction with the user, the expert system must generate questions, declarations, and uncertain declarations. Attributes such as COLOR, LENGTH, and ILLUMINATION can be referenced using the template: "
The LSST Data Mining Research Agenda
NASA Astrophysics Data System (ADS)
Borne, K.; Becla, J.; Davidson, I.; Szalay, A.; Tyson, J. A.
2008-12-01
We describe features of the LSST science database that are amenable to scientific data mining, object classification, outlier identification, anomaly detection, image quality assurance, and survey science validation. The data mining research agenda includes: scalability (at petabytes scales) of existing machine learning and data mining algorithms; development of grid-enabled parallel data mining algorithms; designing a robust system for brokering classifications from the LSST event pipeline (which may produce 10,000 or more event alerts per night) multi-resolution methods for exploration of petascale databases; indexing of multi-attribute multi-dimensional astronomical databases (beyond spatial indexing) for rapid querying of petabyte databases; and more.
MetPetDB: A database for metamorphic geochemistry
NASA Astrophysics Data System (ADS)
Spear, Frank S.; Hallett, Benjamin; Pyle, Joseph M.; Adalı, Sibel; Szymanski, Boleslaw K.; Waters, Anthony; Linder, Zak; Pearce, Shawn O.; Fyffe, Matthew; Goldfarb, Dennis; Glickenhouse, Nickolas; Buletti, Heather
2009-12-01
We present a data model for the initial implementation of MetPetDB, a geochemical database specific to metamorphic rock samples. The database is designed around the concept of preservation of spatial relationships, at all scales, of chemical analyses and their textural setting. Objects in the database (samples) represent physical rock samples; each sample may contain one or more subsamples with associated geochemical and image data. Samples, subsamples, geochemical data, and images are described with attributes (some required, some optional); these attributes also serve as search delimiters. All data in the database are classified as published (i.e., archived or published data), public or private. Public and published data may be freely searched and downloaded. All private data is owned; permission to view, edit, download and otherwise manipulate private data may be granted only by the data owner; all such editing operations are recorded by the database to create a data version log. The sharing of data permissions among a group of collaborators researching a common sample is done by the sample owner through the project manager. User interaction with MetPetDB is hosted by a web-based platform based upon the Java servlet application programming interface, with the PostgreSQL relational database. The database web portal includes modules that allow the user to interact with the database: registered users may save and download public and published data, upload private data, create projects, and assign permission levels to project collaborators. An Image Viewer module provides for spatial integration of image and geochemical data. A toolkit consisting of plotting and geochemical calculation software for data analysis and a mobile application for viewing the public and published data is being developed. Future issues to address include population of the database, integration with other geochemical databases, development of the analysis toolkit, creation of data models for derivative data, and building a community-wide user base. It is believed that this and other geochemical databases will enable more productive collaborations, generate more efficient research efforts, and foster new developments in basic research in the field of solid earth geochemistry.
Huber, Lara
2011-06-01
In the neurosciences digital databases more and more are becoming important tools of data rendering and distributing. This development is due to the growing impact of imaging based trial design in cognitive neuroscience, including morphological as much as functional imaging technologies. As the case of the 'Laboratory of Neuro Imaging' (LONI) is showing, databases are attributed a specific epistemological power: Since the 1990s databasing is seen to foster the integration of neuroscientific data, although local regimes of data production, -manipulation and--interpretation are also challenging this development. Databasing in the neurosciences goes along with the introduction of new structures of integrating local data, hence establishing digital spaces of knowledge (epistemic spaces): At this stage, inherent norms of digital databases are affecting regimes of imaging-based trial design, for example clinical research into Alzheimer's disease.
Chen, R S; Nadkarni, P; Marenco, L; Levin, F; Erdos, J; Miller, P L
2000-01-01
The entity-attribute-value representation with classes and relationships (EAV/CR) provides a flexible and simple database schema to store heterogeneous biomedical data. In certain circumstances, however, the EAV/CR model is known to retrieve data less efficiently than conventionally based database schemas. To perform a pilot study that systematically quantifies performance differences for database queries directed at real-world microbiology data modeled with EAV/CR and conventional representations, and to explore the relative merits of different EAV/CR query implementation strategies. Clinical microbiology data obtained over a ten-year period were stored using both database models. Query execution times were compared for four clinically oriented attribute-centered and entity-centered queries operating under varying conditions of database size and system memory. The performance characteristics of three different EAV/CR query strategies were also examined. Performance was similar for entity-centered queries in the two database models. Performance in the EAV/CR model was approximately three to five times less efficient than its conventional counterpart for attribute-centered queries. The differences in query efficiency became slightly greater as database size increased, although they were reduced with the addition of system memory. The authors found that EAV/CR queries formulated using multiple, simple SQL statements executed in batch were more efficient than single, large SQL statements. This paper describes a pilot project to explore issues in and compare query performance for EAV/CR and conventional database representations. Although attribute-centered queries were less efficient in the EAV/CR model, these inefficiencies may be addressable, at least in part, by the use of more powerful hardware or more memory, or both.
A Data Warehouse to Support Condition Based Maintenance (CBM)
2005-05-01
Application ( VBA ) code sequence to import the original MAST-generated CSV and then create a single output table in DBASE IV format. The DBASE IV format...database architecture (Oracle, Sybase, MS- SQL , etc). This design includes table definitions, comments, specification of table attributes, primary and foreign...built queries and applications. Needs the application developers to construct data views. No SQL programming experience. b. Power Database User - knows
Nicholson, Suzanne W.; Stoeser, Douglas B.; Wilson, Frederic H.; Dicken, Connie L.; Ludington, Steve
2007-01-01
The growth in the use of Geographic nformation Systems (GS) has highlighted the need for regional and national digital geologic maps attributed with age and rock type information. Such spatial data can be conveniently used to generate derivative maps for purposes that include mineral-resource assessment, metallogenic studies, tectonic studies, human health and environmental research. n 1997, the United States Geological Survey’s Mineral Resources Program initiated an effort to develop national digital databases for use in mineral resource and environmental assessments. One primary activity of this effort was to compile a national digital geologic map database, utilizing state geologic maps, to support mineral resource studies in the range of 1:250,000- to 1:1,000,000-scale. Over the course of the past decade, state databases were prepared using a common standard for the database structure, fields, attributes, and data dictionaries. As of late 2006, standardized geological map databases for all conterminous (CONUS) states have been available on-line as USGS Open-File Reports. For Alaska and Hawaii, new state maps are being prepared, and the preliminary work for Alaska is being released as a series of 1:500,000-scale regional compilations. See below for a list of all published databases.
Genetics and attribution issues that confront the microbial forensics field.
Budowle, Bruce
2004-12-02
The commission of an act of bioterrorism or biocrime is a real concern for law enforcement and society. Efforts are underway to develop a strong microbial forensic program to assist in identifying perpetrators of acts of bioterrorism and biocrimes, as well as serve as a deterrent for those who might commit such illicit acts. Genetic analyses of microbial organisms will likely be a powerful tool for attribution of criminal acts. There are some similarities to forensic human DNA analysis practices, such as: molecular biology technology, use of population databases, qualitative conclusions of test results, and the application of QA/QC practices. Differences include: database size and composition, statistical interpretation methods, and confidence/uncertainty in the outcome of an interpretation.
The Unified Database for BM@N experiment data handling
NASA Astrophysics Data System (ADS)
Gertsenberger, Konstantin; Rogachevsky, Oleg
2018-04-01
The article describes the developed Unified Database designed as a comprehensive relational data storage for the BM@N experiment at the Joint Institute for Nuclear Research in Dubna. The BM@N experiment, which is one of the main elements of the first stage of the NICA project, is a fixed target experiment at extracted Nuclotron beams of the Laboratory of High Energy Physics (LHEP JINR). The structure and purposes of the BM@N setup are briefly presented. The article considers the scheme of the Unified Database, its attributes and implemented features in detail. The use of the developed BM@N database provides correct multi-user access to actual information of the experiment for data processing. It stores information on the experiment runs, detectors and their geometries, different configuration, calibration and algorithm parameters used in offline data processing. An important part of any database - user interfaces are presented.
Pretest probability assessment derived from attribute matching
Kline, Jeffrey A; Johnson, Charles L; Pollack, Charles V; Diercks, Deborah B; Hollander, Judd E; Newgard, Craig D; Garvey, J Lee
2005-01-01
Background Pretest probability (PTP) assessment plays a central role in diagnosis. This report compares a novel attribute-matching method to generate a PTP for acute coronary syndrome (ACS). We compare the new method with a validated logistic regression equation (LRE). Methods Eight clinical variables (attributes) were chosen by classification and regression tree analysis of a prospectively collected reference database of 14,796 emergency department (ED) patients evaluated for possible ACS. For attribute matching, a computer program identifies patients within the database who have the exact profile defined by clinician input of the eight attributes. The novel method was compared with the LRE for ability to produce PTP estimation <2% in a validation set of 8,120 patients evaluated for possible ACS and did not have ST segment elevation on ECG. 1,061 patients were excluded prior to validation analysis because of ST-segment elevation (713), missing data (77) or being lost to follow-up (271). Results In the validation set, attribute matching produced 267 unique PTP estimates [median PTP value 6%, 1st–3rd quartile 1–10%] compared with the LRE, which produced 96 unique PTP estimates [median 24%, 1st–3rd quartile 10–30%]. The areas under the receiver operating characteristic curves were 0.74 (95% CI 0.65 to 0.82) for the attribute matching curve and 0.68 (95% CI 0.62 to 0.77) for LRE. The attribute matching system categorized 1,670 (24%, 95% CI = 23–25%) patients as having a PTP < 2.0%; 28 developed ACS (1.7% 95% CI = 1.1–2.4%). The LRE categorized 244 (4%, 95% CI = 3–4%) with PTP < 2.0%; four developed ACS (1.6%, 95% CI = 0.4–4.1%). Conclusion Attribute matching estimated a very low PTP for ACS in a significantly larger proportion of ED patients compared with a validated LRE. PMID:16095534
Guidelines for the Effective Use of Entity-Attribute-Value Modeling for Biomedical Databases
Dinu, Valentin; Nadkarni, Prakash
2007-01-01
Purpose To introduce the goals of EAV database modeling, to describe the situations where Entity-Attribute-Value (EAV) modeling is a useful alternative to conventional relational methods of database modeling, and to describe the fine points of implementation in production systems. Methods We analyze the following circumstances: 1) data are sparse and have a large number of applicable attributes, but only a small fraction will apply to a given entity; 2) numerous classes of data need to be represented, each class has a limited number of attributes, but the number of instances of each class is very small. We also consider situations calling for a mixed approach where both conventional and EAV design are used for appropriate data classes. Results and Conclusions In robust production systems, EAV-modeled databases trade a modest data sub-schema for a complex metadata sub-schema. The need to design the metadata effectively makes EAV design potentially more challenging than conventional design. PMID:17098467
Towards BioDBcore: a community-defined information specification for biological databases
Gaudet, Pascale; Bairoch, Amos; Field, Dawn; Sansone, Susanna-Assunta; Taylor, Chris; Attwood, Teresa K.; Bateman, Alex; Blake, Judith A.; Bult, Carol J.; Cherry, J. Michael; Chisholm, Rex L.; Cochrane, Guy; Cook, Charles E.; Eppig, Janan T.; Galperin, Michael Y.; Gentleman, Robert; Goble, Carole A.; Gojobori, Takashi; Hancock, John M.; Howe, Douglas G.; Imanishi, Tadashi; Kelso, Janet; Landsman, David; Lewis, Suzanna E.; Mizrachi, Ilene Karsch; Orchard, Sandra; Ouellette, B. F. Francis; Ranganathan, Shoba; Richardson, Lorna; Rocca-Serra, Philippe; Schofield, Paul N.; Smedley, Damian; Southan, Christopher; Tan, Tin Wee; Tatusova, Tatiana; Whetzel, Patricia L.; White, Owen; Yamasaki, Chisato
2011-01-01
The present article proposes the adoption of a community-defined, uniform, generic description of the core attributes of biological databases, BioDBCore. The goals of these attributes are to provide a general overview of the database landscape, to encourage consistency and interoperability between resources and to promote the use of semantic and syntactic standards. BioDBCore will make it easier for users to evaluate the scope and relevance of available resources. This new resource will increase the collective impact of the information present in biological databases. PMID:21097465
Towards BioDBcore: a community-defined information specification for biological databases
Gaudet, Pascale; Bairoch, Amos; Field, Dawn; Sansone, Susanna-Assunta; Taylor, Chris; Attwood, Teresa K.; Bateman, Alex; Blake, Judith A.; Bult, Carol J.; Cherry, J. Michael; Chisholm, Rex L.; Cochrane, Guy; Cook, Charles E.; Eppig, Janan T.; Galperin, Michael Y.; Gentleman, Robert; Goble, Carole A.; Gojobori, Takashi; Hancock, John M.; Howe, Douglas G.; Imanishi, Tadashi; Kelso, Janet; Landsman, David; Lewis, Suzanna E.; Karsch Mizrachi, Ilene; Orchard, Sandra; Ouellette, B.F. Francis; Ranganathan, Shoba; Richardson, Lorna; Rocca-Serra, Philippe; Schofield, Paul N.; Smedley, Damian; Southan, Christopher; Tan, Tin W.; Tatusova, Tatiana; Whetzel, Patricia L.; White, Owen; Yamasaki, Chisato
2011-01-01
The present article proposes the adoption of a community-defined, uniform, generic description of the core attributes of biological databases, BioDBCore. The goals of these attributes are to provide a general overview of the database landscape, to encourage consistency and interoperability between resources; and to promote the use of semantic and syntactic standards. BioDBCore will make it easier for users to evaluate the scope and relevance of available resources. This new resource will increase the collective impact of the information present in biological databases. PMID:21205783
Wilson, Frederic H.; Hults, Chad P.; Mull, Charles G.; Karl, Susan M.
2015-12-31
This Alaska compilation is unique in that it is integrated with a rich database of information provided in the spatial datasets and standalone attribute databases. Within the spatial files every line and polygon is attributed to its original source; the references to these sources are contained in related tables, as well as in stand-alone tables. Additional attributes include typical lithology, geologic setting, and age range for the map units. Also included are tables of radiometric ages.
International Data on Radiological Sources
DOE Office of Scientific and Technical Information (OSTI.GOV)
Martha Finck; Margaret Goldberg
2010-07-01
ABSTRACT The mission of radiological dispersal device (RDD) nuclear forensics is to identify the provenance of nuclear and radiological materials used in RDDs and to aid law enforcement in tracking nuclear materials and routes. The application of databases to radiological forensics is to match RDD source material to a source model in the database, provide guidance regarding a possible second device, and aid the FBI by providing a short list of manufacturers and distributors, and ultimately to the last legal owner of the source. The Argonne/Idaho National Laboratory RDD attribution database is a powerful technical tool in radiological forensics. Themore » database (1267 unique vendors) includes all sealed sources and a device registered in the U.S., is complemented by data from the IAEA Catalogue, and is supported by rigorous in-lab characterization of selected sealed sources regarding physical form, radiochemical composition, and age-dating profiles. Close working relationships with global partners in the commercial sealed sources industry provide invaluable technical information and expertise in the development of signature profiles. These profiles are critical to the down-selection of potential candidates in either pre- or post- event RDD attribution. The down-selection process includes a match between an interdicted (or detonated) source and a model in the database linked to one or more manufacturers and distributors.« less
Geodata Modeling and Query in Geographic Information Systems
NASA Technical Reports Server (NTRS)
Adam, Nabil
1996-01-01
Geographic information systems (GIS) deal with collecting, modeling, man- aging, analyzing, and integrating spatial (locational) and non-spatial (attribute) data required for geographic applications. Examples of spatial data are digital maps, administrative boundaries, road networks, and those of non-spatial data are census counts, land elevations and soil characteristics. GIS shares common areas with a number of other disciplines such as computer- aided design, computer cartography, database management, and remote sensing. None of these disciplines however, can by themselves fully meet the requirements of a GIS application. Examples of such requirements include: the ability to use locational data to produce high quality plots, perform complex operations such as network analysis, enable spatial searching and overlay operations, support spatial analysis and modeling, and provide data management functions such as efficient storage, retrieval, and modification of large datasets; independence, integrity, and security of data; and concurrent access to multiple users. It is on the data management issues that we devote our discussions in this monograph. Traditionally, database management technology have been developed for business applications. Such applications require, among other things, capturing the data requirements of high-level business functions and developing machine- level implementations; supporting multiple views of data and yet providing integration that would minimize redundancy and maintain data integrity and security; providing a high-level language for data definition and manipulation; allowing concurrent access to multiple users; and processing user transactions in an efficient manner. The demands on database management systems have been for speed, reliability, efficiency, cost effectiveness, and user-friendliness. Significant progress have been made in all of these areas over the last two decades to the point that many generalized database platforms are now available for developing data intensive applications that run in real-time. While continuous improvement is still being made at a very fast-paced and competitive rate, new application areas such as computer aided design, image processing, VLSI design, and GIS have been identified by many as the next generation of database applications. These new application areas pose serious challenges to the currently available database technology. At the core of these challenges is the nature of data that is manipulated. In traditional database applications, the database objects do not have any spatial dimension, and as such, can be thought of as point data in a multi-dimensional space. For example, each instance of an entity EMPLOYEE will have a unique value corresponding to every attribute such as employee id, employee name, employee address and so on. Thus, every Employee instance can be thought of as a point in a multi-dimensional space where each dimension is represented by an attribute. Furthermore, all operations on such data are one-dimensional. Thus, users may retrieve all entities satisfying one or more constraints. Examples of such constraints include employees with addresses in a certain area code, or salaries within a certain range. Even though constraints can be specified on multiple attributes (dimensions), the search for such data is essentially orthogonal across these dimensions.
Data Mining Research with the LSST
NASA Astrophysics Data System (ADS)
Borne, Kirk D.; Strauss, M. A.; Tyson, J. A.
2007-12-01
The LSST catalog database will exceed 10 petabytes, comprising several hundred attributes for 5 billion galaxies, 10 billion stars, and over 1 billion variable sources (optical variables, transients, or moving objects), extracted from over 20,000 square degrees of deep imaging in 5 passbands with thorough time domain coverage: 1000 visits over the 10-year LSST survey lifetime. The opportunities are enormous for novel scientific discoveries within this rich time-domain ultra-deep multi-band survey database. Data Mining, Machine Learning, and Knowledge Discovery research opportunities with the LSST are now under study, with a potential for new collaborations to develop to contribute to these investigations. We will describe features of the LSST science database that are amenable to scientific data mining, object classification, outlier identification, anomaly detection, image quality assurance, and survey science validation. We also give some illustrative examples of current scientific data mining research in astronomy, and point out where new research is needed. In particular, the data mining research community will need to address several issues in the coming years as we prepare for the LSST data deluge. The data mining research agenda includes: scalability (at petabytes scales) of existing machine learning and data mining algorithms; development of grid-enabled parallel data mining algorithms; designing a robust system for brokering classifications from the LSST event pipeline (which may produce 10,000 or more event alerts per night); multi-resolution methods for exploration of petascale databases; visual data mining algorithms for visual exploration of the data; indexing of multi-attribute multi-dimensional astronomical databases (beyond RA-Dec spatial indexing) for rapid querying of petabyte databases; and more. Finally, we will identify opportunities for synergistic collaboration between the data mining research group and the LSST Data Management and Science Collaboration teams.
Terminological aspects of data elements
DOE Office of Scientific and Technical Information (OSTI.GOV)
Strehlow, R.A.; Kenworthey, W.H. Jr.; Schuldt, R.E.
1991-01-01
The creation and display of data comprise a process that involves a sequence of steps requiring both semantic and systems analysis. An essential early step in this process is the choice, definition, and naming of data element concepts and is followed by the specification of other needed data element concept attributes. The attributes and the values of data element concept remain associated with them from their birth as a concept to a generic data element that serves as a template for final application. Terminology is, therefore, centrally important to the entire data creation process. Smooth mapping from natural language tomore » a database is a critical aspect of database, and consequently, it requires terminology standardization from the outset of database work. In this paper the semantic aspects of data elements are analyzed and discussed. Seven kinds of data element concept information are considered and those that require terminological development and standardization are identified. The four terminological components of a data element are the hierarchical type of a concept, functional dependencies, schematas showing conceptual structures, and definition statements. These constitute the conventional role of terminology in database design. 12 refs., 8 figs., 1 tab.« less
Karst database development in Minnesota: Design and data assembly
Gao, Y.; Alexander, E.C.; Tipping, R.G.
2005-01-01
The Karst Feature Database (KFD) of Minnesota is a relational GIS-based Database Management System (DBMS). Previous karst feature datasets used inconsistent attributes to describe karst features in different areas of Minnesota. Existing metadata were modified and standardized to represent a comprehensive metadata for all the karst features in Minnesota. Microsoft Access 2000 and ArcView 3.2 were used to develop this working database. Existing county and sub-county karst feature datasets have been assembled into the KFD, which is capable of visualizing and analyzing the entire data set. By November 17 2002, 11,682 karst features were stored in the KFD of Minnesota. Data tables are stored in a Microsoft Access 2000 DBMS and linked to corresponding ArcView applications. The current KFD of Minnesota has been moved from a Windows NT server to a Windows 2000 Citrix server accessible to researchers and planners through networked interfaces. ?? Springer-Verlag 2005.
Attribute and topology based change detection in a constellation of previously detected objects
Paglieroni, David W.; Beer, Reginald N.
2016-01-19
A system that applies attribute and topology based change detection to networks of objects that were detected on previous scans of a structure, roadway, or area of interest. The attributes capture properties or characteristics of the previously detected objects, such as location, time of detection, size, elongation, orientation, etc. The topology of the network of previously detected objects is maintained in a constellation database that stores attributes of previously detected objects and implicitly captures the geometrical structure of the network. A change detection system detects change by comparing the attributes and topology of new objects detected on the latest scan to the constellation database of previously detected objects.
Multimodality medical image database for temporal lobe epilepsy
NASA Astrophysics Data System (ADS)
Siadat, Mohammad-Reza; Soltanian-Zadeh, Hamid; Fotouhi, Farshad A.; Elisevich, Kost
2003-05-01
This paper presents the development of a human brain multi-modality database for surgical candidacy determination in temporal lobe epilepsy. The focus of the paper is on content-based image management, navigation and retrieval. Several medical image-processing methods including our newly developed segmentation method are utilized for information extraction/correlation and indexing. The input data includes T1-, T2-Weighted and FLAIR MRI and ictal/interictal SPECT modalities with associated clinical data and EEG data analysis. The database can answer queries regarding issues such as the correlation between the attribute X of the entity Y and the outcome of a temporal lobe epilepsy surgery. The entity Y can be a brain anatomical structure such as the hippocampus. The attribute X can be either a functionality feature of the anatomical structure Y, calculated with SPECT modalities, such as signal average, or a volumetric/morphological feature of the entity Y such as volume or average curvature. The outcome of the surgery can be any surgery assessment such as non-verbal Wechsler memory quotient. A determination is made regarding surgical candidacy by analysis of both textual and image data. The current database system suggests a surgical determination for the cases with relatively small hippocampus and high signal intensity average on FLAIR images within the hippocampus. This indication matches the neurosurgeons expectations/observations. Moreover, as the database gets more populated with patient profiles and individual surgical outcomes, using data mining methods one may discover partially invisible correlations between the contents of different modalities of data and the outcome of the surgery.
NASA Astrophysics Data System (ADS)
Siadat, Mohammad-Reza; Soltanian-Zadeh, Hamid; Fotouhi, Farshad A.; Elisevich, Kost
2003-01-01
This paper presents the development of a human brain multimedia database for surgical candidacy determination in temporal lobe epilepsy. The focus of the paper is on content-based image management, navigation and retrieval. Several medical image-processing methods including our newly developed segmentation method are utilized for information extraction/correlation and indexing. The input data includes T1-, T2-Weighted MRI and FLAIR MRI and ictal and interictal SPECT modalities with associated clinical data and EEG data analysis. The database can answer queries regarding issues such as the correlation between the attribute X of the entity Y and the outcome of a temporal lobe epilepsy surgery. The entity Y can be a brain anatomical structure such as the hippocampus. The attribute X can be either a functionality feature of the anatomical structure Y, calculated with SPECT modalities, such as signal average, or a volumetric/morphological feature of the entity Y such as volume or average curvature. The outcome of the surgery can be any surgery assessment such as memory quotient. A determination is made regarding surgical candidacy by analysis of both textual and image data. The current database system suggests a surgical determination for the cases with relatively small hippocampus and high signal intensity average on FLAIR images within the hippocampus. This indication pretty much fits with the surgeons" expectations/observations. Moreover, as the database gets more populated with patient profiles and individual surgical outcomes, using data mining methods one may discover partially invisible correlations between the contents of different modalities of data and the outcome of the surgery.
Horizontal decomposition of data table for finding one reduct
NASA Astrophysics Data System (ADS)
Hońko, Piotr
2018-04-01
Attribute reduction, being one of the most essential tasks in rough set theory, is a challenge for data that does not fit in the available memory. This paper proposes new definitions of attribute reduction using horizontal data decomposition. Algorithms for computing superreduct and subsequently exact reducts of a data table are developed and experimentally verified. In the proposed approach, the size of subtables obtained during the decomposition can be arbitrarily small. Reducts of the subtables are computed independently from one another using any heuristic method for finding one reduct. Compared with standard attribute reduction methods, the proposed approach can produce superreducts that usually inconsiderably differ from an exact reduct. The approach needs comparable time and much less memory to reduce the attribute set. The method proposed for removing unnecessary attributes from superreducts executes relatively fast for bigger databases.
Mindfulness in nursing: an evolutionary concept analysis.
White, Lacie
2014-02-01
To report an analysis of the concept of mindfulness. Mindfulness is an emerging concept in health care that has significant implications for a variety of clinical populations. Nursing uses this concept in limited ways, and subsequently requires conceptual clarity to further identify its significance, use and applications in nursing. Mindfulness was explored using Rodgers evolutionary method of concept analysis. For this analysis, a sample of 59 English theoretical and research-based articles from the Cumulative Index to Nursing and Allied Health Literature database were obtained. The search was conducted between all-inclusive years of the database, 1981-2012. Data were analysed with particular focus on the attributes, antecedents, consequences, references and related terms that arose in relation to mindfulness in the nursing literature. The analysis found five intricately connected attributes: mindfulness is a transformative process where one develops an increasing ability to 'experience being present', with 'acceptance', 'attention' and 'awareness'. Antecedents, attributes and consequences appeared to inform and strengthen one another over time. Mindfulness is a significant concept for the discipline of nursing with practical applications for nurse well-being, the development and sustainability of therapeutic nursing qualities and holistic health promotion. It is imperative that nurse well-being and self-care become a more prominent focus in nursing research and education. Further development of the concept of mindfulness could support this focus, particularly through rigorous qualitative methodologies. © 2013 John Wiley & Sons Ltd.
The STEP database through the end-users eyes--USABILITY STUDY.
Salunke, Smita; Tuleu, Catherine
2015-08-15
The user-designed database of Safety and Toxicity of Excipients for Paediatrics ("STEP") is created to address the shared need of drug development community to access the relevant information of excipients effortlessly. Usability testing was performed to validate if the database satisfies the need of the end-users. Evaluation framework was developed to assess the usability. The participants performed scenario based tasks and provided feedback and post-session usability ratings. Failure Mode Effect Analysis (FMEA) was performed to prioritize the problems and improvements to the STEP database design and functionalities. The study revealed several design vulnerabilities. Tasks such as limiting the results, running complex queries, location of data and registering to access the database were challenging. The three critical attributes identified to have impact on the usability of the STEP database included (1) content and presentation (2) the navigation and search features (3) potential end-users. Evaluation framework proved to be an effective method for evaluating database effectiveness and user satisfaction. This study provides strong initial support for the usability of the STEP database. Recommendations would be incorporated into the refinement of the database to improve its usability and increase user participation towards the advancement of the database. Copyright © 2015 Elsevier B.V. All rights reserved.
Basic level scene understanding: categories, attributes and structures
Xiao, Jianxiong; Hays, James; Russell, Bryan C.; Patterson, Genevieve; Ehinger, Krista A.; Torralba, Antonio; Oliva, Aude
2013-01-01
A longstanding goal of computer vision is to build a system that can automatically understand a 3D scene from a single image. This requires extracting semantic concepts and 3D information from 2D images which can depict an enormous variety of environments that comprise our visual world. This paper summarizes our recent efforts toward these goals. First, we describe the richly annotated SUN database which is a collection of annotated images spanning 908 different scene categories with object, attribute, and geometric labels for many scenes. This database allows us to systematically study the space of scenes and to establish a benchmark for scene and object recognition. We augment the categorical SUN database with 102 scene attributes for every image and explore attribute recognition. Finally, we present an integrated system to extract the 3D structure of the scene and objects depicted in an image. PMID:24009590
NASA Astrophysics Data System (ADS)
Lee, Sangho; Suh, Jangwon; Park, Hyeong-Dong
2015-03-01
Boring logs are widely used in geological field studies since the data describes various attributes of underground and surface environments. However, it is difficult to manage multiple boring logs in the field as the conventional management and visualization methods are not suitable for integrating and combining large data sets. We developed an iPad application to enable its user to search the boring log rapidly and visualize them using the augmented reality (AR) technique. For the development of the application, a standard borehole database appropriate for a mobile-based borehole database management system was designed. The application consists of three modules: an AR module, a map module, and a database module. The AR module superimposes borehole data on camera imagery as viewed by the user and provides intuitive visualization of borehole locations. The map module shows the locations of corresponding borehole data on a 2D map with additional map layers. The database module provides data management functions for large borehole databases for other modules. Field survey was also carried out using more than 100,000 borehole data.
Lin, Ying-Chi; Wang, Chia-Chi; Chen, Ih-Sheng; Jheng, Jhao-Liang; Li, Jih-Heng; Tung, Chun-Wei
2013-01-01
The unique geographic features of Taiwan are attributed to the rich indigenous and endemic plant species in Taiwan. These plants serve as resourceful bank for biologically active phytochemicals. Given that these plant-derived chemicals are prototypes of potential drugs for diseases, databases connecting the chemical structures and pharmacological activities may facilitate drug development. To enhance the utility of the data, it is desirable to develop a database of chemical compounds and corresponding activities from indigenous plants in Taiwan. A database of anticancer, antiplatelet, and antituberculosis phytochemicals from indigenous plants in Taiwan was constructed. The database, TIPdb, is composed of a standardized format of published anticancer, antiplatelet, and antituberculosis phytochemicals from indigenous plants in Taiwan. A browse function was implemented for users to browse the database in a taxonomy-based manner. Search functions can be utilized to filter records of interest by botanical name, part, chemical class, or compound name. The structured and searchable database TIPdb was constructed to serve as a comprehensive and standardized resource for anticancer, antiplatelet, and antituberculosis compounds search. The manually curated chemical structures and activities provide a great opportunity to develop quantitative structure-activity relationship models for the high-throughput screening of potential anticancer, antiplatelet, and antituberculosis drugs.
Lin, Ying-Chi; Wang, Chia-Chi; Chen, Ih-Sheng; Jheng, Jhao-Liang; Li, Jih-Heng; Tung, Chun-Wei
2013-01-01
The unique geographic features of Taiwan are attributed to the rich indigenous and endemic plant species in Taiwan. These plants serve as resourceful bank for biologically active phytochemicals. Given that these plant-derived chemicals are prototypes of potential drugs for diseases, databases connecting the chemical structures and pharmacological activities may facilitate drug development. To enhance the utility of the data, it is desirable to develop a database of chemical compounds and corresponding activities from indigenous plants in Taiwan. A database of anticancer, antiplatelet, and antituberculosis phytochemicals from indigenous plants in Taiwan was constructed. The database, TIPdb, is composed of a standardized format of published anticancer, antiplatelet, and antituberculosis phytochemicals from indigenous plants in Taiwan. A browse function was implemented for users to browse the database in a taxonomy-based manner. Search functions can be utilized to filter records of interest by botanical name, part, chemical class, or compound name. The structured and searchable database TIPdb was constructed to serve as a comprehensive and standardized resource for anticancer, antiplatelet, and antituberculosis compounds search. The manually curated chemical structures and activities provide a great opportunity to develop quantitative structure-activity relationship models for the high-throughput screening of potential anticancer, antiplatelet, and antituberculosis drugs. PMID:23766708
Demirkus, Meltem; Precup, Doina; Clark, James J; Arbel, Tal
2016-06-01
Recent literature shows that facial attributes, i.e., contextual facial information, can be beneficial for improving the performance of real-world applications, such as face verification, face recognition, and image search. Examples of face attributes include gender, skin color, facial hair, etc. How to robustly obtain these facial attributes (traits) is still an open problem, especially in the presence of the challenges of real-world environments: non-uniform illumination conditions, arbitrary occlusions, motion blur and background clutter. What makes this problem even more difficult is the enormous variability presented by the same subject, due to arbitrary face scales, head poses, and facial expressions. In this paper, we focus on the problem of facial trait classification in real-world face videos. We have developed a fully automatic hierarchical and probabilistic framework that models the collective set of frame class distributions and feature spatial information over a video sequence. The experiments are conducted on a large real-world face video database that we have collected, labelled and made publicly available. The proposed method is flexible enough to be applied to any facial classification problem. Experiments on a large, real-world video database McGillFaces [1] of 18,000 video frames reveal that the proposed framework outperforms alternative approaches, by up to 16.96 and 10.13%, for the facial attributes of gender and facial hair, respectively.
Bergamino, Maurizio; Hamilton, David J; Castelletti, Lara; Barletta, Laura; Castellan, Lucio
2015-03-01
In this study, we describe the development and utilization of a relational database designed to manage the clinical and radiological data of patients with brain tumors. The Brain Tumor Database was implemented using MySQL v.5.0, while the graphical user interface was created using PHP and HTML, thus making it easily accessible through a web browser. This web-based approach allows for multiple institutions to potentially access the database. The BT Database can record brain tumor patient information (e.g. clinical features, anatomical attributes, and radiological characteristics) and be used for clinical and research purposes. Analytic tools to automatically generate statistics and different plots are provided. The BT Database is a free and powerful user-friendly tool with a wide range of possible clinical and research applications in neurology and neurosurgery. The BT Database graphical user interface source code and manual are freely available at http://tumorsdatabase.altervista.org. © The Author(s) 2013.
Heterogeneous Face Attribute Estimation: A Deep Multi-Task Learning Approach.
Han, Hu; K Jain, Anil; Shan, Shiguang; Chen, Xilin
2017-08-10
Face attribute estimation has many potential applications in video surveillance, face retrieval, and social media. While a number of methods have been proposed for face attribute estimation, most of them did not explicitly consider the attribute correlation and heterogeneity (e.g., ordinal vs. nominal and holistic vs. local) during feature representation learning. In this paper, we present a Deep Multi-Task Learning (DMTL) approach to jointly estimate multiple heterogeneous attributes from a single face image. In DMTL, we tackle attribute correlation and heterogeneity with convolutional neural networks (CNNs) consisting of shared feature learning for all the attributes, and category-specific feature learning for heterogeneous attributes. We also introduce an unconstrained face database (LFW+), an extension of public-domain LFW, with heterogeneous demographic attributes (age, gender, and race) obtained via crowdsourcing. Experimental results on benchmarks with multiple face attributes (MORPH II, LFW+, CelebA, LFWA, and FotW) show that the proposed approach has superior performance compared to state of the art. Finally, evaluations on a public-domain face database (LAP) with a single attribute show that the proposed approach has excellent generalization ability.
Research on spatio-temporal database techniques for spatial information service
NASA Astrophysics Data System (ADS)
Zhao, Rong; Wang, Liang; Li, Yuxiang; Fan, Rongshuang; Liu, Ping; Li, Qingyuan
2007-06-01
Geographic data should be described by spatial, temporal and attribute components, but the spatio-temporal queries are difficult to be answered within current GIS. This paper describes research into the development and application of spatio-temporal data management system based upon GeoWindows GIS software platform which was developed by Chinese Academy of Surveying and Mapping (CASM). Faced the current and practical requirements of spatial information application, and based on existing GIS platform, one kind of spatio-temporal data model which integrates vector and grid data together was established firstly. Secondly, we solved out the key technique of building temporal data topology, successfully developed a suit of spatio-temporal database management system adopting object-oriented methods. The system provides the temporal data collection, data storage, data management and data display and query functions. Finally, as a case study, we explored the application of spatio-temporal data management system with the administrative region data of multi-history periods of China as the basic data. With all the efforts above, the GIS capacity of management and manipulation in aspect of time and attribute of GIS has been enhanced, and technical reference has been provided for the further development of temporal geographic information system (TGIS).
Nadkarni, Prakash M.; Brandt, Cynthia M.; Marenco, Luis
2000-01-01
The task of creating and maintaining a front end to a large institutional entity-attribute-value (EAV) database can be cumbersome when using traditional client-server technology. Switching to Web technology as a delivery vehicle solves some of these problems but introduces others. In particular, Web development environments tend to be primitive, and many features that client-server developers take for granted are missing. WebEAV is a generic framework for Web development that is intended to streamline the process of Web application development for databases having a significant EAV component. It also addresses some challenging user interface issues that arise when any complex system is created. The authors describe the architecture of WebEAV and provide an overview of its features with suitable examples. PMID:10887163
Qualitative Comparison of IGRA and ESRL Radiosonde Archived Databases
NASA Technical Reports Server (NTRS)
Walker, John R.
2014-01-01
Multiple databases of atmospheric profile information are freely available to individuals and groups such as the Natural Environments group. Two of the primary database archives provided by NOAA that are most frequently used are those from the Earth Science Research Laboratory (ESRL) and the Integrated Global Radiosonde Archive (IGRA). Inquiries have been made as to why one database is used as opposed to the other, yet to the best of knowledge, no formal comparison has been performed. The goal of this study is to provide a qualitative comparison of the ESRL and IGRA radiosonde databases. For part of this analyses, 14 upper air observation sites were selected. These sites all have the common attribute of having been used or are planned for use in the development of Range Reference Atmospheres (RRAs) in support of NASA's and DOD's current and future goals.
Determining root correspondence between previously and newly detected objects
Paglieroni, David W.; Beer, N Reginald
2014-06-17
A system that applies attribute and topology based change detection to networks of objects that were detected on previous scans of a structure, roadway, or area of interest. The attributes capture properties or characteristics of the previously detected objects, such as location, time of detection, size, elongation, orientation, etc. The topology of the network of previously detected objects is maintained in a constellation database that stores attributes of previously detected objects and implicitly captures the geometrical structure of the network. A change detection system detects change by comparing the attributes and topology of new objects detected on the latest scan to the constellation database of previously detected objects.
Smart Location Database - Service
The Smart Location Database (SLD) summarizes over 80 demographic, built environment, transit service, and destination accessibility attributes for every census block group in the United States. Future updates to the SLD will include additional attributes which summarize the relative location efficiency of a block group when compared to other block groups within the same metropolitan region. EPA also plans to periodically update attributes and add new attributes to reflect latest available data. A log of SLD updates is included in the SLD User Guide. See the user guide for a full description of data sources, data currency, and known limitations: https://edg.epa.gov/data/Public/OP/SLD/SLD_userguide.pdf
Smart Location Database - Download
The Smart Location Database (SLD) summarizes over 80 demographic, built environment, transit service, and destination accessibility attributes for every census block group in the United States. Future updates to the SLD will include additional attributes which summarize the relative location efficiency of a block group when compared to other block groups within the same metropolitan region. EPA also plans to periodically update attributes and add new attributes to reflect latest available data. A log of SLD updates is included in the SLD User Guide. See the user guide for a full description of data sources, data currency, and known limitations: https://edg.epa.gov/data/Public/OP/SLD/SLD_userguide.pdf
Artist Material BRDF Database for Computer Graphics Rendering
NASA Astrophysics Data System (ADS)
Ashbaugh, Justin C.
The primary goal of this thesis was to create a physical library of artist material samples. This collection provides necessary data for the development of a gonio-imaging system for use in museums to more accurately document their collections. A sample set was produced consisting of 25 panels and containing nearly 600 unique samples. Selected materials are representative of those commonly used by artists both past and present. These take into account the variability in visual appearance resulting from the materials and application techniques used. Five attributes of variability were identified including medium, color, substrate, application technique and overcoat. Combinations of these attributes were selected based on those commonly observed in museum collections and suggested by surveying experts in the field. For each sample material, image data is collected and used to measure an average bi-directional reflectance distribution function (BRDF). The results are available as a public-domain image and optical database of artist materials at art-si.org. Additionally, the database includes specifications for each sample along with other information useful for computer graphics rendering such as the rectified sample images and normal maps.
Wang, Yiji; Dix, Theodore
2017-03-01
This study examined processes that might account for why negatively emotional children are at high risk for externalizing behavior problems when raised by mothers with depressive symptoms. Because negative emotionality regulates adaptation to stress, we predicted that it would undermine children's adjustment to mothers' depressive symptoms by increasing child emotions likely to elicit reciprocal negativity from depressed mothers, bias negatively children's attributions about others, and activate difficult-to-control oppositional responses. In a large sample (N = 1,082) evaluated from 6 months to second grade, results showed that, when mothers had depressive symptoms early in the child's development, children who were high in negative emotionality-but not those who were low-displayed increased risk for externalizing problems in second grade. This risk reflected tendencies for negatively emotional children, when raised by mothers with depressive symptoms, to develop hostile attributions about others and poor self-regulation of the negativity these attributions promote. The findings suggest that, when mothers with depressive symptoms raise negatively emotional children, children's risk for externalizing behavior problems may reflect tendencies for high negative emotion in children and reciprocal negativity in the dyad to undermine the development of attributional and self-regulatory processes. (PsycINFO Database Record (c) 2017 APA, all rights reserved).
A review of data quality assessment methods for public health information systems.
Chen, Hong; Hailey, David; Wang, Ning; Yu, Ping
2014-05-14
High quality data and effective data quality assessment are required for accurately evaluating the impact of public health interventions and measuring public health outcomes. Data, data use, and data collection process, as the three dimensions of data quality, all need to be assessed for overall data quality assessment. We reviewed current data quality assessment methods. The relevant study was identified in major databases and well-known institutional websites. We found the dimension of data was most frequently assessed. Completeness, accuracy, and timeliness were the three most-used attributes among a total of 49 attributes of data quality. The major quantitative assessment methods were descriptive surveys and data audits, whereas the common qualitative assessment methods were interview and documentation review. The limitations of the reviewed studies included inattentiveness to data use and data collection process, inconsistency in the definition of attributes of data quality, failure to address data users' concerns and a lack of systematic procedures in data quality assessment. This review study is limited by the coverage of the databases and the breadth of public health information systems. Further research could develop consistent data quality definitions and attributes. More research efforts should be given to assess the quality of data use and the quality of data collection process.
A Review of Data Quality Assessment Methods for Public Health Information Systems
Chen, Hong; Hailey, David; Wang, Ning; Yu, Ping
2014-01-01
High quality data and effective data quality assessment are required for accurately evaluating the impact of public health interventions and measuring public health outcomes. Data, data use, and data collection process, as the three dimensions of data quality, all need to be assessed for overall data quality assessment. We reviewed current data quality assessment methods. The relevant study was identified in major databases and well-known institutional websites. We found the dimension of data was most frequently assessed. Completeness, accuracy, and timeliness were the three most-used attributes among a total of 49 attributes of data quality. The major quantitative assessment methods were descriptive surveys and data audits, whereas the common qualitative assessment methods were interview and documentation review. The limitations of the reviewed studies included inattentiveness to data use and data collection process, inconsistency in the definition of attributes of data quality, failure to address data users’ concerns and a lack of systematic procedures in data quality assessment. This review study is limited by the coverage of the databases and the breadth of public health information systems. Further research could develop consistent data quality definitions and attributes. More research efforts should be given to assess the quality of data use and the quality of data collection process. PMID:24830450
Code of Federal Regulations, 2011 CFR
2011-10-01
... Microsoft Open Database Connectivity (ODBC) standard. ODBC is a Windows technology that allows a database software package to import data from a database created using a different software package. We currently...-compatible format. All databases must be supported with adequate documentation on data attributes, SQL...
Document image database indexing with pictorial dictionary
NASA Astrophysics Data System (ADS)
Akbari, Mohammad; Azimi, Reza
2010-02-01
In this paper we introduce a new approach for information retrieval from Persian document image database without using Optical Character Recognition (OCR).At first an attribute called subword upper contour label is defined then, a pictorial dictionary is constructed based on this attribute for the subwords. By this approach we address two issues in document image retrieval: keyword spotting and retrieval according to the document similarities. The proposed methods have been evaluated on a Persian document image database. The results have proved the ability of this approach in document image information retrieval.
A Mediterranean coastal database for assessing the impacts of sea-level rise and associated hazards
NASA Astrophysics Data System (ADS)
Wolff, Claudia; Vafeidis, Athanasios T.; Muis, Sanne; Lincke, Daniel; Satta, Alessio; Lionello, Piero; Jimenez, Jose A.; Conte, Dario; Hinkel, Jochen
2018-03-01
We have developed a new coastal database for the Mediterranean basin that is intended for coastal impact and adaptation assessment to sea-level rise and associated hazards on a regional scale. The data structure of the database relies on a linear representation of the coast with associated spatial assessment units. Using information on coastal morphology, human settlements and administrative boundaries, we have divided the Mediterranean coast into 13 900 coastal assessment units. To these units we have spatially attributed 160 parameters on the characteristics of the natural and socio-economic subsystems, such as extreme sea levels, vertical land movement and number of people exposed to sea-level rise and extreme sea levels. The database contains information on current conditions and on plausible future changes that are essential drivers for future impacts, such as sea-level rise rates and socio-economic development. Besides its intended use in risk and impact assessment, we anticipate that the Mediterranean Coastal Database (MCD) constitutes a useful source of information for a wide range of coastal applications.
A Mediterranean coastal database for assessing the impacts of sea-level rise and associated hazards
Wolff, Claudia; Vafeidis, Athanasios T.; Muis, Sanne; Lincke, Daniel; Satta, Alessio; Lionello, Piero; Jimenez, Jose A.; Conte, Dario; Hinkel, Jochen
2018-01-01
We have developed a new coastal database for the Mediterranean basin that is intended for coastal impact and adaptation assessment to sea-level rise and associated hazards on a regional scale. The data structure of the database relies on a linear representation of the coast with associated spatial assessment units. Using information on coastal morphology, human settlements and administrative boundaries, we have divided the Mediterranean coast into 13 900 coastal assessment units. To these units we have spatially attributed 160 parameters on the characteristics of the natural and socio-economic subsystems, such as extreme sea levels, vertical land movement and number of people exposed to sea-level rise and extreme sea levels. The database contains information on current conditions and on plausible future changes that are essential drivers for future impacts, such as sea-level rise rates and socio-economic development. Besides its intended use in risk and impact assessment, we anticipate that the Mediterranean Coastal Database (MCD) constitutes a useful source of information for a wide range of coastal applications. PMID:29583140
Semi-automatic feedback using concurrence between mixture vectors for general databases
NASA Astrophysics Data System (ADS)
Larabi, Mohamed-Chaker; Richard, Noel; Colot, Olivier; Fernandez-Maloigne, Christine
2001-12-01
This paper describes how a query system can exploit the basic knowledge by employing semi-automatic relevance feedback to refine queries and runtimes. For general databases, it is often useless to call complex attributes, because we have not sufficient information about images in the database. Moreover, these images can be topologically very different from one to each other and an attribute that is powerful for a database category may be very powerless for the other categories. The idea is to use very simple features, such as color histogram, correlograms, Color Coherence Vectors (CCV), to fill out the signature vector. Then, a number of mixture vectors is prepared depending on the number of very distinctive categories in the database. Knowing that a mixture vector is a vector containing the weight of each attribute that will be used to compute a similarity distance. We post a query in the database using successively all the mixture vectors defined previously. We retain then the N first images for each vector in order to make a mapping using the following information: Is image I present in several mixture vectors results? What is its rank in the results? These informations allow us to switch the system on an unsupervised relevance feedback or user's feedback (supervised feedback).
Information support of monitoring of technical condition of buildings in construction risk area
NASA Astrophysics Data System (ADS)
Skachkova, M. E.; Lepihina, O. Y.; Ignatova, V. V.
2018-05-01
The paper presents the results of the research devoted to the development of a model of information support of monitoring buildings technical condition; these buildings are located in the construction risk area. As a result of the visual and instrumental survey, as well as the analysis of existing approaches and techniques, attributive and cartographic databases have been created. These databases allow monitoring defects and damages of buildings located in a 30-meter risk area from the object under construction. The classification of structures and defects of these buildings under survey is presented. The functional capabilities of the developed model and the field of it practical applications are determined.
Point pattern match-based change detection in a constellation of previously detected objects
Paglieroni, David W.
2016-06-07
A method and system is provided that applies attribute- and topology-based change detection to objects that were detected on previous scans of a medium. The attributes capture properties or characteristics of the previously detected objects, such as location, time of detection, detection strength, size, elongation, orientation, etc. The locations define a three-dimensional network topology forming a constellation of previously detected objects. The change detection system stores attributes of the previously detected objects in a constellation database. The change detection system detects changes by comparing the attributes and topological consistency of newly detected objects encountered during a new scan of the medium to previously detected objects in the constellation database. The change detection system may receive the attributes of the newly detected objects as the objects are detected by an object detection system in real time.
MIPS: analysis and annotation of proteins from whole genomes
Mewes, H. W.; Amid, C.; Arnold, R.; Frishman, D.; Güldener, U.; Mannhaupt, G.; Münsterkötter, M.; Pagel, P.; Strack, N.; Stümpflen, V.; Warfsmann, J.; Ruepp, A.
2004-01-01
The Munich Information Center for Protein Sequences (MIPS-GSF), Neuherberg, Germany, provides protein sequence-related information based on whole-genome analysis. The main focus of the work is directed toward the systematic organization of sequence-related attributes as gathered by a variety of algorithms, primary information from experimental data together with information compiled from the scientific literature. MIPS maintains automatically generated and manually annotated genome-specific databases, develops systematic classification schemes for the functional annotation of protein sequences and provides tools for the comprehensive analysis of protein sequences. This report updates the information on the yeast genome (CYGD), the Neurospora crassa genome (MNCDB), the database of complete cDNAs (German Human Genome Project, NGFN), the database of mammalian protein–protein interactions (MPPI), the database of FASTA homologies (SIMAP), and the interface for the fast retrieval of protein-associated information (QUIPOS). The Arabidopsis thaliana database, the rice database, the plant EST databases (MATDB, MOsDB, SPUTNIK), as well as the databases for the comprehensive set of genomes (PEDANT genomes) are described elsewhere in the 2003 and 2004 NAR database issues, respectively. All databases described, and the detailed descriptions of our projects can be accessed through the MIPS web server (http://mips.gsf.de). PMID:14681354
MIPS: analysis and annotation of proteins from whole genomes.
Mewes, H W; Amid, C; Arnold, R; Frishman, D; Güldener, U; Mannhaupt, G; Münsterkötter, M; Pagel, P; Strack, N; Stümpflen, V; Warfsmann, J; Ruepp, A
2004-01-01
The Munich Information Center for Protein Sequences (MIPS-GSF), Neuherberg, Germany, provides protein sequence-related information based on whole-genome analysis. The main focus of the work is directed toward the systematic organization of sequence-related attributes as gathered by a variety of algorithms, primary information from experimental data together with information compiled from the scientific literature. MIPS maintains automatically generated and manually annotated genome-specific databases, develops systematic classification schemes for the functional annotation of protein sequences and provides tools for the comprehensive analysis of protein sequences. This report updates the information on the yeast genome (CYGD), the Neurospora crassa genome (MNCDB), the database of complete cDNAs (German Human Genome Project, NGFN), the database of mammalian protein-protein interactions (MPPI), the database of FASTA homologies (SIMAP), and the interface for the fast retrieval of protein-associated information (QUIPOS). The Arabidopsis thaliana database, the rice database, the plant EST databases (MATDB, MOsDB, SPUTNIK), as well as the databases for the comprehensive set of genomes (PEDANT genomes) are described elsewhere in the 2003 and 2004 NAR database issues, respectively. All databases described, and the detailed descriptions of our projects can be accessed through the MIPS web server (http://mips.gsf.de).
Ferreira Junior, José Raniery; Oliveira, Marcelo Costa; de Azevedo-Marques, Paulo Mazzoncini
2016-12-01
Lung cancer is the leading cause of cancer-related deaths in the world, and its main manifestation is pulmonary nodules. Detection and classification of pulmonary nodules are challenging tasks that must be done by qualified specialists, but image interpretation errors make those tasks difficult. In order to aid radiologists on those hard tasks, it is important to integrate the computer-based tools with the lesion detection, pathology diagnosis, and image interpretation processes. However, computer-aided diagnosis research faces the problem of not having enough shared medical reference data for the development, testing, and evaluation of computational methods for diagnosis. In order to minimize this problem, this paper presents a public nonrelational document-oriented cloud-based database of pulmonary nodules characterized by 3D texture attributes, identified by experienced radiologists and classified in nine different subjective characteristics by the same specialists. Our goal with the development of this database is to improve computer-aided lung cancer diagnosis and pulmonary nodule detection and classification research through the deployment of this database in a cloud Database as a Service framework. Pulmonary nodule data was provided by the Lung Image Database Consortium and Image Database Resource Initiative (LIDC-IDRI), image descriptors were acquired by a volumetric texture analysis, and database schema was developed using a document-oriented Not only Structured Query Language (NoSQL) approach. The proposed database is now with 379 exams, 838 nodules, and 8237 images, 4029 of them are CT scans and 4208 manually segmented nodules, and it is allocated in a MongoDB instance on a cloud infrastructure.
1984-12-01
OF ENGI- E V ALY UNCLASSIFIED DEC 84 AFIT/CS/MATH/84D-4 F /G 9/2 N somhhmmhlm EohhmhohhohhEE mhhhhhhmhhmhln LI .~ I 28 111251.0 111IL25 1111-4 11...nontransitively dependent on the primary key, then the relation * - R is in the third normal form (3NF). •~ 8 A relation R is in the Boyce/ Codd Normal Form...with attributes AIA2 ... An and functional dependencies F , and if X is a subset of the attributes, then X is a key of R if: 1. X --> AIA A is in the
Fenelon, Joseph M.
2006-01-01
More than 1,200 water-level measurements from 1957 to 2005 in the Rainier Mesa area of the Nevada Test Site were quality assured and analyzed. Water levels were measured from 50 discrete intervals within 18 boreholes and from 4 tunnel sites. An interpretive database was constructed that describes water-level conditions for each water level measured in the Rainier Mesa area. Multiple attributes were assigned to each water-level measurement in the database to describe the hydrologic conditions at the time of measurement. General quality, temporal variability, regional significance, and hydrologic conditions are attributed for each water-level measurement. The database also includes hydrograph narratives that describe the water-level history of each well.
Access control based on attribute certificates for medical intranet applications.
Mavridis, I; Georgiadis, C; Pangalos, G; Khair, M
2001-01-01
Clinical information systems frequently use intranet and Internet technologies. However these technologies have emphasized sharing and not security, despite the sensitive and private nature of much health information. Digital certificates (electronic documents which recognize an entity or its attributes) can be used to control access in clinical intranet applications. To outline the need for access control in distributed clinical database systems, to describe the use of digital certificates and security policies, and to propose the architecture for a system using digital certificates, cryptography and security policy to control access to clinical intranet applications. We have previously developed a security policy, DIMEDAC (Distributed Medical Database Access Control), which is compatible with emerging public key and privilege management infrastructure. In our implementation approach we propose the use of digital certificates, to be used in conjunction with DIMEDAC. Our proposed access control system consists of two phases: the ways users gain their security credentials; and how these credentials are used to access medical data. Three types of digital certificates are used: identity certificates for authentication; attribute certificates for authorization; and access-rule certificates for propagation of access control policy. Once a user is identified and authenticated, subsequent access decisions are based on a combination of identity and attribute certificates, with access-rule certificates providing the policy framework. Access control in clinical intranet applications can be successfully and securely managed through the use of digital certificates and the DIMEDAC security policy.
Footprint Representation of Planetary Remote Sensing Data
NASA Astrophysics Data System (ADS)
Walter, S. H. G.; Gasselt, S. V.; Michael, G.; Neukum, G.
The geometric outline of remote sensing image data, the so called footprint, can be represented as a number of coordinate tuples. These polygons are associated with according attribute information such as orbit name, ground- and image resolution, solar longitude and illumination conditions to generate a powerful base for classification of planetary experiment data. Speed, handling and extended capabilites are the reasons for using geodatabases to store and access these data types. Techniques for such a spatial database of footprint data are demonstrated using the Relational Database Management System (RDBMS) PostgreSQL, spatially enabled by the PostGIS extension. Exemplary, footprints of the HRSC and OMEGA instruments, both onboard ESA's Mars Express Orbiter, are generated and connected to attribute information. The aim is to provide high-resolution footprints of the OMEGA instrument to the science community for the first time and make them available for web-based mapping applications like the "Planetary Interactive GIS-on-the-Web Analyzable Database" (PIG- WAD), produced by the USGS. Map overlays with HRSC or other instruments like MOC and THEMIS (footprint maps are already available for these instruments and can be integrated into the database) allow on-the-fly intersection and comparison as well as extended statistics of the data. Footprint polygons are generated one by one using standard software provided by the instrument teams. Attribute data is calculated and stored together with the geometric information. In the case of HRSC, the coordinates of the footprints are already available in the VICAR label of each image file. Using the VICAR RTL and PostgreSQL's libpq C library they are loaded into the database using the Well-Known Text (WKT) notation by the Open Geospatial Consortium, Inc. (OGC). For the OMEGA instrument, image data is read using IDL routines developed and distributed by the OMEGA team. Image outlines are exported together with relevant attribute data to the industry standard Shapefile format. These files are translated to a Structured Query Language (SQL) command sequence suitable for insertion into the PostGIS/PostgrSQL database using the shp2pgsql data loader provided by the PostGIS software. PostgreSQL's advanced features such as geometry types, rules, operators and functions allow complex spatial queries and on-the-fly processing of data on DBMS level e.g. generalisation of the outlines. Processing done by the DBMS, visualisation via GIS systems and utilisation for web-based applications like mapservers will be demonstrated.
A probabilistic NF2 relational algebra for integrated information retrieval and database systems
DOE Office of Scientific and Technical Information (OSTI.GOV)
Fuhr, N.; Roelleke, T.
The integration of information retrieval (IR) and database systems requires a data model which allows for modelling documents as entities, representing uncertainty and vagueness and performing uncertain inference. For this purpose, we present a probabilistic data model based on relations in non-first-normal-form (NF2). Here, tuples are assigned probabilistic weights giving the probability that a tuple belongs to a relation. Thus, the set of weighted index terms of a document are represented as a probabilistic subrelation. In a similar way, imprecise attribute values are modelled as a set-valued attribute. We redefine the relational operators for this type of relations such thatmore » the result of each operator is again a probabilistic NF2 relation, where the weight of a tuple gives the probability that this tuple belongs to the result. By ordering the tuples according to decreasing probabilities, the model yields a ranking of answers like in most IR models. This effect also can be used for typical database queries involving imprecise attribute values as well as for combinations of database and IR queries.« less
Garcia, Adriana; Masbruch, Melissa D.; Susong, David D.
2014-01-01
The U.S. Geological Survey, as part of the Department of the Interior’s WaterSMART (Sustain and Manage America’s Resources for Tomorrow) initiative, compiled published estimates of groundwater discharge to streams in the Upper Colorado River Basin as a geospatial database. For the purpose of this report, groundwater discharge to streams is the baseflow portion of streamflow that includes contributions of groundwater from various flow paths. Reported estimates of groundwater discharge were assigned as attributes to stream reaches derived from the high-resolution National Hydrography Dataset. A total of 235 estimates of groundwater discharge to streams were compiled and included in the dataset. Feature class attributes of the geospatial database include groundwater discharge (acre-feet per year), method of estimation, citation abbreviation, defined reach, and 8-digit hydrologic unit code(s). Baseflow index (BFI) estimates of groundwater discharge were calculated using an existing streamflow characteristics dataset and were included as an attribute in the geospatial database. A comparison of the BFI estimates to the compiled estimates of groundwater discharge found that the BFI estimates were greater than the reported groundwater discharge estimates.
A novel data storage logic in the cloud
Mátyás, Bence; Szarka, Máté; Járvás, Gábor; Kusper, Gábor; Argay, István; Fialowski, Alice
2016-01-01
Databases which store and manage long-term scientific information related to life science are used to store huge amount of quantitative attributes. Introduction of a new entity attribute requires modification of the existing data tables and the programs that use these data tables. The solution is increasing the virtual data tables while the number of screens remains the same. The main objective of the present study was to introduce a logic called Joker Tao (JT) which provides universal data storage for cloud-based databases. It means all types of input data can be interpreted as an entity and attribute at the same time, in the same data table. PMID:29026521
A novel data storage logic in the cloud.
Mátyás, Bence; Szarka, Máté; Járvás, Gábor; Kusper, Gábor; Argay, István; Fialowski, Alice
2016-01-01
Databases which store and manage long-term scientific information related to life science are used to store huge amount of quantitative attributes. Introduction of a new entity attribute requires modification of the existing data tables and the programs that use these data tables. The solution is increasing the virtual data tables while the number of screens remains the same. The main objective of the present study was to introduce a logic called Joker Tao (JT) which provides universal data storage for cloud-based databases. It means all types of input data can be interpreted as an entity and attribute at the same time, in the same data table.
VOMS/VOMRS utilization patterns and convergence plan
DOE Office of Scientific and Technical Information (OSTI.GOV)
Ceccanti, A.; /INFN, CNAF; Ciaschini, V.
2010-01-01
The Grid community uses two well-established registration services, which allow users to be authenticated under the auspices of Virtual Organizations (VOs). The Virtual Organization Membership Service (VOMS), developed in the context of the Enabling Grid for E-sciencE (EGEE) project, is an Attribute Authority service that issues attributes expressing membership information of a subject within a VO. VOMS allows to partition users in groups, assign them roles and free-form attributes which are then used to drive authorization decisions. The VOMS administrative application, VOMS-Admin, manages and populates the VOMS database with membership information. The Virtual Organization Management Registration Service (VOMRS), developed atmore » Fermilab, extends the basic registration and management functionalities present in VOMS-Admin. It implements a registration workflow that requires VO usage policy acceptance and membership approval by administrators. VOMRS supports management of multiple grid certificates, and handling users' request for group and role assignments, and membership status. VOMRS is capable of interfacing to local systems with personnel information (e.g. the CERN Human Resource Database) and of pulling relevant member information from them. VOMRS synchronizes the relevant subset of information with VOMS. The recent development of new features in VOMS-Admin raises the possibility of rationalizing the support and converging on a single solution by continuing and extending existing collaborations between EGEE and OSG. Such strategy is supported by WLCG, OSG, US CMS, US Atlas, and other stakeholders worldwide. In this paper, we will analyze features in use by major experiments and the use cases for registration addressed by the mature single solution.« less
VOMS/VOMRS utilization patterns and convergence plan
NASA Astrophysics Data System (ADS)
Ceccanti, A.; Ciaschini, V.; Dimou, M.; Garzoglio, G.; Levshina, T.; Traylen, S.; Venturi, V.
2010-04-01
The Grid community uses two well-established registration services, which allow users to be authenticated under the auspices of Virtual Organizations (VOs). The Virtual Organization Membership Service (VOMS), developed in the context of the Enabling Grid for E-sciencE (EGEE) project, is an Attribute Authority service that issues attributes expressing membership information of a subject within a VO. VOMS allows to partition users in groups, assign them roles and free-form attributes which are then used to drive authorization decisions. The VOMS administrative application, VOMS-Admin, manages and populates the VOMS database with membership information. The Virtual Organization Management Registration Service (VOMRS), developed at Fermilab, extends the basic registration and management functionalities present in VOMS-Admin. It implements a registration workflow that requires VO usage policy acceptance and membership approval by administrators. VOMRS supports management of multiple grid certificates, and handling users' request for group and role assignments, and membership status. VOMRS is capable of interfacing to local systems with personnel information (e.g. the CERN Human Resource Database) and of pulling relevant member information from them. VOMRS synchronizes the relevant subset of information with VOMS. The recent development of new features in VOMS-Admin raises the possibility of rationalizing the support and converging on a single solution by continuing and extending existing collaborations between EGEE and OSG. Such strategy is supported by WLCG, OSG, US CMS, US Atlas, and other stakeholders worldwide. In this paper, we will analyze features in use by major experiments and the use cases for registration addressed by the mature single solution.
Analysis of Landslide Hazard Impact Using the Landslide Database for Germany
NASA Astrophysics Data System (ADS)
Klose, M.; Damm, B.
2014-12-01
The Federal Republic of Germany has long been among the few European countries that lack a national landslide database. Systematic collection and inventory of landslide data still shows a comprehensive research history in Germany, but only one focused on development of databases with local or regional coverage. This has changed in recent years with the launch of a database initiative aimed at closing the data gap existing at national level. The present contribution reports on this project that is based on a landslide database which evolved over the last 15 years to a database covering large parts of Germany. A strategy of systematic retrieval, extraction, and fusion of landslide data is at the heart of the methodology, providing the basis for a database with a broad potential of application. The database offers a data pool of more than 4,200 landslide data sets with over 13,000 single data files and dates back to 12th century. All types of landslides are covered by the database, which stores not only core attributes, but also various complementary data, including data on landslide causes, impacts, and mitigation. The current database migration to PostgreSQL/PostGIS is focused on unlocking the full scientific potential of the database, while enabling data sharing and knowledge transfer via a web GIS platform. In this contribution, the goals and the research strategy of the database project are highlighted at first, with a summary of best practices in database development providing perspective. Next, the focus is on key aspects of the methodology, which is followed by the results of different case studies in the German Central Uplands. The case study results exemplify database application in analysis of vulnerability to landslides, impact statistics, and hazard or cost modeling.
Federal Register 2010, 2011, 2012, 2013, 2014
2011-09-28
... Database Incorporating High-Cost Single-Family Securitized Loan Data Fields and Technical Data Field... single-family matrix in FHFA's Public Use Database (PUDB) to include data fields for the high-cost single... of loan attributes in FHFA's databases that could be used, singularly or in some combination, to...
The DREO Elint Browser Utility (DEBU) reference manual
NASA Astrophysics Data System (ADS)
Ford, Barbara; Jones, David
1992-04-01
An electronic intelligent database browsing tool called DEBU has been developed that allows databases such as ELP, Kilting, EWIR, and AFEWC to be reviewed and analyzed from a user-friendly environment on a personal computer. DEBU's basic function is to allow users to examine the contents of user-selected subfiles of user-selected emitters of user-selected databases. DEBU augments this functionality with support for selecting (filtering) and combining subsets of emitters by user-selected attributes such as name, parameter type, or parameter value. DEBU provides facilities for examining histograms and x-y plots of selected parameters, for doing ambiguity analysis and mode level analysis, and for generating and printing a variety of reports. A manual is provided for users of DEBU, including descriptions and illustrations of menus and windows.
Xia, Jun; Tashpolat, Tiyip; Zhang, Fei; Ji, Hong-jiang
2011-07-01
The characteristic of object spectrum is not only the base of the quantification analysis of remote sensing, but also the main content of the basic research of remote sensing. The typical surface object spectral database in arid areas oasis is of great significance for applied research on remote sensing in soil salinization. In the present paper, the authors took the Ugan-Kuqa River Delta Oasis as an example, unified .NET and the SuperMap platform with SQL Server database stored data, used the B/S pattern and the C# language to design and develop the typical surface object spectral information system, and established the typical surface object spectral database according to the characteristics of arid areas oasis. The system implemented the classified storage and the management of typical surface object spectral information and the related attribute data of the study areas; this system also implemented visualized two-way query between the maps and attribute data, the drawings of the surface object spectral response curves and the processing of the derivative spectral data and its drawings. In addition, the system initially possessed a simple spectral data mining and analysis capabilities, and this advantage provided an efficient, reliable and convenient data management and application platform for the Ugan-Kuqa River Delta Oasis's follow-up study in soil salinization. Finally, It's easy to maintain, convinient for secondary development and practically operating in good condition.
Chemical databases evaluated by order theoretical tools.
Voigt, Kristina; Brüggemann, Rainer; Pudenz, Stefan
2004-10-01
Data on environmental chemicals are urgently needed to comply with the future chemicals policy in the European Union. The availability of data on parameters and chemicals can be evaluated by chemometrical and environmetrical methods. Different mathematical and statistical methods are taken into account in this paper. The emphasis is set on a new, discrete mathematical method called METEOR (method of evaluation by order theory). Application of the Hasse diagram technique (HDT) of the complete data-matrix comprising 12 objects (databases) x 27 attributes (parameters + chemicals) reveals that ECOTOX (ECO), environmental fate database (EFD) and extoxnet (EXT)--also called multi-database databases--are best. Most single databases which are specialised are found in a minimal position in the Hasse diagram; these are biocatalysis/biodegradation database (BID), pesticide database (PES) and UmweltInfo (UMW). The aggregation of environmental parameters and chemicals (equal weight) leads to a slimmer data-matrix on the attribute side. However, no significant differences are found in the "best" and "worst" objects. The whole approach indicates a rather bad situation in terms of the availability of data on existing chemicals and hence an alarming signal concerning the new and existing chemicals policies of the EEC.
A spatial-temporal system for dynamic cadastral management.
Nan, Liu; Renyi, Liu; Guangliang, Zhu; Jiong, Xie
2006-03-01
A practical spatio-temporal database (STDB) technique for dynamic urban land management is presented. One of the STDB models, the expanded model of Base State with Amendments (BSA), is selected as the basis for developing the dynamic cadastral management technique. Two approaches, the Section Fast Indexing (SFI) and the Storage Factors of Variable Granularity (SFVG), are used to improve the efficiency of the BSA model. Both spatial graphic data and attribute data, through a succinct engine, are stored in standard relational database management systems (RDBMS) for the actual implementation of the BSA model. The spatio-temporal database is divided into three interdependent sub-databases: present DB, history DB and the procedures-tracing DB. The efficiency of database operation is improved by the database connection in the bottom layer of the Microsoft SQL Server. The spatio-temporal system can be provided at a low-cost while satisfying the basic needs of urban land management in China. The approaches presented in this paper may also be of significance to countries where land patterns change frequently or to agencies where financial resources are limited.
NASA Astrophysics Data System (ADS)
Ries, William; Langridge, Robert; Villamor, Pilar; Litchfield, Nicola; Van Dissen, Russ; Townsend, Dougal; Lee, Julie; Heron, David; Lukovic, Biljana
2014-05-01
In New Zealand, we are currently reconciling multiple digital coverages of mapped active faults into a national coverage at a single scale (1:250,000). This seems at first glance to be a relatively simple task. However, methods used to capture data, the scale of capture, and the initial purpose of the fault mapping, has produced datasets that have very different characteristics. The New Zealand digital active fault database (AFDB) was initially developed as a way of managing active fault locations and fault-related features within a computer-based spatial framework. The data contained within the AFDB comes from a wide range of studies, from plate tectonic (1:500,000) to cadastral (1:2,000) scale. The database was designed to allow capture of field observations and remotely sourced data without a loss in data resolution. This approach has worked well as a method for compiling a centralised database for fault information but not for providing a complete national coverage at a single scale. During the last 15 years other complementary projects have used and also contributed data to the AFDB, most notably the QMAP project (a national series of geological maps completed over 19 years that include coverage of active and inactive faults at 1:250,000). AFDB linework and attributes was incorporated into this series but simplification of linework and attributes has occurred to maintain map clarity at 1:250,000 scale. Also, during this period on-going mapping of active faults has improved upon these data. Other projects of note that have used data from the AFDB include the National Seismic Hazard Model of New Zealand and the Global Earthquake Model (GEM). The main goal of the current project has been to provide the best digital spatial representation of a fault trace at 1:250,000 scale and combine this with the most up to date attributes. In some areas this has required a simplification of very fine detailed data and in some cases new mapping to provide a complete coverage. Where datasets have conflicting line work and/or attributes, data was reviewed through consultation with authors or review of published research to ensure the most to date representation was maintained. The current project aims to provide a coverage that will be consistent between the AFDB and QMAP digital and provide a free download of these data on the AFDB website (http://data.gns.cri.nz/af/).
Access Control based on Attribute Certificates for Medical Intranet Applications
Georgiadis, Christos; Pangalos, George; Khair, Marie
2001-01-01
Background Clinical information systems frequently use intranet and Internet technologies. However these technologies have emphasized sharing and not security, despite the sensitive and private nature of much health information. Digital certificates (electronic documents which recognize an entity or its attributes) can be used to control access in clinical intranet applications. Objectives To outline the need for access control in distributed clinical database systems, to describe the use of digital certificates and security policies, and to propose the architecture for a system using digital certificates, cryptography and security policy to control access to clinical intranet applications. Methods We have previously developed a security policy, DIMEDAC (Distributed Medical Database Access Control), which is compatible with emerging public key and privilege management infrastructure. In our implementation approach we propose the use of digital certificates, to be used in conjunction with DIMEDAC. Results Our proposed access control system consists of two phases: the ways users gain their security credentials; and how these credentials are used to access medical data. Three types of digital certificates are used: identity certificates for authentication; attribute certificates for authorization; and access-rule certificates for propagation of access control policy. Once a user is identified and authenticated, subsequent access decisions are based on a combination of identity and attribute certificates, with access-rule certificates providing the policy framework. Conclusions Access control in clinical intranet applications can be successfully and securely managed through the use of digital certificates and the DIMEDAC security policy. PMID:11720951
SAADA: Astronomical Databases Made Easier
NASA Astrophysics Data System (ADS)
Michel, L.; Nguyen, H. N.; Motch, C.
2005-12-01
Many astronomers wish to share datasets with their community but have not enough manpower to develop databases having the functionalities required for high-level scientific applications. The SAADA project aims at automatizing the creation and deployment process of such databases. A generic but scientifically relevant data model has been designed which allows one to build databases by providing only a limited number of product mapping rules. Databases created by SAADA rely on a relational database supporting JDBC and covered by a Java layer including a lot of generated code. Such databases can simultaneously host spectra, images, source lists and plots. Data are grouped in user defined collections whose content can be seen as one unique set per data type even if their formats differ. Datasets can be correlated one with each other using qualified links. These links help, for example, to handle the nature of a cross-identification (e.g., a distance or a likelihood) or to describe their scientific content (e.g., by associating a spectrum to a catalog entry). The SAADA query engine is based on a language well suited to the data model which can handle constraints on linked data, in addition to classical astronomical queries. These constraints can be applied on the linked objects (number, class and attributes) and/or on the link qualifier values. Databases created by SAADA are accessed through a rich WEB interface or a Java API. We are currently developing an inter-operability module implanting VO protocols.
The MAR databases: development and implementation of databases specific for marine metagenomics
Klemetsen, Terje; Raknes, Inge A; Fu, Juan; Agafonov, Alexander; Balasundaram, Sudhagar V; Tartari, Giacomo; Robertsen, Espen
2018-01-01
Abstract We introduce the marine databases; MarRef, MarDB and MarCat (https://mmp.sfb.uit.no/databases/), which are publicly available resources that promote marine research and innovation. These data resources, which have been implemented in the Marine Metagenomics Portal (MMP) (https://mmp.sfb.uit.no/), are collections of richly annotated and manually curated contextual (metadata) and sequence databases representing three tiers of accuracy. While MarRef is a database for completely sequenced marine prokaryotic genomes, which represent a marine prokaryote reference genome database, MarDB includes all incomplete sequenced prokaryotic genomes regardless level of completeness. The last database, MarCat, represents a gene (protein) catalog of uncultivable (and cultivable) marine genes and proteins derived from marine metagenomics samples. The first versions of MarRef and MarDB contain 612 and 3726 records, respectively. Each record is built up of 106 metadata fields including attributes for sampling, sequencing, assembly and annotation in addition to the organism and taxonomic information. Currently, MarCat contains 1227 records with 55 metadata fields. Ontologies and controlled vocabularies are used in the contextual databases to enhance consistency. The user-friendly web interface lets the visitors browse, filter and search in the contextual databases and perform BLAST searches against the corresponding sequence databases. All contextual and sequence databases are freely accessible and downloadable from https://s1.sfb.uit.no/public/mar/. PMID:29106641
Digital representation of oil and natural gas well pad scars in southwest Wyoming: 2012 update
Garman, Steven L.; McBeth, Jamie L.
2015-01-01
The recent proliferation of oil and natural gas energy development in the Greater Green River Basin of southwest Wyoming has accentuated the need to understand wildlife responses to this development. The location and extent of surface disturbance that is created by oil and natural gas well pad scars are key pieces of information used to assess the effects of energy infrastructure on wildlife populations and habitat. A digital database of oil and natural gas pad scars had previously been generated from 1-meter (m) National Agriculture Imagery Program imagery (NAIP) acquired in 2009 for a 7.7-million hectare (ha) (19,026,700 acres) region of southwest Wyoming. Scars included the pad area where wellheads, pumps, and storage facilities reside and the surrounding area that was scraped and denuded of vegetation during the establishment of the pad. Scars containing tanks, compressors, the storage of oil and gas related equipment, and produced-water ponds were also collected on occasion. This report updates the digital database for the five counties of southwest Wyoming (Carbon, Lincoln, Sublette, Sweetwater, Uinta) within the Wyoming Landscape Conservation Initiative (WLCI) study area and for a limited portion of Fremont, Natrona, and Albany Counties using 2012 1-m NAIP imagery and 2012 oil and natural gas well permit information. This report adds pad scars created since 2009, and updates attributes of all pad scars using the 2012 well permit information. These attributes include the origination year of the pad scar, the number of active and inactive wells on or near each pad scar in 2012, and the overall status of the pad scar (active or inactive). The new 2012 database contains 17,404 pad scars of which 15,532 are attributed as oil and natural gas well pads. Digital data are stored as shapefiles projected to the Universal Transverse Mercator (zones 12 and 13) coordinate system. These data are available from the U.S. Geological Survey (USGS) at http://dx.doi.org/10.3133/ds934.
Development of the Global Earthquake Model’s neotectonic fault database
Christophersen, Annemarie; Litchfield, Nicola; Berryman, Kelvin; Thomas, Richard; Basili, Roberto; Wallace, Laura; Ries, William; Hayes, Gavin P.; Haller, Kathleen M.; Yoshioka, Toshikazu; Koehler, Richard D.; Clark, Dan; Wolfson-Schwehr, Monica; Boettcher, Margaret S.; Villamor, Pilar; Horspool, Nick; Ornthammarath, Teraphan; Zuñiga, Ramon; Langridge, Robert M.; Stirling, Mark W.; Goded, Tatiana; Costa, Carlos; Yeats, Robert
2015-01-01
The Global Earthquake Model (GEM) aims to develop uniform, openly available, standards, datasets and tools for worldwide seismic risk assessment through global collaboration, transparent communication and adapting state-of-the-art science. GEM Faulted Earth (GFE) is one of GEM’s global hazard module projects. This paper describes GFE’s development of a modern neotectonic fault database and a unique graphical interface for the compilation of new fault data. A key design principle is that of an electronic field notebook for capturing observations a geologist would make about a fault. The database is designed to accommodate abundant as well as sparse fault observations. It features two layers, one for capturing neotectonic faults and fold observations, and the other to calculate potential earthquake fault sources from the observations. In order to test the flexibility of the database structure and to start a global compilation, five preexisting databases have been uploaded to the first layer and two to the second. In addition, the GFE project has characterised the world’s approximately 55,000 km of subduction interfaces in a globally consistent manner as a basis for generating earthquake event sets for inclusion in earthquake hazard and risk modelling. Following the subduction interface fault schema and including the trace attributes of the GFE database schema, the 2500-km-long frontal thrust fault system of the Himalaya has also been characterised. We propose the database structure to be used widely, so that neotectonic fault data can make a more complete and beneficial contribution to seismic hazard and risk characterisation globally.
Wyllie, Aileen; DiGiacomo, Michelle; Jackson, Debra; Davidson, Patricia; Phillips, Jane
2016-10-01
To optimise the career development in early career academic nurses by providing an overview of the attributes necessary for success. Evidence of early prospective career planning is necessary to optimise success in the tertiary sector. This is particularly important for nurse academics given the profession's later entry into academia, the ageing nursing workforce and the continuing global shortage of nurses. A qualitative systematic review. Academic Search Complete, CINAHL, Medline, ERIC, Professional Development Collection and Google Scholar databases were searched; resulting in the inclusion of nine qualitative nurse-only focussed studies published between 2004 and 2014. The studies were critically appraised and the data thematically analysed. Three abilities were identified as important to the early career academic nurse: a willingness to adapt to change, an intention to pursue support and embodying resilience. These abilities give rise to attributes that are recommended as key to successful academic career development for those employed on a continuing academic basis. The capacity to rely on one's own capabilities is becoming seen as increasingly important. It is proposed that recognition of these attributes, their skilful application and monitoring outlined in the review are recommended for a successful career in academia. Crown Copyright © 2016. Published by Elsevier Ltd. All rights reserved.
EPA Tribal Areas (4 of 4): Alaska Native Allotments
This dataset is a spatial representation of the Public Land Survey System (PLSS) in Alaska, generated from land survey records. The data represents a seamless spatial portrayal of native allotment land parcels, their legal descriptions, corner positioning and markings, and survey measurements. This data is intended for mapping purposes only and is not a substitute or replacement for the legal land survey records or other legal documents.Measurement and attribute data are collected from survey records using data entry screens into a relational database. The database design is based upon the FGDC Cadastral Content Data Standard. Corner positions are derived by geodetic calculations using measurement records. Closure and edgematching are applied to produce a seamless dataset. The resultant features do not preserve the original geometry of survey measurements, but the record measurements are reported as attributes. Additional boundary data are derived by spatial capture, protraction and GIS processing. The spatial features are stored and managed within the relational database, with active links to the represented measurement and attribute data.
Wollbrett, Julien; Larmande, Pierre; de Lamotte, Frédéric; Ruiz, Manuel
2013-04-15
In recent years, a large amount of "-omics" data have been produced. However, these data are stored in many different species-specific databases that are managed by different institutes and laboratories. Biologists often need to find and assemble data from disparate sources to perform certain analyses. Searching for these data and assembling them is a time-consuming task. The Semantic Web helps to facilitate interoperability across databases. A common approach involves the development of wrapper systems that map a relational database schema onto existing domain ontologies. However, few attempts have been made to automate the creation of such wrappers. We developed a framework, named BioSemantic, for the creation of Semantic Web Services that are applicable to relational biological databases. This framework makes use of both Semantic Web and Web Services technologies and can be divided into two main parts: (i) the generation and semi-automatic annotation of an RDF view; and (ii) the automatic generation of SPARQL queries and their integration into Semantic Web Services backbones. We have used our framework to integrate genomic data from different plant databases. BioSemantic is a framework that was designed to speed integration of relational databases. We present how it can be used to speed the development of Semantic Web Services for existing relational biological databases. Currently, it creates and annotates RDF views that enable the automatic generation of SPARQL queries. Web Services are also created and deployed automatically, and the semantic annotations of our Web Services are added automatically using SAWSDL attributes. BioSemantic is downloadable at http://southgreen.cirad.fr/?q=content/Biosemantic.
2013-01-01
Background In recent years, a large amount of “-omics” data have been produced. However, these data are stored in many different species-specific databases that are managed by different institutes and laboratories. Biologists often need to find and assemble data from disparate sources to perform certain analyses. Searching for these data and assembling them is a time-consuming task. The Semantic Web helps to facilitate interoperability across databases. A common approach involves the development of wrapper systems that map a relational database schema onto existing domain ontologies. However, few attempts have been made to automate the creation of such wrappers. Results We developed a framework, named BioSemantic, for the creation of Semantic Web Services that are applicable to relational biological databases. This framework makes use of both Semantic Web and Web Services technologies and can be divided into two main parts: (i) the generation and semi-automatic annotation of an RDF view; and (ii) the automatic generation of SPARQL queries and their integration into Semantic Web Services backbones. We have used our framework to integrate genomic data from different plant databases. Conclusions BioSemantic is a framework that was designed to speed integration of relational databases. We present how it can be used to speed the development of Semantic Web Services for existing relational biological databases. Currently, it creates and annotates RDF views that enable the automatic generation of SPARQL queries. Web Services are also created and deployed automatically, and the semantic annotations of our Web Services are added automatically using SAWSDL attributes. BioSemantic is downloadable at http://southgreen.cirad.fr/?q=content/Biosemantic. PMID:23586394
47 CFR 52.32 - Allocation of the shared costs of long-term number portability.
Code of Federal Regulations, 2012 CFR
2012-10-01
....21(h), of each regional database, as defined in § 52.21(1), shall recover the shared costs of long-term number portability attributable to that regional database from all telecommunications carriers providing telecommunications service in areas that regional database serves. Pursuant to its duties under...
47 CFR 52.32 - Allocation of the shared costs of long-term number portability.
Code of Federal Regulations, 2010 CFR
2010-10-01
....21(h), of each regional database, as defined in § 52.21(1), shall recover the shared costs of long-term number portability attributable to that regional database from all telecommunications carriers providing telecommunications service in areas that regional database serves. Pursuant to its duties under...
47 CFR 52.32 - Allocation of the shared costs of long-term number portability.
Code of Federal Regulations, 2011 CFR
2011-10-01
....21(h), of each regional database, as defined in § 52.21(1), shall recover the shared costs of long-term number portability attributable to that regional database from all telecommunications carriers providing telecommunications service in areas that regional database serves. Pursuant to its duties under...
47 CFR 52.32 - Allocation of the shared costs of long-term number portability.
Code of Federal Regulations, 2014 CFR
2014-10-01
....21(h), of each regional database, as defined in § 52.21(1), shall recover the shared costs of long-term number portability attributable to that regional database from all telecommunications carriers providing telecommunications service in areas that regional database serves. Pursuant to its duties under...
47 CFR 52.32 - Allocation of the shared costs of long-term number portability.
Code of Federal Regulations, 2013 CFR
2013-10-01
....21(h), of each regional database, as defined in § 52.21(1), shall recover the shared costs of long-term number portability attributable to that regional database from all telecommunications carriers providing telecommunications service in areas that regional database serves. Pursuant to its duties under...
Probing concept of critical thinking in nursing education in Iran: a concept analysis.
Tajvidi, Mansooreh; Ghiyasvandian, Shahrzad; Salsali, Mahvash
2014-06-01
Given the wide disagreement over the definition of critical thinking in different disciplines, defining and standardizing the concept according to the discipline of nursing is essential. Moreover, there is limited scientific evidence regarding critical thinking in the context of nursing in Iran. The aim of this study was to analyze and clarify the concept of critical thinking in nursing education in Iran. We employed the hybrid model to define the concept of critical thinking. The hybrid model has three interconnected phases--the theoretical phase, the fieldwork phase, and the final analytic phase. In the theoretical phase, we searched the online scientific databases (such as Elsevier, Wiley, CINAHL, Proquest, Ovid, and Springer as well as Iranian databases such as SID, Magiran, and Iranmedex). In the fieldwork phase, a purposive sample of 17 nursing faculties, PhD students, clinical instructors, and clinical nurses was recruited. Participants were interviewed by using an interview guide. In the analytical phase we compared the data from the theoretical and the fieldwork phases. The concept of critical thinking had many different antecedents, attributes, and consequences. Antecedents, attributes, and consequences of critical thinking concept identified in the theoretical phase were in some ways different and in some way similar to antecedents, attributes, and consequences identified in the fieldwork phase. Finally critical thinking in nursing education in Iran was clarified. Critical thinking is a logical, situational, purposive, and outcome-oriented thinking process. It is an acquired and evolving ability which develops individually. Such thinking process could lead to the professional accountability, personal development, God's consent, conscience appeasement, and personality development. Copyright © 2014. Published by Elsevier B.V.
The landslide database for Germany: Closing the gap at national level
NASA Astrophysics Data System (ADS)
Damm, Bodo; Klose, Martin
2015-11-01
The Federal Republic of Germany has long been among the few European countries that lack a national landslide database. Systematic collection and inventory of landslide data still has a long research history in Germany, but one focussed on the development of databases with local or regional coverage. This has changed in recent years with the launch of a database initiative aimed at closing the data gap existing at national level. The present paper reports on this project that is based on a landslide database which evolved over the last 15 years to a database covering large parts of Germany. A strategy of systematic retrieval, extraction, and fusion of landslide data is at the heart of the methodology, providing the basis for a database with a broad potential of application. The database offers a data pool of more than 4,200 landslide data sets with over 13,000 single data files and dates back to the 12th century. All types of landslides are covered by the database, which stores not only core attributes, but also various complementary data, including data on landslide causes, impacts, and mitigation. The current database migration to PostgreSQL/PostGIS is focused on unlocking the full scientific potential of the database, while enabling data sharing and knowledge transfer via a web GIS platform. In this paper, the goals and the research strategy of the database project are highlighted at first, with a summary of best practices in database development providing perspective. Next, the focus is on key aspects of the methodology, which is followed by the results of three case studies in the German Central Uplands. The case study results exemplify database application in the analysis of landslide frequency and causes, impact statistics, and landslide susceptibility modeling. Using the example of these case studies, strengths and weaknesses of the database are discussed in detail. The paper concludes with a summary of the database project with regard to previous achievements and the strategic roadmap.
Attributable inpatient costs of recurrent Clostridium difficile infections.
Dubberke, Erik R; Schaefer, Eric; Reske, Kimberly A; Zilberberg, Marya; Hollenbeak, Christopher S; Olsen, Margaret A
2014-11-01
To determine the attributable inpatient costs of recurrent Clostridium difficile infections (CDIs). Retrospective cohort study. Academic, urban, tertiary care hospital. A total of 3,958 patients aged 18 years or more who developed an initial CDI episode from 2003 through 2009. Data were collected electronically from hospital administrative databases and were supplemented with chart review. Patients with an index CDI episode during the study period were followed up for 180 days from the end of their index hospitalization or the end of their index CDI antibiotic treatment (whichever occurred later). Total hospital costs during the outcome period for patients with recurrent versus a single episode of CDI were analyzed using zero-inflated lognormal models. There were 421 persons with recurrent CDI (recurrence rate, 10.6%). Recurrent CDI case patients were significantly more likely than persons without recurrence to have any hospital costs during the outcome period (P < .001). The estimated attributable cost of recurrent CDI was $11,631 (95% confidence interval, $8,937-$14,588). The attributable costs of recurrent CDI are considerable. Patients with recurrent CDI are significantly more likely to have inpatient hospital costs than patients who do not develop recurrences. Better strategies to predict and prevent CDI recurrences are needed.
Supervised Learning for Detection of Duplicates in Genomic Sequence Databases.
Chen, Qingyu; Zobel, Justin; Zhang, Xiuzhen; Verspoor, Karin
2016-01-01
First identified as an issue in 1996, duplication in biological databases introduces redundancy and even leads to inconsistency when contradictory information appears. The amount of data makes purely manual de-duplication impractical, and existing automatic systems cannot detect duplicates as precisely as can experts. Supervised learning has the potential to address such problems by building automatic systems that learn from expert curation to detect duplicates precisely and efficiently. While machine learning is a mature approach in other duplicate detection contexts, it has seen only preliminary application in genomic sequence databases. We developed and evaluated a supervised duplicate detection method based on an expert curated dataset of duplicates, containing over one million pairs across five organisms derived from genomic sequence databases. We selected 22 features to represent distinct attributes of the database records, and developed a binary model and a multi-class model. Both models achieve promising performance; under cross-validation, the binary model had over 90% accuracy in each of the five organisms, while the multi-class model maintains high accuracy and is more robust in generalisation. We performed an ablation study to quantify the impact of different sequence record features, finding that features derived from meta-data, sequence identity, and alignment quality impact performance most strongly. The study demonstrates machine learning can be an effective additional tool for de-duplication of genomic sequence databases. All Data are available as described in the supplementary material.
StarView: The object oriented design of the ST DADS user interface
NASA Technical Reports Server (NTRS)
Williams, J. D.; Pollizzi, J. A.
1992-01-01
StarView is the user interface being developed for the Hubble Space Telescope Data Archive and Distribution Service (ST DADS). ST DADS is the data archive for HST observations and a relational database catalog describing the archived data. Users will use StarView to query the catalog and select appropriate datasets for study. StarView sends requests for archived datasets to ST DADS which processes the requests and returns the database to the user. StarView is designed to be a powerful and extensible user interface. Unique features include an internal relational database to navigate query results, a form definition language that will work with both CRT and X interfaces, a data definition language that will allow StarView to work with any relational database, and the ability to generate adhoc queries without requiring the user to understand the structure of the ST DADS catalog. Ultimately, StarView will allow the user to refine queries in the local database for improved performance and merge in data from external sources for correlation with other query results. The user will be able to create a query from single or multiple forms, merging the selected attributes into a single query. Arbitrary selection of attributes for querying is supported. The user will be able to select how query results are viewed. A standard form or table-row format may be used. Navigation capabilities are provided to aid the user in viewing query results. Object oriented analysis and design techniques were used in the design of StarView to support the mechanisms and concepts required to implement these features. One such mechanism is the Model-View-Controller (MVC) paradigm. The MVC allows the user to have multiple views of the underlying database, while providing a consistent mechanism for interaction regardless of the view. This approach supports both CRT and X interfaces while providing a common mode of user interaction. Another powerful abstraction is the concept of a Query Model. This concept allows a single query to be built form a single or multiple forms before it is submitted to ST DADS. Supporting this concept is the adhoc query generator which allows the user to select and qualify an indeterminate number attributes from the database. The user does not need any knowledge of how the joins across various tables are to be resolved. The adhoc generator calculates the joins automatically and generates the correct SQL query.
International Technical Working Group Round Robin Tests
DOE Office of Scientific and Technical Information (OSTI.GOV)
Dudder, Gordon B.; Hanlen, Richard C.; Herbillion, Georges M.
The goal of nuclear forensics is to develop a preferred approach to support illicit trafficking investigations. This approach must be widely understood and accepted as credible. The principal objectives of the Round Robin Tests are to prioritize forensic techniques and methods, evaluate attribution capabilities, and examine the utility of database. The HEU (Highly Enriched Uranium) Round Robin, and previous Plutonium Round Robin, have made tremendous contributions to fulfilling these goals through a collaborative learning experience that resulted from the outstanding efforts of the nine participating internal laboratories. A prioritized list of techniques and methods has been developed based on thismore » exercise. Current work is focused on the extent to which the techniques and methods can be generalized. The HEU Round Robin demonstrated a rather high level of capability to determine the important characteristics of the materials and processes using analytical methods. When this capability is combined with the appropriate knowledge/database, it results in a significant capability to attribute the source of the materials to a specific process or facility. A number of shortfalls were also identified in the current capabilities including procedures for non-nuclear forensics and the lack of a comprehensive network of data/knowledge bases. The results of the Round Robin will be used to develop guidelines or a ''recommended protocol'' to be made available to the interested authorities and countries to use in real cases.« less
Rule-Based Statistical Calculations on a Database Abstract.
1983-06-01
quadruples 17 L.6.6. Our methds ~ in distributed systems 17 L.6.7. Easy extensions 17 17. The datibms abstract as a database 17 17.1.w S orae mu is 1.7.2...the largest item in the intersection of two sets cannot be any larger that the minima of the maxima of the two sets for some numeric attribute. On the...from "range analysis" of arbitrary numeric attributes. Suppose the length range of tankers is from 300 to 1000 feet and that of American ships 50 to
Detection of driver engagement in secondary tasks from observed naturalistic driving behavior.
Ye, Mengqiu; Osman, Osama A; Ishak, Sherif; Hashemi, Bita
2017-09-01
Distracted driving has long been acknowledged as one of the leading causes of death or injury in roadway crashes. The focus of past research has been mainly on the impact of different causes of distraction on driving behavior. However, only a few studies attempted to address how some driving behavior attributes could be linked to the cause of distraction. In essence, this study takes advantage of the rich SHRP 2 Naturalistic Driving Study (NDS) database to develop a model for detecting the likelihood of a driver's involvement in secondary tasks from distinctive attributes of driving behavior. Five performance attributes, namely speed, longitudinal acceleration, lateral acceleration, yaw rate, and throttle position were used to describe the driving behavior. A model was developed for each of three selected secondary tasks: calling, texting, and passenger interaction. The models were developed using a supervised feed-forward Artificial Neural Network (ANN) architecture to account for the effect of inherent nonlinearity in the relationships between driving behavior and secondary tasks. The results show that the developed ANN models were able to detect the drivers' involvement in calling, texting, and passenger interaction with an overall accuracy of 99.5%, 98.1%, and 99.8%, respectively. These results show that the selected driving performance attributes were effective in detecting the associated secondary tasks with driving behavior. The results are very promising and the developed models could potentially be applied in crash investigations to resolve legal disputes in traffic accidents. Copyright © 2017 Elsevier Ltd. All rights reserved.
BIOPEP database and other programs for processing bioactive peptide sequences.
Minkiewicz, Piotr; Dziuba, Jerzy; Iwaniak, Anna; Dziuba, Marta; Darewicz, Małgorzata
2008-01-01
This review presents the potential for application of computational tools in peptide science based on a sample BIOPEP database and program as well as other programs and databases available via the World Wide Web. The BIOPEP application contains a database of biologically active peptide sequences and a program enabling construction of profiles of the potential biological activity of protein fragments, calculation of quantitative descriptors as measures of the value of proteins as potential precursors of bioactive peptides, and prediction of bonds susceptible to hydrolysis by endopeptidases in a protein chain. Other bioactive and allergenic peptide sequence databases are also presented. Programs enabling the construction of binary and multiple alignments between peptide sequences, the construction of sequence motifs attributed to a given type of bioactivity, searching for potential precursors of bioactive peptides, and the prediction of sites susceptible to proteolytic cleavage in protein chains are available via the Internet as are other approaches concerning secondary structure prediction and calculation of physicochemical features based on amino acid sequence. Programs for prediction of allergenic and toxic properties have also been developed. This review explores the possibilities of cooperation between various programs.
An Algorithm of Association Rule Mining for Microbial Energy Prospection
Shaheen, Muhammad; Shahbaz, Muhammad
2017-01-01
The presence of hydrocarbons beneath earth’s surface produces some microbiological anomalies in soils and sediments. The detection of such microbial populations involves pure bio chemical processes which are specialized, expensive and time consuming. This paper proposes a new algorithm of context based association rule mining on non spatial data. The algorithm is a modified form of already developed algorithm which was for spatial database only. The algorithm is applied to mine context based association rules on microbial database to extract interesting and useful associations of microbial attributes with existence of hydrocarbon reserve. The surface and soil manifestations caused by the presence of hydrocarbon oxidizing microbes are selected from existing literature and stored in a shared database. The algorithm is applied on the said database to generate direct and indirect associations among the stored microbial indicators. These associations are then correlated with the probability of hydrocarbon’s existence. The numerical evaluation shows better accuracy for non-spatial data as compared to conventional algorithms at generating reliable and robust rules. PMID:28393846
The MAR databases: development and implementation of databases specific for marine metagenomics.
Klemetsen, Terje; Raknes, Inge A; Fu, Juan; Agafonov, Alexander; Balasundaram, Sudhagar V; Tartari, Giacomo; Robertsen, Espen; Willassen, Nils P
2018-01-04
We introduce the marine databases; MarRef, MarDB and MarCat (https://mmp.sfb.uit.no/databases/), which are publicly available resources that promote marine research and innovation. These data resources, which have been implemented in the Marine Metagenomics Portal (MMP) (https://mmp.sfb.uit.no/), are collections of richly annotated and manually curated contextual (metadata) and sequence databases representing three tiers of accuracy. While MarRef is a database for completely sequenced marine prokaryotic genomes, which represent a marine prokaryote reference genome database, MarDB includes all incomplete sequenced prokaryotic genomes regardless level of completeness. The last database, MarCat, represents a gene (protein) catalog of uncultivable (and cultivable) marine genes and proteins derived from marine metagenomics samples. The first versions of MarRef and MarDB contain 612 and 3726 records, respectively. Each record is built up of 106 metadata fields including attributes for sampling, sequencing, assembly and annotation in addition to the organism and taxonomic information. Currently, MarCat contains 1227 records with 55 metadata fields. Ontologies and controlled vocabularies are used in the contextual databases to enhance consistency. The user-friendly web interface lets the visitors browse, filter and search in the contextual databases and perform BLAST searches against the corresponding sequence databases. All contextual and sequence databases are freely accessible and downloadable from https://s1.sfb.uit.no/public/mar/. © The Author(s) 2017. Published by Oxford University Press on behalf of Nucleic Acids Research.
Defending against Attribute-Correlation Attacks in Privacy-Aware Information Brokering
NASA Astrophysics Data System (ADS)
Li, Fengjun; Luo, Bo; Liu, Peng; Squicciarini, Anna C.; Lee, Dongwon; Chu, Chao-Hsien
Nowadays, increasing needs for information sharing arise due to extensive collaborations among organizations. Organizations desire to provide data access to their collaborators while preserving full control over the data and comprehensive privacy of their users. A number of information systems have been developed to provide efficient and secure information sharing. However, most of the solutions proposed so far are built atop of conventional data warehousing or distributed database technologies.
NASA Astrophysics Data System (ADS)
Rogers, Steven P.; Hamilton, David B.
1994-06-01
To employ the most readily comprehensible presentation methods and symbology with helmet-mounted displays (HMDs), it is critical to identify the information elements needed to perform each pilot function and to analytically determine the attributes of these elements. The extensive analyses of mission requirements currently performed for pilot-vehicle interface design can be aided and improved by the new capabilities of intelligent systems and relational databases. An intelligent system, named ACIDTEST, has been developed specifically for organizing and applying rules to identify the best display modalities, locations, and formats. The primary objectives of the ACIDTEST system are to provide rapid accessibility to pertinent display research data, to integrate guidelines from many disciplines and identify conflicts among these guidelines, to force a consistent display approach among the design team members, and to serve as an 'audit trail' of design decisions and justifications. A powerful relational database called TAWL ORDIR has been developed to document information requirements and attributes for use by ACIDTEST as well as to greatly augment the applicability of mission analysis data. TAWL ORDIR can be used to rapidly reorganize mission analysis data components for study, perform commonality analyses for groups of tasks, determine the information content requirement for tailored display modes, and identify symbology integration opportunities.
Clinical Documents: Attribute-Values Entity Representation, Context, Page Layout And Communication
Lovis, Christian; Lamb, Alexander; Baud, Robert; Rassinoux, Anne-Marie; Fabry, Paul; Geissbühler, Antoine
2003-01-01
This paper presents how acquisition, storage and communication of clinical documents are implemented at the University Hospitals of Geneva. Careful attention has been given to user-interfaces, in order to support complex layouts, spell checking, templates management with automatic prefilling in order to facilitate acquisition. A dual architecture has been developed for storage using an attributes-values entity unified database and a consolidated, patient-centered, layout-respectful files-based storage, providing both representation power and sinsert (peed of accesses. This architecture allows great flexibility to store a continuum of data types from simple type values up to complex clinical reports. Finally, communication is entirely based on HTTP-XML internally and a HL-7 CDA interface V2 is currently studied for external communication. Some of the problem encountered, mostly concerning the typology of documents and the ontology of clinical attributes are evoked. PMID:14728202
Yamauchi, Takashi; Sasaki, Takeshi; Yoshikawa, Toru; Matsumoto, Shun; Takahashi, Masaya; Suka, Machi; Yanagisawa, Hiroyuki
2018-04-01
This study aimed to clarify whether work-related adverse events in cases involving compensation for mental disorders and suicide differ by sex and industry using a database containing all relevant cases reported from 2010 to 2014 in Japan. A total of 1362 eligible cases involving compensation for mental disorders (422 females and 940 males) were analyzed. Among males, 55.7% of cases were attributed to "long working hours." In both sexes, the frequencies of cases attributed to "long working hours" and other events differed significantly by industry. Among cases involving compensation for suicide, 71.4% were attributed to "long working hours." The frequency distribution of work-related adverse events differed significantly by sex and industry. These differences should be taken into consideration in the development of industry-specific preventive measures for occupational mental disorders.
Desiderata for a Computer-Assisted Audit Tool for Clinical Data Source Verification Audits
Duda, Stephany N.; Wehbe, Firas H.; Gadd, Cynthia S.
2013-01-01
Clinical data auditing often requires validating the contents of clinical research databases against source documents available in health care settings. Currently available data audit software, however, does not provide features necessary to compare the contents of such databases to source data in paper medical records. This work enumerates the primary weaknesses of using paper forms for clinical data audits and identifies the shortcomings of existing data audit software, as informed by the experiences of an audit team evaluating data quality for an international research consortium. The authors propose a set of attributes to guide the development of a computer-assisted clinical data audit tool to simplify and standardize the audit process. PMID:20841814
LinkedOmics: analyzing multi-omics data within and across 32 cancer types.
Vasaikar, Suhas V; Straub, Peter; Wang, Jing; Zhang, Bing
2018-01-04
The LinkedOmics database contains multi-omics data and clinical data for 32 cancer types and a total of 11 158 patients from The Cancer Genome Atlas (TCGA) project. It is also the first multi-omics database that integrates mass spectrometry (MS)-based global proteomics data generated by the Clinical Proteomic Tumor Analysis Consortium (CPTAC) on selected TCGA tumor samples. In total, LinkedOmics has more than a billion data points. To allow comprehensive analysis of these data, we developed three analysis modules in the LinkedOmics web application. The LinkFinder module allows flexible exploration of associations between a molecular or clinical attribute of interest and all other attributes, providing the opportunity to analyze and visualize associations between billions of attribute pairs for each cancer cohort. The LinkCompare module enables easy comparison of the associations identified by LinkFinder, which is particularly useful in multi-omics and pan-cancer analyses. The LinkInterpreter module transforms identified associations into biological understanding through pathway and network analysis. Using five case studies, we demonstrate that LinkedOmics provides a unique platform for biologists and clinicians to access, analyze and compare cancer multi-omics data within and across tumor types. LinkedOmics is freely available at http://www.linkedomics.org. © The Author(s) 2017. Published by Oxford University Press on behalf of Nucleic Acids Research.
Pattern-based, multi-scale segmentation and regionalization of EOSD land cover
NASA Astrophysics Data System (ADS)
Niesterowicz, Jacek; Stepinski, Tomasz F.
2017-10-01
The Earth Observation for Sustainable Development of Forests (EOSD) map is a 25 m resolution thematic map of Canadian forests. Because of its large spatial extent and relatively high resolution the EOSD is difficult to analyze using standard GIS methods. In this paper we propose multi-scale segmentation and regionalization of EOSD as new methods for analyzing EOSD on large spatial scales. Segments, which we refer to as forest land units (FLUs), are delineated as tracts of forest characterized by cohesive patterns of EOSD categories; we delineated from 727 to 91,885 FLUs within the spatial extent of EOSD depending on the selected scale of a pattern. Pattern of EOSD's categories within each FLU is described by 1037 landscape metrics. A shapefile containing boundaries of all FLUs together with an attribute table listing landscape metrics make up an SQL-searchable spatial database providing detailed information on composition and pattern of land cover types in Canadian forest. Shapefile format and extensive attribute table pertaining to the entire legend of EOSD are designed to facilitate broad range of investigations in which assessment of composition and pattern of forest over large areas is needed. We calculated four such databases using different spatial scales of pattern. We illustrate the use of FLU database for producing forest regionalization maps of two Canadian provinces, Quebec and Ontario. Such maps capture the broad scale variability of forest at the spatial scale of the entire province. We also demonstrate how FLU database can be used to map variability of landscape metrics, and thus the character of landscape, over the entire Canada.
Konias, Sokratis; Chouvarda, Ioanna; Vlahavas, Ioannis; Maglaveras, Nicos
2005-09-01
Current approaches for mining association rules usually assume that the mining is performed in a static database, where the problem of missing attribute values does not practically exist. However, these assumptions are not preserved in some medical databases, like in a home care system. In this paper, a novel uncertainty rule algorithm is illustrated, namely URG-2 (Uncertainty Rule Generator), which addresses the problem of mining dynamic databases containing missing values. This algorithm requires only one pass from the initial dataset in order to generate the item set, while new metrics corresponding to the notion of Support and Confidence are used. URG-2 was evaluated over two medical databases, introducing randomly multiple missing values for each record's attribute (rate: 5-20% by 5% increments) in the initial dataset. Compared with the classical approach (records with missing values are ignored), the proposed algorithm was more robust in mining rules from datasets containing missing values. In all cases, the difference in preserving the initial rules ranged between 30% and 60% in favour of URG-2. Moreover, due to its incremental nature, URG-2 saved over 90% of the time required for thorough re-mining. Thus, the proposed algorithm can offer a preferable solution for mining in dynamic relational databases.
Average probability that a "cold hit" in a DNA database search results in an erroneous attribution.
Song, Yun S; Patil, Anand; Murphy, Erin E; Slatkin, Montgomery
2009-01-01
We consider a hypothetical series of cases in which the DNA profile of a crime-scene sample is found to match a known profile in a DNA database (i.e., a "cold hit"), resulting in the identification of a suspect based only on genetic evidence. We show that the average probability that there is another person in the population whose profile matches the crime-scene sample but who is not in the database is approximately 2(N - d)p(A), where N is the number of individuals in the population, d is the number of profiles in the database, and p(A) is the average match probability (AMP) for the population. The AMP is estimated by computing the average of the probabilities that two individuals in the population have the same profile. We show further that if a priori each individual in the population is equally likely to have left the crime-scene sample, then the average probability that the database search attributes the crime-scene sample to a wrong person is (N - d)p(A).
NASA Technical Reports Server (NTRS)
Milos, Frank S.
2011-01-01
In most previous work at NASA Ames Research Center, ablation predictions for carbonaceous materials were obtained using a species thermodynamics database developed by Aerotherm Corporation. This database is derived mostly from the JANAF thermochemical tables. However, the CEA thermodynamics database, also used by NASA, is considered more up to date. In this work, the FIAT code was modified to use CEA-based curve fits for species thermodynamics, then analyses using both the JANAF and CEA thermodynamics were performed for carbon and carbon phenolic materials over a range of test conditions. The ablation predictions are comparable at lower heat fluxes where the dominant mechanism is carbon oxidation. However, the predictions begin to diverge in the sublimation regime, with the CEA model predicting lower recession. The disagreement is more significant for carbon phenolic than for carbon, and this difference is attributed to hydrocarbon species that may contribute to the ablation rate.
GEOGRAPHIC NAMES INFORMATION SYSTEM (GNIS) ...
The Geographic Names Information System (GNIS), developed by the U.S. Geological Survey in cooperation with the U.S. Board on Geographic Names (BGN), contains information about physical and cultural geographic features in the United States and associated areas, both current and historical, but not including roads and highways. The database also contains geographic names in Antarctica. The database holds the Federally recognized name of each feature and defines the location of the feature by state, county, USGS topographic map, and geographic coordinates. Other feature attributes include names or spellings other than the official name, feature designations, feature class, historical and descriptive information, and for some categories of features the geometric boundaries. The database assigns a unique feature identifier, a random number, that is a key for accessing, integrating, or reconciling GNIS data with other data sets. The GNIS is our Nation's official repository of domestic geographic feature names information.
Jaiswal, Kishor
2013-01-01
This memo lays out a procedure for the GEM software to offer an available vulnerability function for any acceptable set of attributes that the user specifies for a particular building category. The memo also provides general guidelines on how to submit the vulnerability or fragility functions to the GEM vulnerability repository, stipulating which attributes modelers must provide so that their vulnerability or fragility functions can be queried appropriately by the vulnerability database. An important objective is to provide users guidance on limitations and applicability by providing the associated modeling assumptions and applicability of each vulnerability or fragility function.
A global organism detection and monitoring system for non-native species
Graham, J.; Newman, G.; Jarnevich, C.; Shory, R.; Stohlgren, T.J.
2007-01-01
Harmful invasive non-native species are a significant threat to native species and ecosystems, and the costs associated with non-native species in the United States is estimated at over $120 Billion/year. While some local or regional databases exist for some taxonomic groups, there are no effective geographic databases designed to detect and monitor all species of non-native plants, animals, and pathogens. We developed a web-based solution called the Global Organism Detection and Monitoring (GODM) system to provide real-time data from a broad spectrum of users on the distribution and abundance of non-native species, including attributes of their habitats for predictive spatial modeling of current and potential distributions. The four major subsystems of GODM provide dynamic links between the organism data, web pages, spatial data, and modeling capabilities. The core survey database tables for recording invasive species survey data are organized into three categories: "Where, Who & When, and What." Organisms are identified with Taxonomic Serial Numbers from the Integrated Taxonomic Information System. To allow users to immediately see a map of their data combined with other user's data, a custom geographic information system (GIS) Internet solution was required. The GIS solution provides an unprecedented level of flexibility in database access, allowing users to display maps of invasive species distributions or abundances based on various criteria including taxonomic classification (i.e., phylum or division, order, class, family, genus, species, subspecies, and variety), a specific project, a range of dates, and a range of attributes (percent cover, age, height, sex, weight). This is a significant paradigm shift from "map servers" to true Internet-based GIS solutions. The remainder of the system was created with a mix of commercial products, open source software, and custom software. Custom GIS libraries were created where required for processing large datasets, accessing the operating system, and to use existing libraries in C++, R, and other languages to develop the tools to track harmful species in space and time. The GODM database and system are crucial for early detection and rapid containment of invasive species. ?? 2007 Elsevier B.V. All rights reserved.
[Development of Hospital Equipment Maintenance Information System].
Zhou, Zhixin
2015-11-01
Hospital equipment maintenance information system plays an important role in improving medical treatment quality and efficiency. By requirement analysis of hospital equipment maintenance, the system function diagram is drawed. According to analysis of input and output data, tables and reports in connection with equipment maintenance process, relationships between entity and attribute is found out, and E-R diagram is drawed and relational database table is established. Software development should meet actual process requirement of maintenance and have a friendly user interface and flexible operation. The software can analyze failure cause by statistical analysis.
The Protein Information Resource: an integrated public resource of functional annotation of proteins
Wu, Cathy H.; Huang, Hongzhan; Arminski, Leslie; Castro-Alvear, Jorge; Chen, Yongxing; Hu, Zhang-Zhi; Ledley, Robert S.; Lewis, Kali C.; Mewes, Hans-Werner; Orcutt, Bruce C.; Suzek, Baris E.; Tsugita, Akira; Vinayaka, C. R.; Yeh, Lai-Su L.; Zhang, Jian; Barker, Winona C.
2002-01-01
The Protein Information Resource (PIR) serves as an integrated public resource of functional annotation of protein data to support genomic/proteomic research and scientific discovery. The PIR, in collaboration with the Munich Information Center for Protein Sequences (MIPS) and the Japan International Protein Information Database (JIPID), produces the PIR-International Protein Sequence Database (PSD), the major annotated protein sequence database in the public domain, containing about 250 000 proteins. To improve protein annotation and the coverage of experimentally validated data, a bibliography submission system is developed for scientists to submit, categorize and retrieve literature information. Comprehensive protein information is available from iProClass, which includes family classification at the superfamily, domain and motif levels, structural and functional features of proteins, as well as cross-references to over 40 biological databases. To provide timely and comprehensive protein data with source attribution, we have introduced a non-redundant reference protein database, PIR-NREF. The database consists of about 800 000 proteins collected from PIR-PSD, SWISS-PROT, TrEMBL, GenPept, RefSeq and PDB, with composite protein names and literature data. To promote database interoperability, we provide XML data distribution and open database schema, and adopt common ontologies. The PIR web site (http://pir.georgetown.edu/) features data mining and sequence analysis tools for information retrieval and functional identification of proteins based on both sequence and annotation information. The PIR databases and other files are also available by FTP (ftp://nbrfa.georgetown.edu/pir_databases). PMID:11752247
Leventhal, John M; Martin, Kimberly D; Asnes, Andrea G
2008-09-01
The goal was to assess the proportion of children with fractures attributable to abuse and the incidence of fractures caused by abuse among children <36 months of age who were hospitalized in the United States. We used the Kids' Inpatient Database, which has discharge data on 80% of acute pediatric hospitalizations in the United States, for 3 time periods (1997, 2000, and 2003). Fractures attributable to abuse in children <36 months of age were identified by both an International Classification of Diseases, Ninth Revision, Clinical Modification code for fracture and a diagnosis external-cause-of-injury code for abuse. Weighted estimates of the incidence were calculated. Among children <36 months of age who were hospitalized with fractures, the proportions of cases attributable to abuse were 11.9% in 1997, 11.9% in 2000, and 12.1% in 2003. The proportions of cases attributable to abuse decreased with increasing age; for example, in 2003, the proportions attributable to abuse were 24.9% for children <12 months of age, 7.2% for children 12 to 23 months of age, and 2.9% for children 24 to 35 months of age. In 2003, the incidence of fractures caused by abuse was 15.3 cases per 100000 children <36 months of age. The incidence was 36.1 cases per 100000 among children <12 months of age; this decreased to 4.8 cases per 100000 among 12- to 23-month-old children and 4.8 cases per 100000 among 24- to 35-month-old children. The Kids' Inpatient Database can be used to provide reasonable estimates of the incidence of hospitalization with fractures attributable to child abuse. For children <12 months of age, the incidence was 36.1 cases per 100000, a rate similar to that of inflicted traumatic brain injury (25-32 cases per 100000).
Emission & Generation Resource Integrated Database (eGRID)
The Emissions & Generation Resource Integrated Database (eGRID) is an integrated source of data on environmental characteristics of electric power generation. Twelve federal databases are represented by eGRID, which provides air emission and resource mix information for thousands of power plants and generating companies. eGRID allows direct comparison of the environmental attributes of electricity from different plants, companies, States, or regions of the power grid.
Assessment of indexing trends with specific and general terms for herbal medicine.
Bartol, Tomaz
2012-12-01
Concepts for medicinal plants are represented by a variety of associated general terms with specific indexing patterns in databases, which may not consistently reflect growth of records. The objectives of this study are to assess the development in databases by identifying general terms that describe herbal medicine with optimal retrieval recall and to identify possible special trends in co-occurrence of specific and general concepts. Different search strategies are tested in cab abstracts, medline and web of science. Specific terms (Origanum and Salvia) are employed. Relevant general terms (e.g. 'Plants, Medicinal', Phytotherapy, Herbal drugs) are identified, along with indexing trends and co-occurrences. Growth trends, in specific (narrower) terms, are similar among databases. General terms, however, exhibit dissimilar trends, sometimes almost opposing one another. Co-occurrence of specific and general terms is changing over time. General terms may not denote definite development of trends as the use of terms differs amongst databases, making it difficult to correctly assess possible numbers of relevant records. Perceived increase can, sometimes, be attributed to an increased occurrence of a more general term alongside the specific one. Thesaurus-controlled databases may yield more hits, because of 'up-posted' (broader) terms. Use of broader terms is helpful as it enhances retrieval of relevant documents. © 2012 The authors. Health Information and Libraries Journal © 2012 Health Libraries Group.
Impact of diabetes on healthcare costs in a population-based cohort: a cost analysis.
Rosella, L C; Lebenbaum, M; Fitzpatrick, T; O'Reilly, D; Wang, J; Booth, G L; Stukel, T A; Wodchis, W P
2016-03-01
To estimate the healthcare costs attributable to diabetes in Ontario, Canada using a propensity-matched control design and health administrative data from the perspective of a single-payer healthcare system. Incident diabetes cases among adults in Ontario were identified from the Ontario Diabetes Database between 2004 and 2012 and matched 1:3 to control subjects without diabetes identified in health administrative databases on the basis of sociodemographics and propensity score. Using a comprehensive source of administrative databases, direct per-person costs (Canadian dollars 2012) were calculated. A cost analysis was performed to calculate the attributable costs of diabetes; i.e. the difference of costs between patients with diabetes and control subjects without diabetes. The study sample included 699 042 incident diabetes cases. The costs attributable to diabetes were greatest in the year after diagnosis [C$3,785 (95% CI 3708, 3862) per person for women and C$3,826 (95% CI 3751, 3901) for men], increasing substantially for older age groups and patients who died during follow-up. After accounting for baseline comorbidities, attributable costs were primarily incurred through inpatient acute hospitalizations, physician visits and prescription medications and assistive devices. The excess healthcare costs attributable to diabetes are substantial and pose a significant clinical and public health challenge. This burden is an important consideration for decision-makers, particularly given increasing concern over the sustainability of the healthcare system, aging population structure and increasing prevalence of diabetic risk factors, such as obesity. © 2015 The Authors. Diabetic Medicine published by John Wiley & Sons Ltd on behalf of Diabetes UK.
Performance Evaluation of a Database System in a Multiple Backend Configurations,
1984-10-01
leaving a systemn process , the * internal performance measuremnents of MMSD have been carried out. Mathodo lo.- gies for constructing test databases...access d i rectory data via the AT, EDIT, and CDT. In designing the test database, one of the key concepts is the choice of the directory attributes in...internal timing. These requests are selected since they retrieve the seIaI lest portion of the test database and the processing time for each request is
Integration of Information Retrieval and Database Management Systems.
ERIC Educational Resources Information Center
Deogun, Jitender S.; Raghavan, Vijay V.
1988-01-01
Discusses the motivation for integrating information retrieval and database management systems, and proposes a probabilistic retrieval model in which records in a file may be composed of attributes (formatted data items) and descriptors (content indicators). The details and resolutions of difficulties involved in integrating such systems are…
Automated Database Schema Design Using Mined Data Dependencies.
ERIC Educational Resources Information Center
Wong, S. K. M.; Butz, C. J.; Xiang, Y.
1998-01-01
Describes a bottom-up procedure for discovering multivalued dependencies in observed data without knowing a priori the relationships among the attributes. The proposed algorithm is an application of technique designed for learning conditional independencies in probabilistic reasoning; a prototype system for automated database schema design has…
Geographic Information System Tools for Conservation Planning: User's Manual
Fox, Timothy J.; Rohweder, Jason J.; Kenow, K.P.; Korschgen, C.E.; DeHaan, H.C.
2003-01-01
Public and private land managers desire better ways to incorporate landscape, species, and habitat relations into their conservation planning processes. We present three tools, developed for the Environmental Systems Research Institute?s ArcView 3.x platform, applicable to many types of wildlife conservation management and planning efforts. These tools provide managers and planners with the ability to rapidly assess landscape attributes and link these attributes with species-habitat information. To use the tools, the user provides a detailed land cover spatial database and develops a matrix to identify species-habitat relations for the landscape of interest. The tools are applicable to any taxa or suite of taxa for which the required data are available. The user also has the ability to interactively make polygon-specific changes to the landscape and re-examine species-habitat relations. The development of these tools has given resource managers the means to evaluate the merits of proposed landscape management scenarios and to choose the scenario that best fits the goals of the managed area.
Flexible network reconstruction from relational databases with Cytoscape and CytoSQL
2010-01-01
Background Molecular interaction networks can be efficiently studied using network visualization software such as Cytoscape. The relevant nodes, edges and their attributes can be imported in Cytoscape in various file formats, or directly from external databases through specialized third party plugins. However, molecular data are often stored in relational databases with their own specific structure, for which dedicated plugins do not exist. Therefore, a more generic solution is presented. Results A new Cytoscape plugin 'CytoSQL' is developed to connect Cytoscape to any relational database. It allows to launch SQL ('Structured Query Language') queries from within Cytoscape, with the option to inject node or edge features of an existing network as SQL arguments, and to convert the retrieved data to Cytoscape network components. Supported by a set of case studies we demonstrate the flexibility and the power of the CytoSQL plugin in converting specific data subsets into meaningful network representations. Conclusions CytoSQL offers a unified approach to let Cytoscape interact with relational databases. Thanks to the power of the SQL syntax, this tool can rapidly generate and enrich networks according to very complex criteria. The plugin is available at http://www.ptools.ua.ac.be/CytoSQL. PMID:20594316
Flexible network reconstruction from relational databases with Cytoscape and CytoSQL.
Laukens, Kris; Hollunder, Jens; Dang, Thanh Hai; De Jaeger, Geert; Kuiper, Martin; Witters, Erwin; Verschoren, Alain; Van Leemput, Koenraad
2010-07-01
Molecular interaction networks can be efficiently studied using network visualization software such as Cytoscape. The relevant nodes, edges and their attributes can be imported in Cytoscape in various file formats, or directly from external databases through specialized third party plugins. However, molecular data are often stored in relational databases with their own specific structure, for which dedicated plugins do not exist. Therefore, a more generic solution is presented. A new Cytoscape plugin 'CytoSQL' is developed to connect Cytoscape to any relational database. It allows to launch SQL ('Structured Query Language') queries from within Cytoscape, with the option to inject node or edge features of an existing network as SQL arguments, and to convert the retrieved data to Cytoscape network components. Supported by a set of case studies we demonstrate the flexibility and the power of the CytoSQL plugin in converting specific data subsets into meaningful network representations. CytoSQL offers a unified approach to let Cytoscape interact with relational databases. Thanks to the power of the SQL syntax, this tool can rapidly generate and enrich networks according to very complex criteria. The plugin is available at http://www.ptools.ua.ac.be/CytoSQL.
An evolutionary concept analysis of futility in health care.
Morata, Lauren
2018-06-01
To report a concept analysis of futility in health care. Each member of the healthcare team: the physician, the nurse, the patient, the family and all others involved perceive futility differently. The current evidence and knowledge in regard to futility in health care manifest a plethora of definitions, meanings and interpretations without consensus. Concept analysis. Databases searched included Medline, Cumulative Index of Nursing and Allied Health Literature, Academic Search Premier, Cochrane Database of Systematic Reviews and PsycINFO. Search terms included "futil*," "concept analysis," "concept," "inefficacious," "non-beneficial," "ineffective" and "fruitless" from 1935-2016 to ensure a historical perspective of the concept. A total of 106 articles were retained to develop the concept. Rogers' evolutionary concept analysis was used to evaluate the concept of futility from ancient medicine to the present. Seven antecedents (the patient/family autonomy, surrogate decision-making movement, the patient-family/physician relationship, physician authority, legislation and court rulings, catastrophic events and advancing medical technology) lead to four major attributes (quantitative, physiologic, qualitative, and disease-specific). Ultimately, futile care could lead to consequences such as litigation, advancing technology, increasing healthcare costs, rationing, moral distress and ethical dilemmas. Futility in health care demonstrates components of a cyclical process and a consensus definition is proposed. A framework is developed to clarify the concept and articulate relationships among attributes, antecedents and consequences. Further testing of the proposed definition and framework are needed. © 2018 John Wiley & Sons Ltd.
Information extraction from Italian medical reports: An ontology-driven approach.
Viani, Natalia; Larizza, Cristiana; Tibollo, Valentina; Napolitano, Carlo; Priori, Silvia G; Bellazzi, Riccardo; Sacchi, Lucia
2018-03-01
In this work, we propose an ontology-driven approach to identify events and their attributes from episodes of care included in medical reports written in Italian. For this language, shared resources for clinical information extraction are not easily accessible. The corpus considered in this work includes 5432 non-annotated medical reports belonging to patients with rare arrhythmias. To guide the information extraction process, we built a domain-specific ontology that includes the events and the attributes to be extracted, with related regular expressions. The ontology and the annotation system were constructed on a development set, while the performance was evaluated on an independent test set. As a gold standard, we considered a manually curated hospital database named TRIAD, which stores most of the information written in reports. The proposed approach performs well on the considered Italian medical corpus, with a percentage of correct annotations above 90% for most considered clinical events. We also assessed the possibility to adapt the system to the analysis of another language (i.e., English), with promising results. Our annotation system relies on a domain ontology to extract and link information in clinical text. We developed an ontology that can be easily enriched and translated, and the system performs well on the considered task. In the future, it could be successfully used to automatically populate the TRIAD database. Copyright © 2017 Elsevier B.V. All rights reserved.
Fonseca, Carissa G; Backhaus, Michael; Bluemke, David A; Britten, Randall D; Chung, Jae Do; Cowan, Brett R; Dinov, Ivo D; Finn, J Paul; Hunter, Peter J; Kadish, Alan H; Lee, Daniel C; Lima, Joao A C; Medrano-Gracia, Pau; Shivkumar, Kalyanam; Suinesiaputra, Avan; Tao, Wenchao; Young, Alistair A
2011-08-15
Integrative mathematical and statistical models of cardiac anatomy and physiology can play a vital role in understanding cardiac disease phenotype and planning therapeutic strategies. However, the accuracy and predictive power of such models is dependent upon the breadth and depth of noninvasive imaging datasets. The Cardiac Atlas Project (CAP) has established a large-scale database of cardiac imaging examinations and associated clinical data in order to develop a shareable, web-accessible, structural and functional atlas of the normal and pathological heart for clinical, research and educational purposes. A goal of CAP is to facilitate collaborative statistical analysis of regional heart shape and wall motion and characterize cardiac function among and within population groups. Three main open-source software components were developed: (i) a database with web-interface; (ii) a modeling client for 3D + time visualization and parametric description of shape and motion; and (iii) open data formats for semantic characterization of models and annotations. The database was implemented using a three-tier architecture utilizing MySQL, JBoss and Dcm4chee, in compliance with the DICOM standard to provide compatibility with existing clinical networks and devices. Parts of Dcm4chee were extended to access image specific attributes as search parameters. To date, approximately 3000 de-identified cardiac imaging examinations are available in the database. All software components developed by the CAP are open source and are freely available under the Mozilla Public License Version 1.1 (http://www.mozilla.org/MPL/MPL-1.1.txt). http://www.cardiacatlas.org a.young@auckland.ac.nz Supplementary data are available at Bioinformatics online.
NASA Astrophysics Data System (ADS)
Medland, A. J.; Zhu, Guowang; Gao, Jian; Sun, Jian
1996-03-01
Feature conversion, also called feature transformation and feature mapping, is defined as the process of converting features from one view of an object to another view of the object. In a relatively simple implementation, for each application the design features are automatically converted into features specific for that application. All modifications have to be made via the design features. This is the approach that has attracted most attention until now. In the ideal situation, however, conversions directly from application views to the design view, and to other applications views, are also possible. In this paper, some difficulties faced in feature conversion are discussed. A new representation scheme of feature-based parts models has been proposed for the purpose of one-way feature conversion. The parts models consist of five different levels of abstraction, extending from an assembly level and its attributes, single parts and their attributes, single features and their attributes, one containing the geometric reference element and finally one for detailed geometry. One implementation of feature conversion for rotational components within GT (Group Technology) has already been undertaken using an automated coding procedure operating on a design-feature database. This database has been generated by a feature-based design system, and the GT coding scheme used in this paper is a specific scheme created for a textile machine manufacturing plant. Such feature conversion techniques presented here are only in their early stages of development and further research is underway.
Multidatabase Query Processing with Uncertainty in Global Keys and Attribute Values.
ERIC Educational Resources Information Center
Scheuermann, Peter; Li, Wen-Syan; Clifton, Chris
1998-01-01
Presents an approach for dynamic database integration and query processing in the absence of information about attribute correspondences and global IDs. Defines different types of equivalence conditions for the construction of global IDs. Proposes a strategy based on ranked role-sets that makes use of an automated semantic integration procedure…
Representing sentence information
NASA Astrophysics Data System (ADS)
Perkins, Walton A., III
1991-03-01
This paper describes a computer-oriented representation for sentence information. Whereas many Artificial Intelligence (AI) natural language systems start with a syntactic parse of a sentence into the linguist's components: noun, verb, adjective, preposition, etc., we argue that it is better to parse the input sentence into 'meaning' components: attribute, attribute value, object class, object instance, and relation. AI systems need a representation that will allow rapid storage and retrieval of information and convenient reasoning with that information. The attribute-of-object representation has proven useful for handling information in relational databases (which are well known for their efficiency in storage and retrieval) and for reasoning in knowledge- based systems. On the other hand, the linguist's syntactic representation of the works in sentences has not been shown to be useful for information handling and reasoning. We think it is an unnecessary and misleading intermediate form. Our sentence representation is semantic based in terms of attribute, attribute value, object class, object instance, and relation. Every sentence is segmented into one or more components with the form: 'attribute' of 'object' 'relation' 'attribute value'. Using only one format for all information gives the system simplicity and good performance as a RISC architecture does for hardware. The attribute-of-object representation is not new; it is used extensively in relational databases and knowledge-based systems. However, we will show that it can be used as a meaning representation for natural language sentences with minor extensions. In this paper we describe how a computer system can parse English sentences into this representation and generate English sentences from this representation. Much of this has been tested with computer implementation.
Quantifying Data Quality for Clinical Trials Using Electronic Data Capture
Nahm, Meredith L.; Pieper, Carl F.; Cunningham, Maureen M.
2008-01-01
Background Historically, only partial assessments of data quality have been performed in clinical trials, for which the most common method of measuring database error rates has been to compare the case report form (CRF) to database entries and count discrepancies. Importantly, errors arising from medical record abstraction and transcription are rarely evaluated as part of such quality assessments. Electronic Data Capture (EDC) technology has had a further impact, as paper CRFs typically leveraged for quality measurement are not used in EDC processes. Methods and Principal Findings The National Institute on Drug Abuse Treatment Clinical Trials Network has developed, implemented, and evaluated methodology for holistically assessing data quality on EDC trials. We characterize the average source-to-database error rate (14.3 errors per 10,000 fields) for the first year of use of the new evaluation method. This error rate was significantly lower than the average of published error rates for source-to-database audits, and was similar to CRF-to-database error rates reported in the published literature. We attribute this largely to an absence of medical record abstraction on the trials we examined, and to an outpatient setting characterized by less acute patient conditions. Conclusions Historically, medical record abstraction is the most significant source of error by an order of magnitude, and should be measured and managed during the course of clinical trials. Source-to-database error rates are highly dependent on the amount of structured data collection in the clinical setting and on the complexity of the medical record, dependencies that should be considered when developing data quality benchmarks. PMID:18725958
Toward designing for trust in database automation
DOE Office of Scientific and Technical Information (OSTI.GOV)
Duez, P. P.; Jamieson, G. A.
Appropriate reliance on system automation is imperative for safe and productive work, especially in safety-critical systems. It is unsafe to rely on automation beyond its designed use; conversely, it can be both unproductive and unsafe to manually perform tasks that are better relegated to automated tools. Operator trust in automated tools mediates reliance, and trust appears to affect how operators use technology. As automated agents become more complex, the question of trust in automation is increasingly important. In order to achieve proper use of automation, we must engender an appropriate degree of trust that is sensitive to changes in operatingmore » functions and context. In this paper, we present research concerning trust in automation in the domain of automated tools for relational databases. Lee and See have provided models of trust in automation. One model developed by Lee and See identifies three key categories of information about the automation that lie along a continuum of attributional abstraction. Purpose-, process-and performance-related information serve, both individually and through inferences between them, to describe automation in such a way as to engender r properly-calibrated trust. Thus, one can look at information from different levels of attributional abstraction as a general requirements analysis for information key to appropriate trust in automation. The model of information necessary to engender appropriate trust in automation [1] is a general one. Although it describes categories of information, it does not provide insight on how to determine the specific information elements required for a given automated tool. We have applied the Abstraction Hierarchy (AH) to this problem in the domain of relational databases. The AH serves as a formal description of the automation at several levels of abstraction, ranging from a very abstract purpose-oriented description to a more concrete description of the resources involved in the automated process. The connection between an AH for an automated tool and a list of information elements at the three levels of attributional abstraction is then direct, providing a method for satisfying information requirements for appropriate trust in automation. In this paper, we will present our method for developing specific information requirements for an automated tool, based on a formal analysis of that tool and the models presented by Lee and See. We will show an example of the application of the AH to automation, in the domain of relational database automation, and the resulting set of specific information elements for appropriate trust in the automated tool. Finally, we will comment on the applicability of this approach to the domain of nuclear plant instrumentation. (authors)« less
McCormick-Huhn, John M; Chen, Hui; Wyble, Bradley P; Dennis, Nancy A
2018-02-01
Previous work has shown mixed evidence regarding age-related deficits for binding in working memory. The current study used the newly developed attribute amnesia effect (H. Chen & Wyble, 2015a) to test the associative-deficit hypothesis during working memory and to probe whether hyper-binding extends to include binding of de-selected information. In studies of attribute amnesia, participants use target attributes (e.g., identity, color) to demonstrate near ceiling levels of reporting of a second target attribute (e.g., location) across a series of trials (H. Chen & Wyble, 2015a, 2016). Yet, despite having just processed the target-defining attribute, they have difficulty reporting it on a surprise trial. This effect provides several predictions for associative binding in aging. The associative-deficit hypothesis predicts age-related decline on the surprise trial, whereas an extension of hyper-binding predicts age-related increase in performance in older adults. In Experiment 1, when working memory load was low, older adults demonstrated attribute amnesia equal to that found in younger adults. When load increased in Experiment 2, older adults again demonstrated attribute amnesia as well as an age deficit for reporting target attributes. In lieu of spontaneous binding, results suggest that expectancy plays a critical role in older adults' propensity to encode and bind target attributes in working memory. Results further suggest that expectancy alone is not enough for older adults to form bound representations when task demands are high. Taken together results revealed a boundary condition of hyper-binding and further provided conditional support for the associative-deficit hypothesis in working memory. (PsycINFO Database Record (c) 2018 APA, all rights reserved).
Reflections on CD-ROM: Bridging the Gap between Technology and Purpose.
ERIC Educational Resources Information Center
Saviers, Shannon Smith
1987-01-01
Provides a technological overview of CD-ROM (Compact Disc-Read Only Memory), an optically-based medium for data storage offering large storage capacity, computer-based delivery system, read-only medium, and economic mass production. CD-ROM database attributes appropriate for information delivery are also reviewed, including large database size,…
Kannry, J L; Wright, L; Shifman, M; Silverstein, S; Miller, P L
1996-01-01
OBJECTIVE: To examine the issues involved in mapping an existing structured controlled vocabulary, the Medical Entities Dictionary (MED) developed at Columbia University, to an institutional vocabulary, the laboratory and pharmacy vocabularies of the Yale New Haven Medical Center. DESIGN: 200 Yale pharmacy terms and 200 Yale laboratory terms were randomly selected from database files containing all of the Yale laboratory and pharmacy terms. These 400 terms were then mapped to the MED in three phases: mapping terms, mapping relationships between terms, and mapping attributes that modify terms. RESULTS: 73% of the Yale pharmacy terms mapped to MED terms. 49% of the Yale laboratory terms mapped to MED terms. After certain obsolete and otherwise inappropriate laboratory terms were eliminated, the latter rate improved to 59%. 23% of the unmatched Yale laboratory terms failed to match because of differences in granularity with MED terms. The Yale and MED pharmacy terms share 12 of 30 distinct attributes. The Yale and MED laboratory terms share 14 of 23 distinct attributes. CONCLUSION: The mapping of an institutional vocabulary to a structured controlled vocabulary requires that the mapping be performed at the level of terms, relationships, and attributes. The mapping process revealed the importance of standardization of local vocabulary subsets, standardization of attribute representation, and term granularity. PMID:8750391
Mohammed, Mohammed A; Panesar, Jagdeep S; Laney, David B; Wilson, Richard
2013-04-01
The use of statistical process control (SPC) charts in healthcare is increasing. The primary purpose of SPC is to distinguish between common-cause variation which is attributable to the underlying process, and special-cause variation which is extrinsic to the underlying process. This is important because improvement under common-cause variation requires action on the process, whereas special-cause variation merits an investigation to first find the cause. Nonetheless, when dealing with attribute or count data (eg, number of emergency admissions) involving very large sample sizes, traditional SPC charts often produce tight control limits with most of the data points appearing outside the control limits. This can give a false impression of common and special-cause variation, and potentially misguide the user into taking the wrong actions. Given the growing availability of large datasets from routinely collected databases in healthcare, there is a need to present a review of this problem (which arises because traditional attribute charts only consider within-subgroup variation) and its solutions (which consider within and between-subgroup variation), which involve the use of the well-established measurements chart and the more recently developed attribute charts based on Laney's innovative approach. We close by making some suggestions for practice.
Multi-source and ontology-based retrieval engine for maize mutant phenotypes
Green, Jason M.; Harnsomburana, Jaturon; Schaeffer, Mary L.; Lawrence, Carolyn J.; Shyu, Chi-Ren
2011-01-01
Model Organism Databases, including the various plant genome databases, collect and enable access to massive amounts of heterogeneous information, including sequence data, gene product information, images of mutant phenotypes, etc, as well as textual descriptions of many of these entities. While a variety of basic browsing and search capabilities are available to allow researchers to query and peruse the names and attributes of phenotypic data, next-generation search mechanisms that allow querying and ranking of text descriptions are much less common. In addition, the plant community needs an innovative way to leverage the existing links in these databases to search groups of text descriptions simultaneously. Furthermore, though much time and effort have been afforded to the development of plant-related ontologies, the knowledge embedded in these ontologies remains largely unused in available plant search mechanisms. Addressing these issues, we have developed a unique search engine for mutant phenotypes from MaizeGDB. This advanced search mechanism integrates various text description sources in MaizeGDB to aid a user in retrieving desired mutant phenotype information. Currently, descriptions of mutant phenotypes, loci and gene products are utilized collectively for each search, though expansion of the search mechanism to include other sources is straightforward. The retrieval engine, to our knowledge, is the first engine to exploit the content and structure of available domain ontologies, currently the Plant and Gene Ontologies, to expand and enrich retrieval results in major plant genomic databases. Database URL: http:www.PhenomicsWorld.org/QBTA.php PMID:21558151
Maternal sensitivity: a concept analysis.
Shin, Hyunjeong; Park, Young-Joo; Ryu, Hosihn; Seomun, Gyeong-Ae
2008-11-01
The aim of this paper is to report a concept analysis of maternal sensitivity. Maternal sensitivity is a broad concept encompassing a variety of interrelated affective and behavioural caregiving attributes. It is used interchangeably with the terms maternal responsiveness or maternal competency, with no consistency of use. There is a need to clarify the concept of maternal sensitivity for research and practice. A search was performed on the CINAHL and Ovid MEDLINE databases using 'maternal sensitivity', 'maternal responsiveness' and 'sensitive mothering' as key words. The searches yielded 54 records for the years 1981-2007. Rodgers' method of evolutionary concept analysis was used to analyse the material. Four critical attributes of maternal sensitivity were identified: (a) dynamic process involving maternal abilities; (b) reciprocal give-and-take with the infant; (c) contingency on the infant's behaviour and (d) quality of maternal behaviours. Maternal identity and infant's needs and cues are antecedents for these attributes. The consequences are infant's comfort, mother-infant attachment and infant development. In addition, three positive affecting factors (social support, maternal-foetal attachment and high self-esteem) and three negative affecting factors (maternal depression, maternal stress and maternal anxiety) were identified. A clear understanding of the concept of maternal sensitivity could be useful for developing ways to enhance maternal sensitivity and to maximize the developmental potential of infants. Knowledge of the attributes of maternal sensitivity identified in this concept analysis may be helpful for constructing measuring items or dimensions.
NASA Astrophysics Data System (ADS)
Thessen, Anne E.; McGinnis, Sean; North, Elizabeth W.
2016-02-01
Process studies and coupled-model validation efforts in geosciences often require integration of multiple data types across time and space. For example, improved prediction of hydrocarbon fate and transport is an important societal need which fundamentally relies upon synthesis of oceanography and hydrocarbon chemistry. Yet, there are no publically accessible databases which integrate these diverse data types in a georeferenced format, nor are there guidelines for developing such a database. The objective of this research was to analyze the process of building one such database to provide baseline information on data sources and data sharing and to document the challenges and solutions that arose during this major undertaking. The resulting Deepwater Horizon Database was approximately 2.4 GB in size and contained over 8 million georeferenced data points collected from industry, government databases, volunteer networks, and individual researchers. The major technical challenges that were overcome were reconciliation of terms, units, and quality flags which were necessary to effectively integrate the disparate data sets. Assembling this database required the development of relationships with individual researchers and data managers which often involved extensive e-mail contacts. The average number of emails exchanged per data set was 7.8. Of the 95 relevant data sets that were discovered, 38 (40%) were obtained, either in whole or in part. Over one third (36%) of the requests for data went unanswered. The majority of responses were received after the first request (64%) and within the first week of the first request (67%). Although fewer than half of the potentially relevant datasets were incorporated into the database, the level of sharing (40%) was high compared to some other disciplines where sharing can be as low as 10%. Our suggestions for building integrated databases include budgeting significant time for e-mail exchanges, being cognizant of the cost versus benefits of pursuing reticent data providers, and building trust through clear, respectful communication and with flexible and appropriate attributions.
Cultural Aspects of Secrecy in Global Economy
NASA Astrophysics Data System (ADS)
Serradell-Lopez, Enric; Cavaller, Victor
The main objective of this paper is to provide greater understanding of the nature of secrecy in firms. It presents an effort to develop some links between management of the secrecy and its relationship with culture. Using measures from Hofstede's work, we have linked some dimensions of national culture with CIS 3 UE survey database. The results show that some attributes of the culture as Masculinity and Uncertainty Avoidance have impact on the tendency of the firms for not to patent and maintain secrecy of their innovations.
An integrated data-analysis and database system for AMS 14C
NASA Astrophysics Data System (ADS)
Kjeldsen, Henrik; Olsen, Jesper; Heinemeier, Jan
2010-04-01
AMSdata is the name of a combined database and data-analysis system for AMS 14C and stable-isotope work that has been developed at Aarhus University. The system (1) contains routines for data analysis of AMS and MS data, (2) allows a flexible and accurate description of sample extraction and pretreatment, also when samples are split into several fractions, and (3) keeps track of all measured, calculated and attributed data. The structure of the database is flexible and allows an unlimited number of measurement and pretreatment procedures. The AMS 14C data analysis routine is fairly advanced and flexible, and it can be easily optimized for different kinds of measuring processes. Technically, the system is based on a Microsoft SQL server and includes stored SQL procedures for the data analysis. Microsoft Office Access is used for the (graphical) user interface, and in addition Excel, Word and Origin are exploited for input and output of data, e.g. for plotting data during data analysis.
Spatiotemporal conceptual platform for querying archaeological information systems
NASA Astrophysics Data System (ADS)
Partsinevelos, Panagiotis; Sartzetaki, Mary; Sarris, Apostolos
2015-04-01
Spatial and temporal distribution of archaeological sites has been shown to associate with several attributes including marine, water, mineral and food resources, climate conditions, geomorphological features, etc. In this study, archeological settlement attributes are evaluated under various associations in order to provide a specialized query platform in a geographic information system (GIS). Towards this end, a spatial database is designed to include a series of archaeological findings for a secluded geographic area of Crete in Greece. The key categories of the geodatabase include the archaeological type (palace, burial site, village, etc.), temporal information of the habitation/usage period (pre Minoan, Minoan, Byzantine, etc.), and the extracted geographical attributes of the sites (distance to sea, altitude, resources, etc.). Most of the related spatial attributes are extracted with readily available GIS tools. Additionally, a series of conceptual data attributes are estimated, including: Temporal relation of an era to a future one in terms of alteration of the archaeological type, topologic relations of various types and attributes, spatial proximity relations between various types. These complex spatiotemporal relational measures reveal new attributes towards better understanding of site selection for prehistoric and/or historic cultures, yet their potential combinations can become numerous. Therefore, after the quantification of the above mentioned attributes, they are classified as of their importance for archaeological site location modeling. Under this new classification scheme, the user may select a geographic area of interest and extract only the important attributes for a specific archaeological type. These extracted attributes may then be queried against the entire spatial database and provide a location map of possible new archaeological sites. This novel type of querying is robust since the user does not have to type a standard SQL query but graphically select an area of interest. In addition, according to the application at hand, novel spatiotemporal attributes and relations can be supported, towards the understanding of historical settlement patterns.
Case retrieval in medical databases by fusing heterogeneous information.
Quellec, Gwénolé; Lamard, Mathieu; Cazuguel, Guy; Roux, Christian; Cochener, Béatrice
2011-01-01
A novel content-based heterogeneous information retrieval framework, particularly well suited to browse medical databases and support new generation computer aided diagnosis (CADx) systems, is presented in this paper. It was designed to retrieve possibly incomplete documents, consisting of several images and semantic information, from a database; more complex data types such as videos can also be included in the framework. The proposed retrieval method relies on image processing, in order to characterize each individual image in a document by their digital content, and information fusion. Once the available images in a query document are characterized, a degree of match, between the query document and each reference document stored in the database, is defined for each attribute (an image feature or a metadata). A Bayesian network is used to recover missing information if need be. Finally, two novel information fusion methods are proposed to combine these degrees of match, in order to rank the reference documents by decreasing relevance for the query. In the first method, the degrees of match are fused by the Bayesian network itself. In the second method, they are fused by the Dezert-Smarandache theory: the second approach lets us model our confidence in each source of information (i.e., each attribute) and take it into account in the fusion process for a better retrieval performance. The proposed methods were applied to two heterogeneous medical databases, a diabetic retinopathy database and a mammography screening database, for computer aided diagnosis. Precisions at five of 0.809 ± 0.158 and 0.821 ± 0.177, respectively, were obtained for these two databases, which is very promising.
Quality Attribute-Guided Evaluation of NoSQL Databases: A Case Study
2015-01-16
evaluations of NoSQL databases specifically, and big data systems in general, that have become apparent during our study. Keywords—NoSQL, distributed...technology, namely that of big data , software systems [1]. At the heart of big data systems are a collection of database technologies that are more...born organizations such as Google and Amazon [3][4], along with those of numerous other big data innovators, have created a variety of open source and
Smith, W Brad; Cuenca Lara, Rubí Angélica; Delgado Caballero, Carina Edith; Godínez Valdivia, Carlos Isaías; Kapron, Joseph S; Leyva Reyes, Juan Carlos; Meneses Tovar, Carmen Lourdes; Miles, Patrick D; Oswalt, Sonja N; Ramírez Salgado, Mayra; Song, Xilong Alex; Stinson, Graham; Villela Gaytán, Sergio Armando
2018-05-21
Forests cannot be managed sustainably without reliable data to inform decisions. National Forest Inventories (NFI) tend to report national statistics, with sub-national stratification based on domestic ecological classification systems. It is becoming increasingly important to be able to report statistics on ecosystems that span international borders, as global change and globalization expand stakeholders' spheres of concern. The state of a transnational ecosystem can only be properly assessed by examining the entire ecosystem. In global forest resource assessments, it may be useful to break national statistics down by ecosystem, especially for large countries. The Inventory and Monitoring Working Group (IMWG) of the North American Forest Commission (NAFC) has begun developing a harmonized North American Forest Database (NAFD) for managing forest inventory data, enabling consistent, continental-scale forest assessment supporting ecosystem-level reporting and relational queries. The first iteration of the database contains data describing 1.9 billion ha, including 677.5 million ha of forest. Data harmonization is made challenging by the existence of definitions and methodologies tailored to suit national circumstances, emerging from each country's professional forestry development. This paper reports the methods used to synchronize three national forest inventories, starting with a small suite of variables and attributes.
ERIC Educational Resources Information Center
Kancherla, Vijaya; Amendah, Djesika D.; Grosse, Scott D.; Yeargin-Allsopp, Marshalyn; Van Naarden Braun, Kim
2012-01-01
This study estimated medical expenditures attributable to cerebral palsy (CP) among children enrolled in Medicaid, stratified by the presence of co-occurring intellectual disability (ID), relative to children without CP or ID. The MarketScan[R] Medicaid Multi-State database was used to identify children with CP for 2003-2005 by using the…
Morphological biosignatures and the search for life on Mars.
Cady, Sherry L; Farmer, Jack D; Grotzinger, John P; Schopf, J William; Steele, Andrew
2003-01-01
This report provides a rationale for the advances in instrumentation and understanding needed to assess claims of ancient and extraterrestrial life made on the basis of morphological biosignatures. Morphological biosignatures consist of bona fide microbial fossils as well as microbially influenced sedimentary structures. To be recognized as evidence of life, microbial fossils must contain chemical and structural attributes uniquely indicative of microbial cells or cellular or extracellular processes. When combined with various research strategies, high-resolution instruments can reveal such attributes and elucidate how morphological fossils form and become altered, thereby improving the ability to recognize them in the geological record on Earth or other planets. Also, before fossilized microbially influenced sedimentary structures can provide evidence of life, criteria to distinguish their biogenic from non-biogenic attributes must be established. This topic can be advanced by developing process-based models. A database of images and spectroscopic data that distinguish the suite of bona fide morphological biosignatures from their abiotic mimics will avoid detection of false-positives for life. The use of high-resolution imaging and spectroscopic instruments, in conjunction with an improved knowledge base of the attributes that demonstrate life, will maximize our ability to recognize and assess the biogenicity of extraterrestrial and ancient terrestrial life.
FDA perspective on specifications for biotechnology products--from IND to PLA.
Murano, G
1997-01-01
Quality standards are obligatory throughout development, approval and post-marketing phases of biotechnology-derived products, thus assuring product identity, purity, and potency/strength. The process of developing and setting specifications should be based on sound science and should represent a logical progression of actions based on the use of experiential data spanning manufacturing process validation, consistency in production, and characterization of relevant product properties/attributes, by multiple analytical means. This interactive process occurs in phases, varying in rigour. It is best described as encompassing a framework which starts with the implementation of realistic/practical operational quality limits, progressing to the establishment/adoption of more stringent specifications. The historical database is generated from preclinical, toxicology and early clinical lots. This supports the clinical development programme which, as it progresses, allows for further assay method validation/refinement, adoption/addition due to relevant or newly recognized product attributes or rejection due to irrelevance. In the next phase, (licensing/approval) specifications are set through extended experience and validation of both the preparative and analytical processes, to include availability of suitable reference standards and extensive product characterization throughout its proposed dating period. Subsequent to product approval, the incremental database of test results serves as a natural continuum for further evolving/refining specifications. While there is considerable latitude in the kinds of testing modalities finally adopted to establish product quality on a routine basis, for both drugs and drug products, it is important that the selection takes into consideration relevant (significant) product characteristics that appropriately reflect on identity, purity and potency.
Design of Integrated Database on Mobile Information System: A Study of Yogyakarta Smart City App
NASA Astrophysics Data System (ADS)
Nurnawati, E. K.; Ermawati, E.
2018-02-01
An integration database is a database which acts as the data store for multiple applications and thus integrates data across these applications (in contrast to an Application Database). An integration database needs a schema that takes all its client applications into account. The benefit of the schema that sharing data among applications does not require an extra layer of integration services on the applications. Any changes to data made in a single application are made available to all applications at the time of database commit - thus keeping the applications’ data use better synchronized. This study aims to design and build an integrated database that can be used by various applications in a mobile device based system platforms with the based on smart city system. The built-in database can be used by various applications, whether used together or separately. The design and development of the database are emphasized on the flexibility, security, and completeness of attributes that can be used together by various applications to be built. The method used in this study is to choice of the appropriate database logical structure (patterns of data) and to build the relational-database models (Design Databases). Test the resulting design with some prototype apps and analyze system performance with test data. The integrated database can be utilized both of the admin and the user in an integral and comprehensive platform. This system can help admin, manager, and operator in managing the application easily and efficiently. This Android-based app is built based on a dynamic clientserver where data is extracted from an external database MySQL. So if there is a change of data in the database, then the data on Android applications will also change. This Android app assists users in searching of Yogyakarta (as smart city) related information, especially in culture, government, hotels, and transportation.
Rouigari, Maedeh; Dehbashi, Moein; Ghaedi, Kamran; Pourhossein, Meraj
2018-07-01
For the first time, we used molecular signaling pathway enrichment analysis to determine possible involvement of miR-126 and IRS-1 in neurotrophin pathway. In this prospective study, Validated and predicted targets (targetome) of miR-126 were collected following searching miRtarbase (http://mirtarbase.mbc.nctu.edu.tw/) and miRWalk 2.0 databases, respectively. Then, approximate expression of miR-126 targeting in Glioma tissue was examined using UniGene database (http://www.ncbi. nlm.nih.gov/unigene). In silico molecular pathway enrichment analysis was carried out by DAVID 6.7 database (http://david. abcc.ncifcrf.gov/) to explore which signaling pathway is related to miR-126 targeting and how miR-126 attributes to glioma development. MiR-126 exerts a variety of functions in cancer pathogenesis via suppression of expression of target gene including PI3K, KRAS, EGFL7, IRS-1 and VEGF. Our bioinformatic studies implementing DAVID database, showed the involvement of miR-126 target genes in several signaling pathways including cancer pathogenesis, neurotrophin functions, Glioma formation, insulin function, focal adhesion production, chemokine synthesis and secretion and regulation of the actin cytoskeleton. Taken together, we concluded that miR-126 enhances the formation of glioma cancer stem cell probably via down regulation of IRS-1 in neurotrophin signaling pathway. Copyright© by Royan Institute. All rights reserved.
A manual-based psychodynamic therapy for treatment-resistant borderline personality disorder.
Gregory, Robert J; Remen, Anna L
2008-03-01
The authors introduce a manual-based treatment, labeled dynamic deconstructive psychotherapy, developed for those patients with borderline personality disorder who are most difficult to engage in therapy, such as those having co-occurring substance use disorders. This treatment model is based on the hypothesis that borderline pathology and related behaviors reflect impairment in specific neurocognitive functions, including association, attribution, and alterity that form the basis for a coherent and differentiated self. Dynamic deconstructive psychotherapy aims to activate and remediate neurocognitive self-capacities by facilitating elaboration of affect-laden interpersonal experiences and integration of attributions, as well as providing novel experiences in the patient-therapist relationship that promote self-other differentiation. Treatment involves weekly individual sessions for a predetermined period of time and follows sequential stages. (PsycINFO Database Record (c) 2010 APA, all rights reserved).
Magnetic Fields for All: The GPIPS Community Web-Access Portal
NASA Astrophysics Data System (ADS)
Carveth, Carol; Clemens, D. P.; Pinnick, A.; Pavel, M.; Jameson, K.; Taylor, B.
2007-12-01
The new GPIPS website portal provides community users with an intuitive and powerful interface to query the data products of the Galactic Plane Infrared Polarization Survey. The website, which was built using PHP for the front end and MySQL for the database back end, allows users to issue queries based on galactic or equatorial coordinates, GPIPS-specific identifiers, polarization information, magnitude information, and several other attributes. The returns are presented in HTML tables, with the added option of either downloading or being emailed an ASCII file including the same or more information from the database. Other functionalities of the website include providing details of the status of the Survey (which fields have been observed or are planned to be observed), techniques involved in data collection and analysis, and descriptions of the database contents and names. For this initial launch of the website, users may access the GPIPS polarization point source catalog and the deep coadd photometric point source catalog. Future planned developments include a graphics-based method for querying the database, as well as tools to combine neighboring GPIPS images into larger image files for both polarimetry and photometry. This work is partially supported by NSF grant AST-0607500.
Paradise: A Parallel Information System for EOSDIS
NASA Technical Reports Server (NTRS)
DeWitt, David
1996-01-01
The Paradise project was begun-in 1993 in order to explore the application of the parallel and object-oriented database system technology developed as a part of the Gamma, Exodus. and Shore projects to the design and development of a scaleable, geo-spatial database system for storing both massive spatial and satellite image data sets. Paradise is based on an object-relational data model. In addition to the standard attribute types such as integers, floats, strings and time, Paradise also provides a set of and multimedia data types, designed to facilitate the storage and querying of complex spatial and multimedia data sets. An individual tuple can contain any combination of this rich set of data types. For example, in the EOSDIS context, a tuple might mix terrain and map data for an area along with the latest satellite weather photo of the area. The use of a geo-spatial metaphor simplifies the task of fusing disparate forms of data from multiple data sources including text, image, map, and video data sets.
Managerial coaching: a concept analysis.
Batson, Vicki D; Yoder, Linda H
2012-07-01
This article presents a report of a concept analysis of managerial coaching. Managerial coaching has been identified as a means for managers to give support to staff nurses, however, no clear delineation of what behaviours and attributes constitute managerial coaching or differentiate it from other career development relationships is provided in the current nursing literature. The CINAHL, ProQuest, Business Source Complete and PscyhIFNO databases were searched for articles published between 1980-2009 using the keywords coaching, managerial coaching, nurse manager support, nursing leadership, self-efficacy, work environment and empowerment. A hybrid approach was used, incorporating both Walker and Avant's method of concept analysis and King's conceptual system and Theory of Goal Attainment to explore the meaning of managerial coaching. Inclusive years of search ranged from 1980-2009. Managerial coaching is a specific dyadic relationship between the nurse manager and staff nurse intended to improve skills and knowledge as they relate to expected job performance. Antecedents and consequences are categorized at the individual and organizational level. Defining attributes, empirical referents and a model case are presented. The theoretical definition for this concept helps to differentiate it from other types of career development relationships and will give a basis for nurse managers to understand what skills and attributes are necessary to establish an effective managerial coaching relationship with staff nurses. Conceptualization will also assist in developing empirical studies examining managerial coaching behaviours in the work environment. © 2012 Blackwell Publishing Ltd.
High-Performance Data Analytics Beyond the Relational and Graph Data Models with GEMS
DOE Office of Scientific and Technical Information (OSTI.GOV)
Castellana, Vito G.; Minutoli, Marco; Bhatt, Shreyansh
Graphs represent an increasingly popular data model for data-analytics, since they can naturally represent relationships and interactions between entities. Relational databases and their pure table-based data model are not well suitable to store and process sparse data. Consequently, graph databases have gained interest in the last few years and the Resource Description Framework (RDF) became the standard data model for graph data. Nevertheless, while RDF is well suited to analyze the relationships between the entities, it is not efficient in representing their attributes and properties. In this work we propose the adoption of a new hybrid data model, based onmore » attributed graphs, that aims at overcoming the limitations of the pure relational and graph data models. We present how we have re-designed the GEMS data-analytics framework to fully take advantage of the proposed hybrid data model. To improve analysts productivity, in addition to a C++ API for applications development, we adopt GraQL as input query language. We validate our approach implementing a set of queries on net-flow data and we compare our framework performance against Neo4j. Experimental results show significant performance improvement over Neo4j, up to several orders of magnitude when increasing the size of the input data.« less
Kalish, Yuval; Luria, Gil
2016-10-01
Research into leadership emergence typically focuses on the attributes of the emergent leader. By considering also the attributes of perceivers and the passage of time, we develop a more complete theory of leadership emergence in short-lived groups. Using expectation states theory as an overarching theoretical framework, and integrating it with the surface- and deep-level diversity literature and with theories of self-serving biases, we examine the predictors of leadership emergence in short timeframes. We conduct a field study in a military assessment boot camp (a pilot study, n = 60; and a main study, n = 89). We use cross-sectional and longitudinal exponential random graph models to analyze data on participants' abilities and on their perceptions of who, in their respective groups, were "leaders." We find that the criteria by which people perceive leadership in others change over time, from easily noticeable attributes to covert leadership-relevant attributes, and that people also rely on leadership-relevant attributes that they possess at high levels to inform their perceptions of leadership in others. The integration of expectation states theory, attribute salience over time and theories of self-serving bias is needed for a full understanding of leadership emergence in groups, because perceivers' own abilities are instrumental in shaping their perceptions of emergent leadership over time. Theoretical and practical implications are discussed. (PsycINFO Database Record (c) 2016 APA, all rights reserved).
Perception: a concept analysis.
McDonald, Susan M
2012-02-01
Concept analysis methodology by Walker and Avant (2005) was used to define, describe, and delimit the concept of perception. Nursing literature in the Medline database was searched for definitions of "perception." Definitions, uses, and defining attributes of perception were identified; model and contrary cases were developed; and antecedents, consequences, and empirical referents were determined. An operational definition for the concept was developed. Nurses need to be cognizant of how perceptual differences impact the delivery of nursing care. In research, a mixed methodology approach may yield a richer description of the phenomenon and provide useful information for clinical practice. © 2011, The Author. International Journal of Nursing Knowledge © 2011, NANDA International.
Malware detection and analysis
DOE Office of Scientific and Technical Information (OSTI.GOV)
Chiang, Ken; Lloyd, Levi; Crussell, Jonathan
Embodiments of the invention describe systems and methods for malicious software detection and analysis. A binary executable comprising obfuscated malware on a host device may be received, and incident data indicating a time when the binary executable was received and identifying processes operating on the host device may be recorded. The binary executable is analyzed via a scalable plurality of execution environments, including one or more non-virtual execution environments and one or more virtual execution environments, to generate runtime data and deobfuscation data attributable to the binary executable. At least some of the runtime data and deobfuscation data attributable tomore » the binary executable is stored in a shared database, while at least some of the incident data is stored in a private, non-shared database.« less
Privacy-Preserving Classifier Learning
NASA Astrophysics Data System (ADS)
Brickell, Justin; Shmatikov, Vitaly
We present an efficient protocol for the privacy-preserving, distributed learning of decision-tree classifiers. Our protocol allows a user to construct a classifier on a database held by a remote server without learning any additional information about the records held in the database. The server does not learn anything about the constructed classifier, not even the user’s choice of feature and class attributes.
Rember, William C.; Bennett, Earl H.
2001-01-01
he paper geologic map of the east part of the Pullman 1·x 2· degree quadrangle, Idaho (Rember and Bennett, 1979) was scanned and initially attributed by Optronics Specialty Co., Inc. (Northridge, CA) and remitted to the U.S. Geological Survey for further attribution and publication of the geospatial digital files. The resulting digital geologic map GIS can be queried in many ways to produce a variety of geologic maps. This digital geospatial database is one of many being created by the U.S. Geological Survey as an ongoing effort to provide geologic information in a geographic information system (GIS) for use in spatial analysis. Digital base map data files (topography, roads, towns, rivers and lakes, and others.) are not included: they may be obtained from a variety of commercial and government sources. This database is not meant to be used or displayed at any scale larger than 1:250,000 (for example, 1:100,000 or 1:24,000). The digital geologic map graphics and plot files (pull250k.gra/.hp /.eps) that are provided in the digital package are representations of the digital database.
New Zealand's National Landslide Database
NASA Astrophysics Data System (ADS)
Rosser, B.; Dellow, S.; Haubrook, S.; Glassey, P.
2016-12-01
Since 1780, landslides have caused an average of about 3 deaths a year in New Zealand and have cost the economy an average of at least NZ$250M/a (0.1% GDP). To understand the risk posed by landslide hazards to society, a thorough knowledge of where, when and why different types of landslides occur is vital. The main objective for establishing the database was to provide a centralised national-scale, publically available database to collate landslide information that could be used for landslide hazard and risk assessment. Design of a national landslide database for New Zealand required consideration of both existing landslide data stored in a variety of digital formats, and future data, yet to be collected. Pre-existing databases were developed and populated with data reflecting the needs of the landslide or hazard project, and the database structures of the time. Bringing these data into a single unified database required a new structure capable of storing and delivering data at a variety of scales and accuracy and with different attributes. A "unified data model" was developed to enable the database to hold old and new landslide data irrespective of scale and method of capture. The database contains information on landslide locations and where available: 1) the timing of landslides and the events that may have triggered them; 2) the type of landslide movement; 3) the volume and area; 4) the source and debris tail; and 5) the impacts caused by the landslide. Information from a variety of sources including aerial photographs (and other remotely sensed data), field reconnaissance and media accounts has been collated and is presented for each landslide along with metadata describing the data sources and quality. There are currently nearly 19,000 landslide records in the database that include point locations, polygons of landslide source and deposit areas, and linear features. Several large datasets are awaiting upload which will bring the total number of landslides to over 100,000. The geo-spatial database is publicly available via the Internet. Software components, including the underlying database (PostGIS), Web Map Server (GeoServer) and web application use open-source software. The hope is that others will add relevant information to the database as well as download the data contained in it.
Community detection in sequence similarity networks based on attribute clustering
Chowdhary, Janamejaya; Loeffler, Frank E.; Smith, Jeremy C.
2017-07-24
Networks are powerful tools for the presentation and analysis of interactions in multi-component systems. A commonly studied mesoscopic feature of networks is their community structure, which arises from grouping together similar nodes into one community and dissimilar nodes into separate communities. Here in this paper, the community structure of protein sequence similarity networks is determined with a new method: Attribute Clustering Dependent Communities (ACDC). Sequence similarity has hitherto typically been quantified by the alignment score or its expectation value. However, pair alignments with the same score or expectation value cannot thus be differentiated. To overcome this deficiency, the method constructs,more » for pair alignments, an extended alignment metric, the link attribute vector, which includes the score and other alignment characteristics. Rescaling components of the attribute vectors qualitatively identifies a systematic variation of sequence similarity within protein superfamilies. The problem of community detection is then mapped to clustering the link attribute vectors, selection of an optimal subset of links and community structure refinement based on the partition density of the network. ACDC-predicted communities are found to be in good agreement with gold standard sequence databases for which the "ground truth" community structures (or families) are known. ACDC is therefore a community detection method for sequence similarity networks based entirely on pair similarity information. A serial implementation of ACDC is available from https://cmb.ornl.gov/resources/developments« less
Community detection in sequence similarity networks based on attribute clustering
DOE Office of Scientific and Technical Information (OSTI.GOV)
Chowdhary, Janamejaya; Loeffler, Frank E.; Smith, Jeremy C.
Networks are powerful tools for the presentation and analysis of interactions in multi-component systems. A commonly studied mesoscopic feature of networks is their community structure, which arises from grouping together similar nodes into one community and dissimilar nodes into separate communities. Here in this paper, the community structure of protein sequence similarity networks is determined with a new method: Attribute Clustering Dependent Communities (ACDC). Sequence similarity has hitherto typically been quantified by the alignment score or its expectation value. However, pair alignments with the same score or expectation value cannot thus be differentiated. To overcome this deficiency, the method constructs,more » for pair alignments, an extended alignment metric, the link attribute vector, which includes the score and other alignment characteristics. Rescaling components of the attribute vectors qualitatively identifies a systematic variation of sequence similarity within protein superfamilies. The problem of community detection is then mapped to clustering the link attribute vectors, selection of an optimal subset of links and community structure refinement based on the partition density of the network. ACDC-predicted communities are found to be in good agreement with gold standard sequence databases for which the "ground truth" community structures (or families) are known. ACDC is therefore a community detection method for sequence similarity networks based entirely on pair similarity information. A serial implementation of ACDC is available from https://cmb.ornl.gov/resources/developments« less
The need for high-quality whole-genome sequence databases in microbial forensics.
Sjödin, Andreas; Broman, Tina; Melefors, Öjar; Andersson, Gunnar; Rasmusson, Birgitta; Knutsson, Rickard; Forsman, Mats
2013-09-01
Microbial forensics is an important part of a strengthened capability to respond to biocrime and bioterrorism incidents to aid in the complex task of distinguishing between natural outbreaks and deliberate acts. The goal of a microbial forensic investigation is to identify and criminally prosecute those responsible for a biological attack, and it involves a detailed analysis of the weapon--that is, the pathogen. The recent development of next-generation sequencing (NGS) technologies has greatly increased the resolution that can be achieved in microbial forensic analyses. It is now possible to identify, quickly and in an unbiased manner, previously undetectable genome differences between closely related isolates. This development is particularly relevant for the most deadly bacterial diseases that are caused by bacterial lineages with extremely low levels of genetic diversity. Whole-genome analysis of pathogens is envisaged to be increasingly essential for this purpose. In a microbial forensic context, whole-genome sequence analysis is the ultimate method for strain comparisons as it is informative during identification, characterization, and attribution--all 3 major stages of the investigation--and at all levels of microbial strain identity resolution (ie, it resolves the full spectrum from family to isolate). Given these capabilities, one bottleneck in microbial forensics investigations is the availability of high-quality reference databases of bacterial whole-genome sequences. To be of high quality, databases need to be curated and accurate in terms of sequences, metadata, and genetic diversity coverage. The development of whole-genome sequence databases will be instrumental in successfully tracing pathogens in the future.
Evolving Spiking Neural Networks for Recognition of Aged Voices.
Silva, Marco; Vellasco, Marley M B R; Cataldo, Edson
2017-01-01
The aging of the voice, known as presbyphonia, is a natural process that can cause great change in vocal quality of the individual. This is a relevant problem to those people who use their voices professionally, and its early identification can help determine a suitable treatment to avoid its progress or even to eliminate the problem. This work focuses on the development of a new model for the identification of aging voices (independently of their chronological age), using as input attributes parameters extracted from the voice and glottal signals. The proposed model, named Quantum binary-real evolving Spiking Neural Network (QbrSNN), is based on spiking neural networks (SNNs), with an unsupervised training algorithm, and a Quantum-Inspired Evolutionary Algorithm that automatically determines the most relevant attributes and the optimal parameters that configure the SNN. The QbrSNN model was evaluated in a database composed of 120 records, containing samples from three groups of speakers. The results obtained indicate that the proposed model provides better accuracy than other approaches, with fewer input attributes. Copyright © 2017 The Voice Foundation. Published by Elsevier Inc. All rights reserved.
SAbDab: the structural antibody database
Dunbar, James; Krawczyk, Konrad; Leem, Jinwoo; Baker, Terry; Fuchs, Angelika; Georges, Guy; Shi, Jiye; Deane, Charlotte M.
2014-01-01
Structural antibody database (SAbDab; http://opig.stats.ox.ac.uk/webapps/sabdab) is an online resource containing all the publicly available antibody structures annotated and presented in a consistent fashion. The data are annotated with several properties including experimental information, gene details, correct heavy and light chain pairings, antigen details and, where available, antibody–antigen binding affinity. The user can select structures, according to these attributes as well as structural properties such as complementarity determining region loop conformation and variable domain orientation. Individual structures, datasets and the complete database can be downloaded. PMID:24214988
Road Traffic Accident Analysis of Ajmer City Using Remote Sensing and GIS Technology
NASA Astrophysics Data System (ADS)
Bhalla, P.; Tripathi, S.; Palria, S.
2014-12-01
With advancement in technology, new and sophisticated models of vehicle are available and their numbers are increasing day by day. A traffic accident has multi-facet characteristics associated with it. In India 93% of crashes occur due to Human induced factor (wholly or partly). For proper traffic accident analysis use of GIS technology has become an inevitable tool. The traditional accident database is a summary spreadsheet format using codes and mileposts to denote location, type and severity of accidents. Geo-referenced accident database is location-referenced. It incorporates a GIS graphical interface with the accident information to allow for query searches on various accident attributes. Ajmer city, headquarter of Ajmer district, Rajasthan has been selected as the study area. According to Police records, 1531 accidents occur during 2009-2013. Maximum accident occurs in 2009 and the maximum death in 2013. Cars, jeeps, auto, pickup and tempo are mostly responsible for accidents and that the occurrence of accidents is mostly concentrated between 4PM to 10PM. GIS has proved to be a good tool for analyzing multifaceted nature of accidents. While road safety is a critical issue, yet it is handled in an adhoc manner. This Study is a demonstration of application of GIS for developing an efficient database on road accidents taking Ajmer City as a study. If such type of database is developed for other cities, a proper analysis of accidents can be undertaken and suitable management strategies for traffic regulation can be successfully proposed.
Wickham, James; Riitters, Kurt; Vogt, Peter; Costanza, Jennifer; Neale, Anne
2017-11-01
Landscape context is an important factor in restoration ecology, but the use of landscape context for site prioritization has not been as fully developed. We used morphological image processing to identify candidate ecological restoration areas based on their proximity to existing natural vegetation. We identified 1,102,720 candidate ecological restoration areas across the continental United States. Candidate ecological restoration areas were concentrated in the Great Plains and eastern United States. We populated the database of candidate ecological restoration areas with 17 attributes related to site content and context, including factors such as soil fertility and roads (site content), and number and area of potentially conjoined vegetated regions (site context) to facilitate its use for site prioritization. We demonstrate the utility of the database in the state of North Carolina, U.S.A. for a restoration objective related to restoration of water quality (mandated by the U.S. Clean Water Act), wetlands, and forest. The database will be made publicly available on the U.S. Environmental Protection Agency's EnviroAtlas website (http://enviroatlas.epa.gov) for stakeholders interested in ecological restoration.
Wickham, James; Riitters, Kurt; Vogt, Peter; Costanza, Jennifer; Neale, Anne
2018-01-01
Landscape context is an important factor in restoration ecology, but the use of landscape context for site prioritization has not been as fully developed. We used morphological image processing to identify candidate ecological restoration areas based on their proximity to existing natural vegetation. We identified 1,102,720 candidate ecological restoration areas across the continental United States. Candidate ecological restoration areas were concentrated in the Great Plains and eastern United States. We populated the database of candidate ecological restoration areas with 17 attributes related to site content and context, including factors such as soil fertility and roads (site content), and number and area of potentially conjoined vegetated regions (site context) to facilitate its use for site prioritization. We demonstrate the utility of the database in the state of North Carolina, U.S.A. for a restoration objective related to restoration of water quality (mandated by the U.S. Clean Water Act), wetlands, and forest. The database will be made publicly available on the U.S. Environmental Protection Agency's EnviroAtlas website (http://enviroatlas.epa.gov) for stakeholders interested in ecological restoration. PMID:29683130
A binary linear programming formulation of the graph edit distance.
Justice, Derek; Hero, Alfred
2006-08-01
A binary linear programming formulation of the graph edit distance for unweighted, undirected graphs with vertex attributes is derived and applied to a graph recognition problem. A general formulation for editing graphs is used to derive a graph edit distance that is proven to be a metric, provided the cost function for individual edit operations is a metric. Then, a binary linear program is developed for computing this graph edit distance, and polynomial time methods for determining upper and lower bounds on the solution of the binary program are derived by applying solution methods for standard linear programming and the assignment problem. A recognition problem of comparing a sample input graph to a database of known prototype graphs in the context of a chemical information system is presented as an application of the new method. The costs associated with various edit operations are chosen by using a minimum normalized variance criterion applied to pairwise distances between nearest neighbors in the database of prototypes. The new metric is shown to perform quite well in comparison to existing metrics when applied to a database of chemical graphs.
Rdesign: A data dictionary with relational database design capabilities in Ada
NASA Technical Reports Server (NTRS)
Lekkos, Anthony A.; Kwok, Teresa Ting-Yin
1986-01-01
Data Dictionary is defined to be the set of all data attributes, which describe data objects in terms of their intrinsic attributes, such as name, type, size, format and definition. It is recognized as the data base for the Information Resource Management, to facilitate understanding and communication about the relationship between systems applications and systems data usage and to help assist in achieving data independence by permitting systems applications to access data knowledge of the location or storage characteristics of the data in the system. A research and development effort to use Ada has produced a data dictionary with data base design capabilities. This project supports data specification and analysis and offers a choice of the relational, network, and hierarchical model for logical data based design. It provides a highly integrated set of analysis and design transformation tools which range from templates for data element definition, spreadsheet for defining functional dependencies, normalization, to logical design generator.
Nosql for Storage and Retrieval of Large LIDAR Data Collections
NASA Astrophysics Data System (ADS)
Boehm, J.; Liu, K.
2015-08-01
Developments in LiDAR technology over the past decades have made LiDAR to become a mature and widely accepted source of geospatial information. This in turn has led to an enormous growth in data volume. The central idea for a file-centric storage of LiDAR point clouds is the observation that large collections of LiDAR data are typically delivered as large collections of files, rather than single files of terabyte size. This split of the dataset, commonly referred to as tiling, was usually done to accommodate a specific processing pipeline. It makes therefore sense to preserve this split. A document oriented NoSQL database can easily emulate this data partitioning, by representing each tile (file) in a separate document. The document stores the metadata of the tile. The actual files are stored in a distributed file system emulated by the NoSQL database. We demonstrate the use of MongoDB a highly scalable document oriented NoSQL database for storing large LiDAR files. MongoDB like any NoSQL database allows for queries on the attributes of the document. As a specialty MongoDB also allows spatial queries. Hence we can perform spatial queries on the bounding boxes of the LiDAR tiles. Inserting and retrieving files on a cloud-based database is compared to native file system and cloud storage transfer speed.
Hydrologic Derivatives for Modeling and Analysis—A new global high-resolution database
Verdin, Kristine L.
2017-07-17
The U.S. Geological Survey has developed a new global high-resolution hydrologic derivative database. Loosely modeled on the HYDRO1k database, this new database, entitled Hydrologic Derivatives for Modeling and Analysis, provides comprehensive and consistent global coverage of topographically derived raster layers (digital elevation model data, flow direction, flow accumulation, slope, and compound topographic index) and vector layers (streams and catchment boundaries). The coverage of the data is global, and the underlying digital elevation model is a hybrid of three datasets: HydroSHEDS (Hydrological data and maps based on SHuttle Elevation Derivatives at multiple Scales), GMTED2010 (Global Multi-resolution Terrain Elevation Data 2010), and the SRTM (Shuttle Radar Topography Mission). For most of the globe south of 60°N., the raster resolution of the data is 3 arc-seconds, corresponding to the resolution of the SRTM. For the areas north of 60°N., the resolution is 7.5 arc-seconds (the highest resolution of the GMTED2010 dataset) except for Greenland, where the resolution is 30 arc-seconds. The streams and catchments are attributed with Pfafstetter codes, based on a hierarchical numbering system, that carry important topological information. This database is appropriate for use in continental-scale modeling efforts. The work described in this report was conducted by the U.S. Geological Survey in cooperation with the National Aeronautics and Space Administration Goddard Space Flight Center.
Context indexing of digital cardiac ultrasound records in PACS
NASA Astrophysics Data System (ADS)
Lobodzinski, S. Suave; Meszaros, Georg N.
1998-07-01
Recent wide adoption of the DICOM 3.0 standard by ultrasound equipment vendors created a need for practical clinical implementations of cardiac imaging study visualization, management and archiving, DICOM 3.0 defines only a logical and physical format for exchanging image data (still images, video, patient and study demographics). All DICOM compliant imaging studies must presently be archived on a 650 Mb recordable compact disk. This is a severe limitation for ultrasound applications where studies of 3 to 10 minutes long are a common practice. In addition, DICOM digital echocardiography objects require physiological signal indexing, content segmentation and characterization. Since DICOM 3.0 is an interchange standard only, it does not define how to database composite video objects. The goal of this research was therefore to address the issues of efficient storage, retrieval and management of DICOM compliant cardiac video studies in a distributed PACS environment. Our Web based implementation has the advantage of accommodating both DICOM defined entity-relation modules (equipment data, patient data, video format, etc.) in standard relational database tables and digital indexed video with its attributes in an object relational database. Object relational data model facilitates content indexing of full motion cardiac imaging studies through bi-directional hyperlink generation that tie searchable video attributes and related objects to individual video frames in the temporal domain. Benefits realized from use of bi-directionally hyperlinked data models in an object relational database include: (1) real time video indexing during image acquisition, (2) random access and frame accurate instant playback of previously recorded full motion imaging data, and (3) time savings from faster and more accurate access to data through multiple navigation mechanisms such as multidimensional queries on an index, queries on a hyperlink attribute, free search and browsing.
MimoSA: a system for minimotif annotation
2010-01-01
Background Minimotifs are short peptide sequences within one protein, which are recognized by other proteins or molecules. While there are now several minimotif databases, they are incomplete. There are reports of many minimotifs in the primary literature, which have yet to be annotated, while entirely novel minimotifs continue to be published on a weekly basis. Our recently proposed function and sequence syntax for minimotifs enables us to build a general tool that will facilitate structured annotation and management of minimotif data from the biomedical literature. Results We have built the MimoSA application for minimotif annotation. The application supports management of the Minimotif Miner database, literature tracking, and annotation of new minimotifs. MimoSA enables the visualization, organization, selection and editing functions of minimotifs and their attributes in the MnM database. For the literature components, Mimosa provides paper status tracking and scoring of papers for annotation through a freely available machine learning approach, which is based on word correlation. The paper scoring algorithm is also available as a separate program, TextMine. Form-driven annotation of minimotif attributes enables entry of new minimotifs into the MnM database. Several supporting features increase the efficiency of annotation. The layered architecture of MimoSA allows for extensibility by separating the functions of paper scoring, minimotif visualization, and database management. MimoSA is readily adaptable to other annotation efforts that manually curate literature into a MySQL database. Conclusions MimoSA is an extensible application that facilitates minimotif annotation and integrates with the Minimotif Miner database. We have built MimoSA as an application that integrates dynamic abstract scoring with a high performance relational model of minimotif syntax. MimoSA's TextMine, an efficient paper-scoring algorithm, can be used to dynamically rank papers with respect to context. PMID:20565705
Dietary intake and main sources of plant lignans in five European countries
Tetens, Inge; Turrini, Aida; Tapanainen, Heli; Christensen, Tue; Lampe, Johanna W.; Fagt, Sisse; Håkansson, Niclas; Lundquist, Annamari; Hallund, Jesper; Valsta, Liisa M.
2013-01-01
Background Dietary intakes of plant lignans have been hypothesized to be inversely associated with the risk of developing cardiovascular disease and cancer. Earlier studies were based on a Finnish lignan database (Fineli®) with two lignan precursors, secoisolariciresinol (SECO) and matairesinol (MAT). More recently, a Dutch database, including SECO and MAT and the newly recognized lignan precursors lariciresinol (LARI) and pinoresinol (PINO), was compiled. The objective was to re-estimate and re-evaluate plant lignan intakes and to identify the main sources of plant lignans in five European countries using the Finnish and Dutch lignan databases, respectively. Methods Forty-two food groups known to contribute to the total lignan intake were selected and attributed a value for SECO and MAT from the Finnish lignan database (Fineli®) or for SECO, MAT, LARI, and PINO from the Dutch database. Total intake of lignans was estimated from food consumption data for adult men and women (19–79 years) from Denmark, Finland, Italy, Sweden, United Kingdom, and the contribution of aggregated food groups calculated using the Dutch lignin database. Results Mean dietary lignan intakes estimated using the Dutch database ranged from 1 to 2 mg/day, which was approximately four-fold higher than the intakes estimated from the Fineli® database. When LARI and PINO were included in the estimation of the total lignan intakes, cereals, grain products, vegetables, fruit and berries were the most important dietary sources of lignans. Conclusion Total lignin intake was approximately four-fold higher in the Dutch lignin database, which includes the lignin precursors LARI and PINO, compared to estimates based on the Finnish database based only on SECO and MAT. The main sources of lignans according to the Dutch database in the five countries studied were cereals and grain products, vegetables, fruit, berries, and beverages. PMID:23766759
Integrating query of relational and textual data in clinical databases: a case study.
Fisk, John M; Mutalik, Pradeep; Levin, Forrest W; Erdos, Joseph; Taylor, Caroline; Nadkarni, Prakash
2003-01-01
The authors designed and implemented a clinical data mart composed of an integrated information retrieval (IR) and relational database management system (RDBMS). Using commodity software, which supports interactive, attribute-centric text and relational searches, the mart houses 2.8 million documents that span a five-year period and supports basic IR features such as Boolean searches, stemming, and proximity and fuzzy searching. Results are relevance-ranked using either "total documents per patient" or "report type weighting." Non-curated medical text has a significant degree of malformation with respect to spelling and punctuation, which creates difficulties for text indexing and searching. Presently, the IR facilities of RDBMS packages lack the features necessary to handle such malformed text adequately. A robust IR+RDBMS system can be developed, but it requires integrating RDBMSs with third-party IR software. RDBMS vendors need to make their IR offerings more accessible to non-programmers.
NASA Astrophysics Data System (ADS)
Xu, Mingzhu; Gao, Zhiqiang; Ning, Jicai
2014-10-01
To improve the access efficiency of geoscience data, efficient data model and storage solutions should be used. Geoscience data is usually classified by format or coordinate system in existing storage solutions. When data is large, it is not conducive to search the geographic features. In this study, a geographical information integration system of Shandong province, China was developed based on the technology of ArcGIS Engine, .NET, and SQL Server. It uses Geodatabase spatial data model and ArcSDE to organize and store spatial and attribute data and establishes geoscience database of Shangdong. Seven function modules were designed: map browse, database and subject management, layer control, map query, spatial analysis and map symbolization. The system's characteristics of can be browsed and managed by geoscience subjects make the system convenient for geographic researchers and decision-making departments to use the data.
Giardine, Belinda; Borg, Joseph; Higgs, Douglas R; Peterson, Kenneth R; Philipsen, Sjaak; Maglott, Donna; Singleton, Belinda K; Anstee, David J; Basak, A Nazli; Clark, Barnaby; Costa, Flavia C; Faustino, Paula; Fedosyuk, Halyna; Felice, Alex E; Francina, Alain; Galanello, Renzo; Gallivan, Monica V E; Georgitsi, Marianthi; Gibbons, Richard J; Giordano, Piero C; Harteveld, Cornelis L; Hoyer, James D; Jarvis, Martin; Joly, Philippe; Kanavakis, Emmanuel; Kollia, Panagoula; Menzel, Stephan; Miller, Webb; Moradkhani, Kamran; Old, John; Papachatzopoulou, Adamantia; Papadakis, Manoussos N; Papadopoulos, Petros; Pavlovic, Sonja; Perseu, Lucia; Radmilovic, Milena; Riemer, Cathy; Satta, Stefania; Schrijver, Iris; Stojiljkovic, Maja; Thein, Swee Lay; Traeger-Synodinos, Jan; Tully, Ray; Wada, Takahito; Waye, John S; Wiemann, Claudia; Zukic, Branka; Chui, David H K; Wajcman, Henri; Hardison, Ross C; Patrinos, George P
2011-03-20
We developed a series of interrelated locus-specific databases to store all published and unpublished genetic variation related to hemoglobinopathies and thalassemia and implemented microattribution to encourage submission of unpublished observations of genetic variation to these public repositories. A total of 1,941 unique genetic variants in 37 genes, encoding globins and other erythroid proteins, are currently documented in these databases, with reciprocal attribution of microcitations to data contributors. Our project provides the first example of implementing microattribution to incentivise submission of all known genetic variation in a defined system. It has demonstrably increased the reporting of human variants, leading to a comprehensive online resource for systematically describing human genetic variation in the globin genes and other genes contributing to hemoglobinopathies and thalassemias. The principles established here will serve as a model for other systems and for the analysis of other common and/or complex human genetic diseases.
NASA Astrophysics Data System (ADS)
Gasser, Deta; Viola, Giulio; Bingen, Bernard
2016-04-01
Since 2010, the Geological Survey of Norway has been implementing and continuously developing a digital workflow for geological bedrock mapping in Norway, from fieldwork to final product. Our workflow is based on the ESRI ArcGIS platform, and we use rugged Windows computers in the field. Three different hardware solutions have been tested over the past 5 years (2010-2015). (1) Panasonic Toughbook CE-19 (2.3 kg), (2) Panasonic Toughbook CF H2 Field (1.6 kg) and (3) Motion MC F5t tablet (1.5 kg). For collection of point observations in the field we mainly use the SIGMA Mobile application in ESRI ArcGIS developed by the British Geological Survey, which allows the mappers to store georeferenced comments, structural measurements, sample information, photographs, sketches, log information etc. in a Microsoft Access database. The application is freely downloadable from the BGS websites. For line- and polygon work we use our in-house database, which is currently under revision. Our line database consists of three feature classes: (1) bedrock boundaries, (2) bedrock lineaments, and (3) bedrock lines, with each feature class having up to 24 different attribute fields. Our polygon database consists of one feature class with 38 attribute fields enabling to store various information concerning lithology, stratigraphic order, age, metamorphic grade and tectonic subdivision. The polygon and line databases are coupled via topology in ESRI ArcGIS, which allows us to edit them simultaneously. This approach has been applied in two large-scale 1:50 000 bedrock mapping projects, one in the Kongsberg domain of the Sveconorwegian orogen, and the other in the greater Trondheim area (Orkanger) in the Caledonian belt. The mapping projects combined collection of high-resolution geophysical data, digital acquisition of field data, and collection of geochronological, geochemical and petrological data. During the Kongsberg project, some 25000 field observation points were collected by eight geologists. For the Orkanger project, some 2100 field observation points were collected by three geologists. Several advantages of the applied digital approach became clear during these projects: (1) The systematic collection of geological field data in a common format allows easy access and exchange of data among different geologists, (2) Easier access to background information such as geophysics and DEMS in the field, (3) Faster workflow from field data collection to final map product. Obvious disadvantages include: (1) Heavy(ish) and expensive hardware, (2) Battery life and other technical issues in the field, (3) Need for a central field observation point storage inhouse (large amounts of data!), and (4) Acceptance of- and training in a common workflow from all involved geologists.
Quality Attribute-Guided Evaluation of NoSQL Databases: An Experience Report
2014-10-18
detailed technical evaluations of NoSQL databases specifically, and big data systems in general, that have become apparent during our study... big data , software systems [Agarwal 2011]. Internet-born organizations such as Google and Amazon are at the cutting edge of this revolution...Chang 2008], along with those of numerous other big data innovators, have made a variety of open source and commercial data management technologies
The research and development of water resources management information system based on ArcGIS
NASA Astrophysics Data System (ADS)
Cui, Weiqun; Gao, Xiaoli; Li, Yuzhi; Cui, Zhencai
According to that there are large amount of data, complexity of data type and format in the water resources management, we built the water resources calculation model and established the water resources management information system based on the advanced ArcGIS and Visual Studio.NET development platform. The system can integrate the spatial data and attribute data organically, and manage them uniformly. It can analyze spatial data, inquire by map and data bidirectionally, provide various charts and report forms automatically, link multimedia information, manage database etc. . So it can provide spatial and static synthetical information services for study, management and decision of water resources, regional geology and eco-environment etc..
NASA Astrophysics Data System (ADS)
McEver, Jimmie; Davis, Paul K.; Bigelow, James H.
2000-06-01
We have developed and used families of multiresolution and multiple-perspective models (MRM and MRMPM), both in our substantive analytic work for the Department of Defense and to learn more about how such models can be designed and implemented. This paper is a brief case history of our experience with a particular family of models addressing the use of precision fires in interdicting and halting an invading army. Our models were implemented as closed-form analytic solutions, in spreadsheets, and in the more sophisticated AnalyticaTM environment. We also drew on an entity-level simulation for data. The paper reviews the importance of certain key attributes of development environments (visual modeling, interactive languages, friendly use of array mathematics, facilities for experimental design and configuration control, statistical analysis tools, graphical visualization tools, interactive post-processing, and relational database tools). These can go a long way towards facilitating MRMPM work, but many of these attributes are not yet widely available (or available at all) in commercial model-development tools--especially for use with personal computers. We conclude with some lessons learned from our experience.
Spontaneous belief attribution in younger siblings of children on the autism spectrum.
Gliga, Teodora; Senju, Atsushi; Pettinato, Michèle; Charman, Tony; Johnson, Mark H
2014-03-01
The recent development in the measurements of spontaneous mental state understanding, employing eye-movements instead of verbal responses, has opened new opportunities for understanding the developmental origin of "mind-reading" impairments frequently described in autism spectrum disorders (ASDs). Our main aim was to characterize the relationship between mental state understanding and the broader autism phenotype, early in childhood. An eye-tracker was used to capture anticipatory looking as a measure of false beliefs attribution in 3-year-old children with a family history of autism (at-risk participants, n = 47) and controls (control participants, n = 39). Unlike controls, the at-risk group, independent of their clinical outcome (ASD, broader autism phenotype or typically developing), performed at chance. Performance was not related to children's verbal or general IQ, nor was it explained by children "missing out" on crucial information, as shown by an analysis of visual scanning during the task. We conclude that difficulties with using mental state understanding for action prediction may be an endophenotype of autism spectrum disorders. PsycINFO Database Record (c) 2014 APA, all rights reserved.
Resilience in American Indian and Alaska Native Public Health: An Underexplored Framework.
Teufel-Shone, Nicolette I; Tippens, Julie A; McCrary, Hilary C; Ehiri, John E; Sanderson, Priscilla R
2018-02-01
To conduct a systematic literature review to assess the conceptualization, application, and measurement of resilience in American Indian and Alaska Native (AIAN) health promotion. We searched 9 literature databases to document how resilience is discussed, fostered, and evaluated in studies of AIAN health promotion in the United States. The article had to (1) be in English; (2) peer reviewed, published from January 1, 1980, to July 31, 2015; (3) identify the target population as predominantly AIANs in the United States; (4) describe a nonclinical intervention or original research that identified resilience as an outcome or resource; and (5) discuss resilience as related to cultural, social, and/or collective strengths. Sixty full texts were retrieved and assessed for inclusion by 3 reviewers. Data were extracted by 2 reviewers and verified for relevance to inclusion criteria by the third reviewer. Attributes of resilience that appeared repeatedly in the literature were identified. Findings were categorized across the lifespan (age group of participants), divided by attributes, and further defined by specific domains within each attribute. Nine articles (8 studies) met the criteria. Currently, resilience research in AIAN populations is limited to the identification of attributes and pilot interventions focused on individual resilience. Resilience models are not used to guide health promotion programming; collective resilience is not explored. Attributes of AIAN resilience should be considered in the development of health interventions. Attention to collective resilience is recommended to leverage existing assets in AIAN communities.
Datasets for U.S. mortality, U.S. populations, standard populations, county attributes, and expected survival. Plus SEER-linked databases (SEER-Medicare, SEER-Medicare Health Outcomes Survey [SEER-MHOS], SEER-Consumer Assessment of Healthcare Providers and Systems [SEER-CAHPS]).
Zhang, Liming; Yu, Dongsheng; Shi, Xuezheng; Xu, Shengxiang; Xing, Shihe; Zhao, Yongcong
2014-01-01
Soil organic carbon (SOC) models were often applied to regions with high heterogeneity, but limited spatially differentiated soil information and simulation unit resolution. This study, carried out in the Tai-Lake region of China, defined the uncertainty derived from application of the DeNitrification-DeComposition (DNDC) biogeochemical model in an area with heterogeneous soil properties and different simulation units. Three different resolution soil attribute databases, a polygonal capture of mapping units at 1∶50,000 (P5), a county-based database of 1∶50,000 (C5) and county-based database of 1∶14,000,000 (C14), were used as inputs for regional DNDC simulation. The P5 and C5 databases were combined with the 1∶50,000 digital soil map, which is the most detailed soil database for the Tai-Lake region. The C14 database was combined with 1∶14,000,000 digital soil map, which is a coarse database and is often used for modeling at a national or regional scale in China. The soil polygons of P5 database and county boundaries of C5 and C14 databases were used as basic simulation units. Results project that from 1982 to 2000, total SOC change in the top layer (0–30 cm) of the 2.3 M ha of paddy soil in the Tai-Lake region was +1.48 Tg C, −3.99 Tg C and −15.38 Tg C based on P5, C5 and C14 databases, respectively. With the total SOC change as modeled with P5 inputs as the baseline, which is the advantages of using detailed, polygon-based soil dataset, the relative deviation of C5 and C14 were 368% and 1126%, respectively. The comparison illustrates that DNDC simulation is strongly influenced by choice of fundamental geographic resolution as well as input soil attribute detail. The results also indicate that improving the framework of DNDC is essential in creating accurate models of the soil carbon cycle. PMID:24523922
Delayed Instantiation Bulk Operations for Management of Distributed, Object-Based Storage Systems
2009-08-01
source and destination object sets, while they have attribute pages to indicate that history . Fourth, we allow for operations to occur on any objects...client dialogue to the PostgreSQL database where server-side functions implement the service logic for the requests. The translation is done...to satisfy client requests, and performs delayed instantiation bulk operations. It is built around a PostgreSQL database with tables for storing
DOE Office of Scientific and Technical Information (OSTI.GOV)
J.M. Fenelon
2005-10-05
Almost 4,000 water-level measurements in 216 wells in the Yucca Flat area from 1951 to 2003 were quality assured and analyzed. An interpretative database was developed that describes water-level conditions for each water level measured in Yucca Flat. Multiple attributes were assigned to each water-level measurement in the database to describe the hydrologic conditions at the time of measurement. General quality, temporal variability, regional significance, and hydrologic conditions are attributed for each water-level measurement. The database also includes narratives that discuss the water-level history of each well. Water levels in 34 wells were analyzed for variability and for statistically significantmore » trends. An attempt was made to identify the cause of many of the water-level fluctuations or trends. Potential causes include equilibration following well construction or development, pumping in the monitoring well, withdrawals from a nearby supply well, recharge from precipitation, earthquakes, underground nuclear tests, land subsidence, barometric pressure, and Earth tides. Some of the naturally occurring fluctuations in water levels may result from variations in recharge. The magnitude of the overall water-level change for these fluctuations generally is less than 2 feet. Long-term steady-state hydrographs for most of the wells open to carbonate rock have a very similar pattern. Carbonate-rock wells without the characteristic pattern are directly west of the Yucca and Topgallant faults in the southwestern part of Yucca Flat. Long-term steady-state hydrographs from wells open to volcanic tuffs or the Eleana confining unit have a distinctly different pattern from the general water-level pattern of the carbonate-rock aquifers. Anthropogenic water-level fluctuations were caused primarily by water withdrawals and nuclear testing. Nuclear tests affected water levels in many wells. Trends in these wells are attributed to test-cavity infilling or the effects of depressurization following nuclear testing. The magnitude of the overall water-level change for wells with anthropogenic trends can be large, ranging from several feet to hundreds of feet. Vertical water-level differences at 27 sites in Yucca Flat with multiple open intervals were compared. Large vertical differences were noted in volcanic rocks and in boreholes where water levels were affected by nuclear tests. Small vertical differences were noted within the carbonate-rock and valley-fill aquifers. Vertical hydraulic gradients generally are downward in volcanic rocks and from pre-Tertiary clastic rocks toward volcanic- or carbonate-rock units.« less
Fenelon, Joseph M.
2005-01-01
Almost 4,000 water-level measurements in 216 wells in the Yucca Flat area from 1951 to 2003 were quality assured and analyzed. An interpretative database was developed that describes water-level conditions for each water level measured in Yucca Flat. Multiple attributes were assigned to each water-level measurement in the database to describe the hydrologic conditions at the time of measurement. General quality, temporal variability, regional significance, and hydrologic conditions are attributed for each water-level measurement. The database also includes narratives that discuss the water-level history of each well. Water levels in 34 wells were analyzed for variability and for statistically significant trends. An attempt was made to identify the cause of many of the water-level fluctuations or trends. Potential causes include equilibration following well construction or development, pumping in the monitoring well, withdrawals from a nearby supply well, recharge from precipitation, earthquakes, underground nuclear tests, land subsidence, barometric pressure, and Earth tides. Some of the naturally occurring fluctuations in water levels may result from variations in recharge. The magnitude of the overall water-level change for these fluctuations generally is less than 2 feet. Long-term steady-state hydrographs for most of the wells open to carbonate rock have a very similar pattern. Carbonate-rock wells without the characteristic pattern are directly west of the Yucca and Topgallant faults in the southwestern part of Yucca Flat. Long-term steady-state hydrographs from wells open to volcanic tuffs or the Eleana confining unit have a distinctly different pattern from the general water-level pattern of the carbonate-rock aquifers. Anthropogenic water-level fluctuations were caused primarily by water withdrawals and nuclear testing. Nuclear tests affected water levels in many wells. Trends in these wells are attributed to test-cavity infilling or the effects of depressurization following nuclear testing. The magnitude of the overall water-level change for wells with anthropogenic trends can be large, ranging from several feet to hundreds of feet. Vertical water-level differences at 27 sites in Yucca Flat with multiple open intervals were compared. Large vertical differences were noted in volcanic rocks and in boreholes where water levels were affected by nuclear tests. Small vertical differences were noted within the carbonate-rock and valley-fill aquifers. Vertical hydraulic gradients generally are downward in volcanic rocks and from pre-Tertiary clastic rocks toward volcanic- or carbonate-rock units.
A web-based, relational database for studying glaciers in the Italian Alps
NASA Astrophysics Data System (ADS)
Nigrelli, G.; Chiarle, M.; Nuzzi, A.; Perotti, L.; Torta, G.; Giardino, M.
2013-02-01
Glaciers are among the best terrestrial indicators of climate change and thus glacier inventories have attracted a growing, worldwide interest in recent years. In Italy, the first official glacier inventory was completed in 1925 and 774 glacial bodies were identified. As the amount of data continues to increase, and new techniques become available, there is a growing demand for computer tools that can efficiently manage the collected data. The Research Institute for Geo-hydrological Protection of the National Research Council, in cooperation with the Departments of Computer Science and Earth Sciences of the University of Turin, created a database that provides a modern tool for storing, processing and sharing glaciological data. The database was developed according to the need of storing heterogeneous information, which can be retrieved through a set of web search queries. The database's architecture is server-side, and was designed by means of an open source software. The website interface, simple and intuitive, was intended to meet the needs of a distributed public: through this interface, any type of glaciological data can be managed, specific queries can be performed, and the results can be exported in a standard format. The use of a relational database to store and organize a large variety of information about Italian glaciers collected over the last hundred years constitutes a significant step forward in ensuring the safety and accessibility of such data. Moreover, the same benefits also apply to the enhanced operability for handling information in the future, including new and emerging types of data formats, such as geographic and multimedia files. Future developments include the integration of cartographic data, such as base maps, satellite images and vector data. The relational database described in this paper will be the heart of a new geographic system that will merge data, data attributes and maps, leading to a complete description of Italian glacial environments.
NASA Astrophysics Data System (ADS)
Qiu, Xin; Cheng, Irene; Yang, Fuquan; Horb, Erin; Zhang, Leiming; Harner, Tom
2018-03-01
Two speciated and spatially resolved emissions databases for polycyclic aromatic compounds (PACs) in the Athabasca oil sands region (AOSR) were developed. The first database was derived from volatile organic compound (VOC) emissions data provided by the Cumulative Environmental Management Association (CEMA) and the second database was derived from additional data collected within the Joint Canada-Alberta Oil Sands Monitoring (JOSM) program. CALPUFF modelling results for atmospheric polycyclic aromatic hydrocarbons (PAHs), alkylated PAHs, and dibenzothiophenes (DBTs), obtained using each of the emissions databases, are presented and compared with measurements from a passive air monitoring network. The JOSM-derived emissions resulted in better model-measurement agreement in the total PAH concentrations and for most PAH species concentrations compared to results using CEMA-derived emissions. At local sites near oil sands mines, the percent error of the model compared to observations decreased from 30 % using the CEMA-derived emissions to 17 % using the JOSM-derived emissions. The improvement at local sites was likely attributed to the inclusion of updated tailings pond emissions estimated from JOSM activities. In either the CEMA-derived or JOSM-derived emissions scenario, the model underestimated PAH concentrations by a factor of 3 at remote locations. Potential reasons for the disagreement include forest fire emissions, re-emissions of previously deposited PAHs, and long-range transport not considered in the model. Alkylated PAH and DBT concentrations were also significantly underestimated. The CALPUFF model is expected to predict higher concentrations because of the limited chemistry and deposition modelling. Thus the model underestimation of PACs is likely due to gaps in the emissions database for these compounds and uncertainties in the methodology for estimating the emissions. Future work is required that focuses on improving the PAC emissions estimation and speciation methodologies and reducing the uncertainties in VOC emissions which are subsequently used in PAC emissions estimation.
Risk as an attribute in discrete choice experiments: a systematic review of the literature.
Harrison, Mark; Rigby, Dan; Vass, Caroline; Flynn, Terry; Louviere, Jordan; Payne, Katherine
2014-01-01
Discrete choice experiments (DCEs) are used to elicit preferences of current and future patients and healthcare professionals about how they value different aspects of healthcare. Risk is an integral part of most healthcare decisions. Despite the use of risk attributes in DCEs consistently being highlighted as an area for further research, current methods of incorporating risk attributes in DCEs have not been reviewed explicitly. This study aimed to systematically identify published healthcare DCEs that incorporated a risk attribute, summarise and appraise methods used to present and analyse risk attributes, and recommend best practice regarding including, analysing and transparently reporting the methodology supporting risk attributes in future DCEs. The Web of Science, MEDLINE, EMBASE, PsycINFO and Econlit databases were searched on 18 April 2013 for DCEs that included a risk attribute published since 1995, and on 23 April 2013 to identify studies assessing risk communication in the general (non-DCE) health literature. Healthcare-related DCEs with a risk attribute mentioned or suggested in the title/abstract were obtained and retained in the final review if a risk attribute meeting our definition was included. Extracted data were tabulated and critically appraised to summarise the quality of reporting, and the format, presentation and interpretation of the risk attribute were summarised. This review identified 117 healthcare DCEs that incorporated at least one risk attribute. Whilst there was some evidence of good practice incorporated into the presentation of risk attributes, little evidence was found that developing methods and recommendations from other disciplines about effective methods and validation of risk communication were systematically applied to DCEs. In general, the reviewed DCE studies did not thoroughly report the methodology supporting the explanation of risk in training materials, the impact of framing risk, or exploring the validity of risk communication. The primary limitation of this review was that the methods underlying presentation, format and analysis of risk attributes could only be appraised to the extent that they were reported. Improvements in reporting and transparency of risk presentation from conception to the analysis of DCEs are needed. To define best practice, further research is needed to test how the process of communicating risk affects the way in which people value risk attributes in DCEs.
Service user involvement in mental health care: an evolutionary concept analysis.
Millar, Samantha L; Chambers, Mary; Giles, Melanie
2016-04-01
The concept of service user involvement is an evolving concept in the mental health-care literature. This study sought to explore and analyse the concept of service user involvement as used in within the field of mental health care. An evolutionary concept analysis was conducted using a literature-based sample extracted from an electronic database search. One hundred and thirty-four papers met the inclusion criteria and were analysed to discover key attributes, antecedents and consequences of service user involvement and to produce a definition of the concept. Five key attributes of service user involvement within the context of mental health care were identified: a person-centred approach, informed decision making, advocacy, obtaining service user views and feedback and working in partnership. Clarity of the attributes and definition of the concept of service user involvement aims to promote understanding of the concept among key stakeholders including mental health professionals, service users and community and voluntary organizations. The findings of the research have utility in the areas of theory and policy development, research on service user involvement in mental health care and service user involvement in mental health practice. Directions for further research regarding the concept are identified. © 2015 John Wiley & Sons Ltd.
NASA Astrophysics Data System (ADS)
Bhanumurthy, V.; Venugopala Rao, K.; Srinivasa Rao, S.; Ram Mohan Rao, K.; Chandra, P. Satya; Vidhyasagar, J.; Diwakar, P. G.; Dadhwal, V. K.
2014-11-01
Geographical Information Science (GIS) is now graduated from traditional desktop system to Internet system. Internet GIS is emerging as one of the most promising technologies for addressing Emergency Management. Web services with different privileges are playing an important role in dissemination of the emergency services to the decision makers. Spatial database is one of the most important components in the successful implementation of Emergency Management. It contains spatial data in the form of raster, vector, linked with non-spatial information. Comprehensive data is required to handle emergency situation in different phases. These database elements comprise core data, hazard specific data, corresponding attribute data, and live data coming from the remote locations. Core data sets are minimum required data including base, thematic, infrastructure layers to handle disasters. Disaster specific information is required to handle a particular disaster situation like flood, cyclone, forest fire, earth quake, land slide, drought. In addition to this Emergency Management require many types of data with spatial and temporal attributes that should be made available to the key players in the right format at right time. The vector database needs to be complemented with required resolution satellite imagery for visualisation and analysis in disaster management. Therefore, the database is interconnected and comprehensive to meet the requirement of an Emergency Management. This kind of integrated, comprehensive and structured database with appropriate information is required to obtain right information at right time for the right people. However, building spatial database for Emergency Management is a challenging task because of the key issues such as availability of data, sharing policies, compatible geospatial standards, data interoperability etc. Therefore, to facilitate using, sharing, and integrating the spatial data, there is a need to define standards to build emergency database systems. These include aspects such as i) data integration procedures namely standard coding scheme, schema, meta data format, spatial format ii) database organisation mechanism covering data management, catalogues, data models iii) database dissemination through a suitable environment, as a standard service for effective service dissemination. National Database for Emergency Management (NDEM) is such a comprehensive database for addressing disasters in India at the national level. This paper explains standards for integrating, organising the multi-scale and multi-source data with effective emergency response using customized user interfaces for NDEM. It presents standard procedure for building comprehensive emergency information systems for enabling emergency specific functions through geospatial technologies.
Prakash, Peralam Yegneswaran; Irinyi, Laszlo; Halliday, Catriona; Chen, Sharon; Robert, Vincent
2017-01-01
ABSTRACT The increase in public online databases dedicated to fungal identification is noteworthy. This can be attributed to improved access to molecular approaches to characterize fungi, as well as to delineate species within specific fungal groups in the last 2 decades, leading to an ever-increasing complexity of taxonomic assortments and nomenclatural reassignments. Thus, well-curated fungal databases with substantial accurate sequence data play a pivotal role for further research and diagnostics in the field of mycology. This minireview aims to provide an overview of currently available online databases for the taxonomy and identification of human and animal-pathogenic fungi and calls for the establishment of a cloud-based dynamic data network platform. PMID:28179406
Roudier, B; Davit, B; Schütz, H; Cardot, J-M
2015-01-01
The in vitro-in vivo correlation (IVIVC) (Food and Drug Administration 1997) aims to predict performances in vivo of a pharmaceutical formulation based on its in vitro characteristics. It is a complex process that (i) incorporates in a gradual and incremental way a large amount of information and (ii) requires information from different properties (formulation, analytical, clinical) and associated dedicated treatments (statistics, modeling, simulation). These results in many studies that are initiated and integrated into the specifications (quality target product profile, QTPP). This latter defines the appropriate experimental designs (quality by design, QbD) (Food and Drug Administration 2011, 2012) whose main objectives are determination (i) of key factors of development and manufacturing (critical process parameters, CPPs) and (ii) of critical points of physicochemical nature relating to active ingredients (API) and critical quality attribute (CQA) which may have implications in terms of efficiency, safety, and inoffensiveness for the patient, due to their non-inclusion. These processes generate a very large amount of data that is necessary to structure. In this context, the storage of information in a database (DB) and the management of this database (database management system, DBMS) become an important issue for the management of projects and IVIVC and more generally for development of new pharmaceutical forms. This article describes the implementation of a prototype object-oriented database (OODB) considered as a tool, which is helpful for decision taking, responding in a structured and consistent way to the issues of project management of IVIVC (including bioequivalence and bioavailability) (Food and Drug Administration 2003) necessary for the implementation of QTPP.
Assessing the SunGuide and STEWARD databases.
DOT National Transportation Integrated Search
2017-02-01
This project evaluated the feasibility of using the existing software and data bases as platforms : for analyzing the attributes of electric vehicles within present and future transportation : infrastructure projects and models. The Florida based Sun...
Turnbull, Bev; Royal, Bernadette; Purnell, Margaret
2011-01-01
As learning paradigms shift to student-centred active learning, development of effective skills in locating and retrieving information using electronic sources is integral to promoting lifelong learning. Recency of information that is evidence based is a critical factor in a dynamic field such as health. A changing demographic is evident among nursing students with greater numbers of mature age students who may not possess the computer skills often assumed with school leavers, and whose study preference is mostly by external mode. Development of interdisciplinary partnerships between faculties and librarians can provide the attributes and innovation of new and improved ways to better support student learning, whether or not students attend on campus. The Health Online Tutorial, an online database searching tool developed through a collaborative, interdisciplinary partnership at Charles Darwin University is one such example.
The role of sensory perception in the development and targeting of tobacco products.
Carpenter, Carrie M; Wayne, Geoffrey Ferris; Connolly, Gregory N
2007-01-01
To examine tobacco industry research on smoking-related sensory effects, including differences in sensory perception across smoker groups, and to determine whether this research informed targeted product development and impacted the development of commercial tobacco products. We searched previously secret internal tobacco industry documents available online through document databases housed at Tobacco Documents Online, the British American Tobacco Document Archive and the Legacy Tobacco Documents Library. We identified relevant documents using a snowball sampling method to first search the databases using an initial set of key words and to then establish further search terms. Sensory research is a priority within the tobacco industry directly impacting commercial markets both in the United States and internationally. Sensory factors contribute to smoker satisfaction and product acceptance, and play an important role in controlling puffing behavior. Cigarette manufacturers have capitalized on distinct sensory preferences across gender, age and ethnic groups by tailoring products for specific populations. Regulation of tobacco products is needed to address product changes that are used to reinforce or contribute to tobacco dependence; for instance, the incorporation of additives that target attributes such as smoothness, harshness and aftertaste. Greater understanding of the role of sensory effects on smoking behavior may also help to inform the development of tobacco treatment options that support long-term tobacco abstinence.
TMDB: a literature-curated database for small molecular compounds found from tea.
Yue, Yi; Chu, Gang-Xiu; Liu, Xue-Shi; Tang, Xing; Wang, Wei; Liu, Guang-Jin; Yang, Tao; Ling, Tie-Jun; Wang, Xiao-Gang; Zhang, Zheng-Zhu; Xia, Tao; Wan, Xiao-Chun; Bao, Guan-Hu
2014-09-16
Tea is one of the most consumed beverages worldwide. The healthy effects of tea are attributed to a wealthy of different chemical components from tea. Thousands of studies on the chemical constituents of tea had been reported. However, data from these individual reports have not been collected into a single database. The lack of a curated database of related information limits research in this field, and thus a cohesive database system should necessarily be constructed for data deposit and further application. The Tea Metabolome database (TMDB), a manually curated and web-accessible database, was developed to provide detailed, searchable descriptions of small molecular compounds found in Camellia spp. esp. in the plant Camellia sinensis and compounds in its manufactured products (different kinds of tea infusion). TMDB is currently the most complete and comprehensive curated collection of tea compounds data in the world. It contains records for more than 1393 constituents found in tea with information gathered from 364 published books, journal articles, and electronic databases. It also contains experimental 1H NMR and 13C NMR data collected from the purified reference compounds or collected from other database resources such as HMDB. TMDB interface allows users to retrieve tea compounds entries by keyword search using compound name, formula, occurrence, and CAS register number. Each entry in the TMDB contains an average of 24 separate data fields including its original plant species, compound structure, formula, molecular weight, name, CAS registry number, compound types, compound uses including healthy benefits, reference literatures, NMR, MS data, and the corresponding ID from databases such as HMDB and Pubmed. Users can also contribute novel regulatory entries by using a web-based submission page. The TMDB database is freely accessible from the URL of http://pcsb.ahau.edu.cn:8080/TCDB/index.jsp. The TMDB is designed to address the broad needs of tea biochemists, natural products chemists, nutritionists, and members of tea related research community. The TMDB database provides a solid platform for collection, standardization, and searching of compounds information found in tea. As such this database will be a comprehensive repository for tea biochemistry and tea health research community.
NASA Astrophysics Data System (ADS)
Liu, G.; Wu, C.; Li, X.; Song, P.
2013-12-01
The 3D urban geological information system has been a major part of the national urban geological survey project of China Geological Survey in recent years. Large amount of multi-source and multi-subject data are to be stored in the urban geological databases. There are various models and vocabularies drafted and applied by industrial companies in urban geological data. The issues such as duplicate and ambiguous definition of terms and different coding structure increase the difficulty of information sharing and data integration. To solve this problem, we proposed a national standard-driven information classification and coding method to effectively store and integrate urban geological data, and we applied the data dictionary technology to achieve structural and standard data storage. The overall purpose of this work is to set up a common data platform to provide information sharing service. Research progresses are as follows: (1) A unified classification and coding method for multi-source data based on national standards. Underlying national standards include GB 9649-88 for geology and GB/T 13923-2006 for geography. Current industrial models are compared with national standards to build a mapping table. The attributes of various urban geological data entity models are reduced to several categories according to their application phases and domains. Then a logical data model is set up as a standard format to design data file structures for a relational database. (2) A multi-level data dictionary for data standardization constraint. Three levels of data dictionary are designed: model data dictionary is used to manage system database files and enhance maintenance of the whole database system; attribute dictionary organizes fields used in database tables; term and code dictionary is applied to provide a standard for urban information system by adopting appropriate classification and coding methods; comprehensive data dictionary manages system operation and security. (3) An extension to system data management function based on data dictionary. Data item constraint input function is making use of the standard term and code dictionary to get standard input result. Attribute dictionary organizes all the fields of an urban geological information database to ensure the consistency of term use for fields. Model dictionary is used to generate a database operation interface automatically with standard semantic content via term and code dictionary. The above method and technology have been applied to the construction of Fuzhou Urban Geological Information System, South-East China with satisfactory results.
[Construction of chemical information database based on optical structure recognition technique].
Lv, C Y; Li, M N; Zhang, L R; Liu, Z M
2018-04-18
To create a protocol that could be used to construct chemical information database from scientific literature quickly and automatically. Scientific literature, patents and technical reports from different chemical disciplines were collected and stored in PDF format as fundamental datasets. Chemical structures were transformed from published documents and images to machine-readable data by using the name conversion technology and optical structure recognition tool CLiDE. In the process of molecular structure information extraction, Markush structures were enumerated into well-defined monomer molecules by means of QueryTools in molecule editor ChemDraw. Document management software EndNote X8 was applied to acquire bibliographical references involving title, author, journal and year of publication. Text mining toolkit ChemDataExtractor was adopted to retrieve information that could be used to populate structured chemical database from figures, tables, and textual paragraphs. After this step, detailed manual revision and annotation were conducted in order to ensure the accuracy and completeness of the data. In addition to the literature data, computing simulation platform Pipeline Pilot 7.5 was utilized to calculate the physical and chemical properties and predict molecular attributes. Furthermore, open database ChEMBL was linked to fetch known bioactivities, such as indications and targets. After information extraction and data expansion, five separate metadata files were generated, including molecular structure data file, molecular information, bibliographical references, predictable attributes and known bioactivities. Canonical simplified molecular input line entry specification as primary key, metadata files were associated through common key nodes including molecular number and PDF number to construct an integrated chemical information database. A reasonable construction protocol of chemical information database was created successfully. A total of 174 research articles and 25 reviews published in Marine Drugs from January 2015 to June 2016 collected as essential data source, and an elementary marine natural product database named PKU-MNPD was built in accordance with this protocol, which contained 3 262 molecules and 19 821 records. This data aggregation protocol is of great help for the chemical information database construction in accuracy, comprehensiveness and efficiency based on original documents. The structured chemical information database can facilitate the access to medical intelligence and accelerate the transformation of scientific research achievements.
Methods to achieve accurate projection of regional and global raster databases
Usery, E. Lynn; Seong, Jeong Chang; Steinwand, Dan
2002-01-01
Modeling regional and global activities of climatic and human-induced change requires accurate geographic data from which we can develop mathematical and statistical tabulations of attributes and properties of the environment. Many of these models depend on data formatted as raster cells or matrices of pixel values. Recently, it has been demonstrated that regional and global raster datasets are subject to significant error from mathematical projection and that these errors are of such magnitude that model results may be jeopardized (Steinwand, et al., 1995; Yang, et al., 1996; Usery and Seong, 2001; Seong and Usery, 2001). There is a need to develop methods of projection that maintain the accuracy of these datasets to support regional and global analyses and modeling
An interactive system for computer-aided diagnosis of breast masses.
Wang, Xingwei; Li, Lihua; Liu, Wei; Xu, Weidong; Lederman, Dror; Zheng, Bin
2012-10-01
Although mammography is the only clinically accepted imaging modality for screening the general population to detect breast cancer, interpreting mammograms is difficult with lower sensitivity and specificity. To provide radiologists "a visual aid" in interpreting mammograms, we developed and tested an interactive system for computer-aided detection and diagnosis (CAD) of mass-like cancers. Using this system, an observer can view CAD-cued mass regions depicted on one image and then query any suspicious regions (either cued or not cued by CAD). CAD scheme automatically segments the suspicious region or accepts manually defined region and computes a set of image features. Using content-based image retrieval (CBIR) algorithm, CAD searches for a set of reference images depicting "abnormalities" similar to the queried region. Based on image retrieval results and a decision algorithm, a classification score is assigned to the queried region. In this study, a reference database with 1,800 malignant mass regions and 1,800 benign and CAD-generated false-positive regions was used. A modified CBIR algorithm with a new function of stretching the attributes in the multi-dimensional space and decision scheme was optimized using a genetic algorithm. Using a leave-one-out testing method to classify suspicious mass regions, we compared the classification performance using two CBIR algorithms with either equally weighted or optimally stretched attributes. Using the modified CBIR algorithm, the area under receiver operating characteristic curve was significantly increased from 0.865 ± 0.006 to 0.897 ± 0.005 (p < 0.001). This study demonstrated the feasibility of developing an interactive CAD system with a large reference database and achieving improved performance.
Rodrigues, David; Prada, Marília; Gaspar, Rui; Garrido, Margarida V; Lopes, Diniz
2018-02-01
The use of emoticons and emoji is increasingly popular across a variety of new platforms of online communication. They have also become popular as stimulus materials in scientific research. However, the assumption that emoji/emoticon users' interpretations always correspond to the developers'/researchers' intended meanings might be misleading. This article presents subjective norms of emoji and emoticons provided by everyday users. The Lisbon Emoji and Emoticon Database (LEED) comprises 238 stimuli: 85 emoticons and 153 emoji (collected from iOS, Android, Facebook, and Emojipedia). The sample included 505 Portuguese participants recruited online. Each participant evaluated a random subset of 20 stimuli for seven dimensions: aesthetic appeal, familiarity, visual complexity, concreteness, valence, arousal, and meaningfulness. Participants were additionally asked to attribute a meaning to each stimulus. The norms obtained include quantitative descriptive results (means, standard deviations, and confidence intervals) and a meaning analysis for each stimulus. We also examined the correlations between the dimensions and tested for differences between emoticons and emoji, as well as between the two major operating systems-Android and iOS. The LEED constitutes a readily available normative database (available at www.osf.io/nua4x ) with potential applications to different research domains.
Biological and ecological traits of marine species
Claus, Simon; Dekeyzer, Stefanie; Vandepitte, Leen; Tuama, Éamonn Ó; Lear, Dan; Tyler-Walters, Harvey
2015-01-01
This paper reviews the utility and availability of biological and ecological traits for marine species so as to prioritise the development of a world database on marine species traits. In addition, the ‘status’ of species for conservation, that is, whether they are introduced or invasive, of fishery or aquaculture interest, harmful, or used as an ecological indicator, were reviewed because these attributes are of particular interest to society. Whereas traits are an enduring characteristic of a species and/or population, a species status may vary geographically and over time. Criteria for selecting traits were that they could be applied to most taxa, were easily available, and their inclusion would result in new research and/or management applications. Numerical traits were favoured over categorical. Habitat was excluded as it can be derived from a selection of these traits. Ten traits were prioritized for inclusion in the most comprehensive open access database on marine species (World Register of Marine Species), namely taxonomic classification, environment, geography, depth, substratum, mobility, skeleton, diet, body size and reproduction. These traits and statuses are being added to the database and new use cases may further subdivide and expand upon them. PMID:26312188
Vázquez, José Juan; Panadero, Sonia; Zúñiga, Claudia
2017-01-01
The study analyzes the differences in causal attributions of homelessness and attributions of responsibility among the members of 3 groups: homeless group, consisting of a representative sample of homeless people in Madrid, Spain (n = 188); domiciled service-users group, consisting of people at risk of homelessness (n = 164); and domiciled nonservice-users group, consisting of people at no imminent risk of homelessness (n = 180). The domiciled service-users group and domiciled nonservice-users group were matched to the homeless group or sex, age, and nationality. The article also analyzes homeless people's causal attributions as regards their own situation. The results show that compared with the domiciled nonservice-users group, a higher percentage of members of the homeless group and domiciled service-users group attributed homelessness to individualistic causes and they blamed homeless people for their situation to a greater extent. The results also show that there was no "actor-observer bias" in causal attributions for homelessness in Madrid. (PsycINFO Database Record (c) 2017 APA, all rights reserved).
Methods for assessing the quality of data in public health information systems: a critical review.
Chen, Hong; Yu, Ping; Hailey, David; Wang, Ning
2014-01-01
The quality of data in public health information systems can be ensured by effective data quality assessment. In order to conduct effective data quality assessment, measurable data attributes have to be precisely defined. Then reliable and valid measurement methods for data attributes have to be used to measure each attribute. We conducted a systematic review of data quality assessment methods for public health using major databases and well-known institutional websites. 35 studies were eligible for inclusion in the study. A total of 49 attributes of data quality were identified from the literature. Completeness, accuracy and timeliness were the three most frequently assessed attributes of data quality. Most studies directly examined data values. This is complemented by exploring either data users' perception or documentation quality. However, there are limitations of current data quality assessment methods: a lack of consensus on attributes measured; inconsistent definition of the data quality attributes; a lack of mixed methods for assessing data quality; and inadequate attention to reliability and validity. Removal of these limitations is an opportunity for further improvement.
Attribute amnesia reflects a lack of memory consolidation for attended information.
Chen, Hui; Wyble, Brad
2016-02-01
A recently reported phenomenon, termed attribute amnesia, challenged the commonly held belief that attention plays the determining role in controlling how information is remembered, by showing that participants fail to remember a specific attended attribute (e.g., the target-defining color), even when they had just used that attribute to perform a task (Chen & Wyble, 2015a). The main purpose of the present study sought to better understand the mechanism underlying this phenomenon. The results revealed that attribute amnesia was nearly eliminated once participants were forced to store and hold attended information for a brief time, suggesting that this amnesia effect most likely reflects a lack of memory consolidation for an attended attribute that had been processed to some certain level. In addition, we demonstrated that the effect is not particular to the use of location report or the repetition of targets. One additional finding is that amnesia was markedly absent for location memory, indicating an important difference between memories for locations and attributes such as color or identity. (PsycINFO Database Record (c) 2016 APA, all rights reserved).
Design and development of linked data from the National Map
Usery, E. Lynn; Varanka, Dalia E.
2012-01-01
The development of linked data on the World-Wide Web provides the opportunity for the U.S. Geological Survey (USGS) to supply its extensive volumes of geospatial data, information, and knowledge in a machine interpretable form and reach users and applications that heretofore have been unavailable. To pilot a process to take advantage of this opportunity, the USGS is developing an ontology for The National Map and converting selected data from nine research test areas to a Semantic Web format to support machine processing and linked data access. In a case study, the USGS has developed initial methods for legacy vector and raster formatted geometry, attributes, and spatial relationships to be accessed in a linked data environment maintaining the capability to generate graphic or image output from semantic queries. The description of an initial USGS approach to developing ontology, linked data, and initial query capability from The National Map databases is presented.
Toward the establishment of design guidelines for effective 3D perspective interfaces
NASA Astrophysics Data System (ADS)
Fitzhugh, Elisabeth; Dixon, Sharon; Aleva, Denise; Smith, Eric; Ghrayeb, Joseph; Douglas, Lisa
2009-05-01
The propagation of information operation technologies, with correspondingly vast amounts of complex network information to be conveyed, significantly impacts operator workload. Information management research is rife with efforts to develop schemes to aid operators to identify, review, organize, and retrieve the wealth of available data. Data may take on such distinct forms as intelligence libraries, logistics databases, operational environment models, or network topologies. Increased use of taxonomies and semantic technologies opens opportunities to employ network visualization as a display mechanism for diverse information aggregations. The broad applicability of network visualizations is still being tested, but in current usage, the complexity of densely populated abstract networks suggests the potential utility of 3D. Employment of 2.5D in network visualization, using classic perceptual cues, creates a 3D experience within a 2D medium. It is anticipated that use of 3D perspective (2.5D) will enhance user ability to visually inspect large, complex, multidimensional networks. Current research for 2.5D visualizations demonstrates that display attributes, including color, shape, size, lighting, atmospheric effects, and shadows, significantly impact operator experience. However, guidelines for utilization of attributes in display design are limited. This paper discusses pilot experimentation intended to identify potential problem areas arising from these cues and determine how best to optimize perceptual cue settings. Development of optimized design guidelines will ensure that future experiments, comparing network displays with other visualizations, are not confounded or impeded by suboptimal attribute characterization. Current experimentation is anticipated to support development of cost-effective, visually effective methods to implement 3D in military applications.
Asadi, S S; Vuppala, Padmaja; Reddy, M Anji
2005-01-01
A preliminary survey of area under Zone-III of MCH was undertaken to assess the ground water quality, demonstrate its spatial distribution and correlate with the land use patterns using advance techniques of remote sensing and geographical information system (GIS). Twenty-seven ground water samples were collected and their chemical analysis was done to form the attribute database. Water quality index was calculated from the measured parameters, based on which the study area was classified into five groups with respect to suitability of water for drinking purpose. Thematic maps viz., base map, road network, drainage and land use/land cover were prepared from IRS ID PAN + LISS III merged satellite imagery forming the spatial database. Attribute database was integrated with spatial sampling locations map in Arc/Info and maps showing spatial distribution of water quality parameters were prepared in Arc View. Results indicated that high concentrations of total dissolved solids (TDS), nitrates, fluorides and total hardness were observed in few industrial and densely populated areas indicating deteriorated water quality while the other areas exhibited moderate to good water quality.
Building Inventory Database on the Urban Scale Using GIS for Earthquake Risk Assessment
NASA Astrophysics Data System (ADS)
Kaplan, O.; Avdan, U.; Guney, Y.; Helvaci, C.
2016-12-01
The majority of the existing buildings are not safe against earthquakes in most of the developing countries. Before a devastating earthquake, existing buildings need to be assessed and the vulnerable ones must be determined. Determining the seismic performance of existing buildings which is usually made with collecting the attributes of existing buildings, making the analysis and the necessary queries, and producing the result maps is very hard and complicated procedure that can be simplified with Geographic Information System (GIS). The aim of this study is to produce a building inventory database using GIS for assessing the earthquake risk of existing buildings. In this paper, a building inventory database for 310 buildings, located in Eskisehir, Turkey, was produced in order to assess the earthquake risk of the buildings. The results from this study show that 26% of the buildings have high earthquake risk, 33% of the buildings have medium earthquake risk and the 41% of the buildings have low earthquake risk. The produced building inventory database can be very useful especially for governments in dealing with the problem of determining seismically vulnerable buildings in the large existing building stocks. With the help of this kind of methods, determination of the buildings, which may collapse and cause life and property loss during a possible future earthquake, will be very quick, cheap and reliable.
RoadPlex: A Mobile VGI Game to Collect and Validate Data for POIs
NASA Astrophysics Data System (ADS)
Kashian, A.; Rajabifard, A.; Richter, K. F.
2014-11-01
By increasing the popularity of smart phones equipped with GPS sensors, more volunteers are expected to join VGI (Volunteered Geographic Information) activities and therefore more positional data will be collected in shorter time. Current statistics from open databases such OpenStreetMap reveals that although there have been exponential growth in the number of contributed POIs (Points of Interest), the lack of detailed attribute information is immediately visible. The process of adding attribute information to VGI databases is usually considered as a boring task and it is believed that contributors do not experience a similar level of satisfaction when they add such detailed information compared to tasks like adding new roads or copying building boundaries from satellite imageries. In other crowdsourcing projects, different approaches are taken for engaging contributors in problem solving by embedding the tasks inside a game. In the literature, this concept is known as "gamification" or "games with purpose" which encapsulate the idea of entertaining contributors while they are completing a particular defined task. Same concept is used to design a mobile application called "RoadPlex" which aims to collect general or specific attribute information for POIs The increased number of contributions in the past few months confirms that the design characteristics and the methodology of the game are appealing to players. Such growth enables us to evaluate the quality of the generated data through mining the database of answered questions. This paper reflects the some contribution results and emphasises the importance of using gamification concept in the domain of VGI.
Nolan, Daniel J.; Ginsberg, Michael; Israely, Edo; Palikuqi, Brisa; Poulos, Michael G.; James, Daylon; Ding, Bi-Sen; Schachterle, William; Liu, Ying; Rosenwaks, Zev; Butler, Jason M.; Xiang, Jenny; Rafii, Arash; Shido, Koji; Rabbany, Sina Y.; Elemento, Olivier; Rafii, Shahin
2013-01-01
SUMMARY Microvascular endothelial cells (ECs) within different tissues are endowed with distinct but as yet unrecognized structural, phenotypic, and functional attributes. We devised EC purification, cultivation, profiling, and transplantation models that establish tissue-specific molecular libraries of ECs devoid of lymphatic ECs or parenchymal cells. These libraries identify attributes that confer ECs with their organotypic features. We show that clusters of transcription factors, angiocrine growth factors, adhesion molecules, and chemokines are expressed in unique combinations by ECs of each organ. Furthermore, ECs respond distinctly in tissue regeneration models, hepatectomy, and myeloablation. To test the data set, we developed a transplantation model that employs generic ECs differentiated from embryonic stem cells. Transplanted generic ECs engraft into regenerating tissues and acquire features of organotypic ECs. Collectively, we demonstrate the utility of informational databases of ECs toward uncovering the extravascular and intrinsic signals that define EC heterogeneity. These factors could be exploited therapeutically to engineer tissue-specific ECs for regeneration. PMID:23871589
Nursing professionalism: An evolutionary concept analysis
Ghadirian, Fataneh; Salsali, Mahvash; Cheraghi, Mohammad Ali
2014-01-01
Background: Professionalism is an important feature of the professional jobs. Dynamic nature and the various interpretations of this term lead to multiple definitions of this concept. The aim of this paper is to identify the core attributes of the nursing professionalism. Materials and Methods: We followed Rodgers’ evolutionary method of concept analysis. Texts published in scientific databases about nursing professionalism between 1980 and 2011 were assessed. After applying the selection criteria, the final sample consisting of 4 books and 213 articles was selected, examined, and analyzed in depth. Two experts checked the process of analysis and monitored and reviewed them. Results: The analysis showed that nursing professionalism is determined by three attributes of cognitive, attitudinal, and psychomotor. In addition, the most important antecedents concepts were demographic, experiential, educational, environmental, and attitudinal factors. Conclusion: Nursing professionalism is an inevitable, complex, varied, and dynamic process. In this study, the importance, scope, and concept of professionalism in nursing, the concept of a beginning for further research and development, and expanding the nursing knowledge are explained and clarified. PMID:24554953
Prakash, Peralam Yegneswaran; Irinyi, Laszlo; Halliday, Catriona; Chen, Sharon; Robert, Vincent; Meyer, Wieland
2017-04-01
The increase in public online databases dedicated to fungal identification is noteworthy. This can be attributed to improved access to molecular approaches to characterize fungi, as well as to delineate species within specific fungal groups in the last 2 decades, leading to an ever-increasing complexity of taxonomic assortments and nomenclatural reassignments. Thus, well-curated fungal databases with substantial accurate sequence data play a pivotal role for further research and diagnostics in the field of mycology. This minireview aims to provide an overview of currently available online databases for the taxonomy and identification of human and animal-pathogenic fungi and calls for the establishment of a cloud-based dynamic data network platform. Copyright © 2017 American Society for Microbiology.
Joy and happiness: a simultaneous and evolutionary concept analysis.
Cottrell, Laura
2016-07-01
To report a simultaneous and evolutionary analysis of the concepts of joy and long-term happiness. Joy and happiness are underrepresented in the nursing literature, though negative concepts are well represented. When mentioned in the literature, neither joy nor happiness is adequately defined, explained, or clearly understood. To promote further investigation of these concepts in nursing and to explore their relationship with health and healing, conceptual clarity is an essential first step. Concept analysis. The following databases were searched, without time restrictions, for articles in English: Academic Search Complete, Anthropology Plus; ATLA Religious Database with ATLASerials; Cumulative Index of Nursing and Allied Health Literature (CINAHL); Education Research Complete; Humanities International Complete; Psych EXTRA; and SocINDEX with Full Text. The final sample size consists of 61 articles and one book, published between 1978-2014. An adapted combination of Rodgers' Evolutionary Model and Haase et al.'s Simultaneous Concept Analysis (SCA) method. Though both are positive concepts, joy and happiness have significant differences. Attributes of joy describe a spontaneous, sudden and transient concept associated with connection, awareness, and freedom. Attributes of happiness describe a pursued, long-lasting, stable mental state associated with virtue and self-control. Further exploration of joy and happiness is necessary to ascertain their relationship with health and their value to nursing practice and theory development. Nurses are encouraged to consider the value of positive concepts to all areas of nursing. © 2016 John Wiley & Sons Ltd.
Data mining of text as a tool in authorship attribution
NASA Astrophysics Data System (ADS)
Visa, Ari J. E.; Toivonen, Jarmo; Autio, Sami; Maekinen, Jarno; Back, Barbro; Vanharanta, Hannu
2001-03-01
It is common that text documents are characterized and classified by keywords that the authors use to give them. Visa et al. have developed a new methodology based on prototype matching. The prototype is an interesting document or a part of an extracted, interesting text. This prototype is matched with the document database of the monitored document flow. The new methodology is capable of extracting the meaning of the document in a certain degree. Our claim is that the new methodology is also capable of authenticating the authorship. To verify this claim two tests were designed. The test hypothesis was that the words and the word order in the sentences could authenticate the author. In the first test three authors were selected. The selected authors were William Shakespeare, Edgar Allan Poe, and George Bernard Shaw. Three texts from each author were examined. Every text was one by one used as a prototype. The two nearest matches with the prototype were noted. The second test uses the Reuters-21578 financial news database. A group of 25 short financial news reports from five different authors are examined. Our new methodology and the interesting results from the two tests are reported in this paper. In the first test, for Shakespeare and for Poe all cases were successful. For Shaw one text was confused with Poe. In the second test the Reuters-21578 financial news were identified by the author relatively well. The resolution is that our text mining methodology seems to be capable of authorship attribution.
The Chinchilla Research Resource Database: resource for an otolaryngology disease model
Shimoyama, Mary; Smith, Jennifer R.; De Pons, Jeff; Tutaj, Marek; Khampang, Pawjai; Hong, Wenzhou; Erbe, Christy B.; Ehrlich, Garth D.; Bakaletz, Lauren O.; Kerschner, Joseph E.
2016-01-01
The long-tailed chinchilla (Chinchilla lanigera) is an established animal model for diseases of the inner and middle ear, among others. In particular, chinchilla is commonly used to study diseases involving viral and bacterial pathogens and polymicrobial infections of the upper respiratory tract and the ear, such as otitis media. The value of the chinchilla as a model for human diseases prompted the sequencing of its genome in 2012 and the more recent development of the Chinchilla Research Resource Database (http://crrd.mcw.edu) to provide investigators with easy access to relevant datasets and software tools to enhance their research. The Chinchilla Research Resource Database contains a complete catalog of genes for chinchilla and, for comparative purposes, human. Chinchilla genes can be viewed in the context of their genomic scaffold positions using the JBrowse genome browser. In contrast to the corresponding records at NCBI, individual gene reports at CRRD include functional annotations for Disease, Gene Ontology (GO) Biological Process, GO Molecular Function, GO Cellular Component and Pathway assigned to chinchilla genes based on annotations from the corresponding human orthologs. Data can be retrieved via keyword and gene-specific searches. Lists of genes with similar functional attributes can be assembled by leveraging the hierarchical structure of the Disease, GO and Pathway vocabularies through the Ontology Search and Browser tool. Such lists can then be further analyzed for commonalities using the Gene Annotator (GA) Tool. All data in the Chinchilla Research Resource Database is freely accessible and downloadable via the CRRD FTP site or using the download functions available in the search and analysis tools. The Chinchilla Research Resource Database is a rich resource for researchers using, or considering the use of, chinchilla as a model for human disease. Database URL: http://crrd.mcw.edu PMID:27173523
The intrapsychics of gender: a model of self-socialization.
Tobin, Desiree D; Menon, Meenakshi; Menon, Madhavi; Spatta, Brooke C; Hodges, Ernest V E; Perry, David G
2010-04-01
This article outlines a model of the structure and the dynamics of gender cognition in childhood. The model incorporates 3 hypotheses featured in different contemporary theories of childhood gender cognition and unites them under a single theoretical framework. Adapted from Greenwald et al. (2002), the model distinguishes three constructs: gender identity, gender stereotypes, and attribute self-perceptions. The model specifies 3 causal processes among the constructs: Gender identity and stereotypes interactively influence attribute self-perceptions (stereotype emulation hypothesis); gender identity and attribute self-perceptions interactively influence gender stereotypes (stereotype construction hypothesis); and gender stereotypes and attribute self-perceptions interactively influence identity (identity construction hypothesis). The model resolves nagging ambiguities in terminology, organizes diverse hypotheses and empirical findings under a unifying conceptual umbrella, and stimulates many new research directions. PsycINFO Database Record (c) 2010 APA, all rights reserved.
EPA Office of Water (OW): 2002 Impaired Waters Baseline NHDPlus Indexed Dataset
This dataset consists of geospatial and attribute data identifying the spatial extent of state-reported impaired waters (EPA's Integrated Reporting categories 4a, 4b, 4c and 5)* available in EPA's Reach Address Database (RAD) at the time of extraction. For the 2002 baseline reporting year, EPA compiled state-submitted GIS data to create a seamless and nationally consistent picture of the Nation's impaired waters for measuring progress. EPA's Assessment and TMDL Tracking and Implementation System (ATTAINS) is a national compilation of states' 303(d) listings and TMDL development information, spanning several years of tracking over 40,000 impaired waters.
Integrating Query of Relational and Textual Data in Clinical Databases: A Case Study
Fisk, John M.; Mutalik, Pradeep; Levin, Forrest W.; Erdos, Joseph; Taylor, Caroline; Nadkarni, Prakash
2003-01-01
Objectives: The authors designed and implemented a clinical data mart composed of an integrated information retrieval (IR) and relational database management system (RDBMS). Design: Using commodity software, which supports interactive, attribute-centric text and relational searches, the mart houses 2.8 million documents that span a five-year period and supports basic IR features such as Boolean searches, stemming, and proximity and fuzzy searching. Measurements: Results are relevance-ranked using either “total documents per patient” or “report type weighting.” Results: Non-curated medical text has a significant degree of malformation with respect to spelling and punctuation, which creates difficulties for text indexing and searching. Presently, the IR facilities of RDBMS packages lack the features necessary to handle such malformed text adequately. Conclusion: A robust IR+RDBMS system can be developed, but it requires integrating RDBMSs with third-party IR software. RDBMS vendors need to make their IR offerings more accessible to non-programmers. PMID:12509355
NASA Astrophysics Data System (ADS)
Brenden, T. O.; Clark, R. D.; Wiley, M. J.; Seelbach, P. W.; Wang, L.
2005-05-01
Remote sensing and geographic information systems have made it possible to attribute variables for streams at increasingly detailed resolutions (e.g., individual river reaches). Nevertheless, management decisions still must be made at large scales because land and stream managers typically lack sufficient resources to manage on an individual reach basis. Managers thus require a method for identifying stream management units that are ecologically similar and that can be expected to respond similarly to management decisions. We have developed a spatially-constrained clustering algorithm that can merge neighboring river reaches with similar ecological characteristics into larger management units. The clustering algorithm is based on the Cluster Affinity Search Technique (CAST), which was developed for clustering gene expression data. Inputs to the clustering algorithm are the neighbor relationships of the reaches that comprise the digital river network, the ecological attributes of the reaches, and an affinity value, which identifies the minimum similarity for merging river reaches. In this presentation, we describe the clustering algorithm in greater detail and contrast its use with other methods (expert opinion, classification approach, regular clustering) for identifying management units using several Michigan watersheds as a backdrop.
Professional nursing values: A concept analysis.
Schmidt, Bonnie J; McArthur, Erin C
2018-01-01
The aim of this concept analysis is to clarify the meaning of professional nursing values. In a time of increasing ethical dilemmas, it is essential that nurses internalize professional values to develop and maintain a professional identity. However, nursing organizations and researchers provide different conceptions of professional nursing values, leading to a lack of clarity as to the meaning and attributes of this construct. Walker and Avant's (2011) method was used to guide an analysis of this concept. Resources published from 1973 to 2016 were identified via electronic databases and hand-searching of reference lists. A review of the literature was completed and the data were analyzed to identify uses of the concept; the defining attributes of the concept; borderline, related, contrary, and illegitimate examples; antecedents and consequences; and empirical referents. Professional nursing values were defined as important professional nursing principles of human dignity, integrity, altruism, and justice that serve as a framework for standards, professional practice, and evaluation. Further research is needed in the development and testing of professional nursing values theory, and the reassessment of values instruments. Core professional values that are articulated may help unify the profession and demonstrate the value of nursing to the public. © 2017 Wiley Periodicals, Inc.
Sartorius, B; Sartorius, K; Aldous, C; Madiba, T E; Stefan, C; Noakes, T
2016-01-01
Introduction Linkages between carbohydrates, obesity and cancer continue to demonstrate conflicting results. Evidence suggests inconclusive direct linkages between carbohydrates and specific cancers. Conversely, obesity has been strongly linked to a wide range of cancers. The purpose of the study is to explore linkages between carbohydrate intake and cancer types using a two-step approach. First the study will evaluate the linkages between carbohydrate intake and obesity, potentially stratified by metabolic syndrome status. Second, the estimated attributable fraction of obesity ascribed to carbohydrate intake will be multiplied against obesity attributable fractions for cancer types to give estimated overall attributable fraction for carbohydrate versus cancer type. Methods and analysis We will perform a comprehensive search to identify all possible published and unpublished studies that have assessed risk factors for obesity including dietary carbohydrate intake. Scientific databases, namely PubMed MEDLINE, EMBASE, EBSCOhost and ISI Web of Science will be searched. Following study selection, paper/data acquisition, and data extraction and synthesis, we will appraise the quality of studies and risk of bias, as well as assess heterogeneity. Meta-weighted attributable fractions of obesity due to carbohydrate intake will be estimated after adjusting for other potential confounding factors (eg, physical inactivity, other dietary intake). Furthermore, previously published systematic reviews assessing the cancer-specific risk associated with obesity will also be drawn. These estimates will be linked with the attributability of carbohydrate intake in part 1 to estimate the cancer-specific burden that can be attributed to dietary carbohydrates. This systematic review protocol has been developed according to the ‘Preferred Reporting Items for Systematic review and Meta-Analysis Protocols (PRISMA-P) 2015’. Ethics and dissemination The current study will be based on published literature and data, and, as such, ethics approval is not required. The final results of this two part systematic review (plus multiplicative calculations) will be published in a relevant international peer-reviewed journal. Trial registration number PROSPERO CRD42015023257. PMID:26729382
NASA Astrophysics Data System (ADS)
Buśko, Małgorzata
2017-06-01
According to the original wording of the Regulation on the register of land and buildings of 2001, in the real estate cadastre there was one attribute associated with the use of a building structure - its intended use, which was applicable until the amendment to the Regulation was introduced in 2013. Then, additional attributes were added, i.e. the type of the building according to the Classification of Fixed Assets (KST), the class of the building according to the Polish Classification of Types of Constructions (PKOB) and, at the same time, the main functional use and other functions of the building remained in the Regulation as well. The record data on buildings are captured for the real estate cadastre from other data sets, for example those maintained by architectural and construction authorities. At the same time, the data contained in the cadastre, after they have been entered or changed in the database, are transferred to other registers, such as tax records, or land and mortgage court registers. This study is the result of the analysis of the laws applicable to the specific units and registers. A list of discrepancies in the attributes occurring in the different registers was prepared. The practical part of the study paid particular attention to the legal bases and procedures for entering the function of a building in the real estate cadastre, which is extremely significant, as it is the attribute determining the property tax basis.
User’s Guide and Metadata for the PICES Nonindigenous Species Information System
The database, the "PICES Nonindigenous Species Information System", was constucted to synthesize the global distributions, environmental tolerances, and natural history attributes of the nonindigenous species in the North Pacific and Hawaii. The User's Guide provides th...
Methods for Estimating Annual Wastewater Nutrient Loads in the Southeastern United States
McMahon, Gerard; Tervelt, Larinda; Donehoo, William
2007-01-01
This report describes an approach for estimating annual total nitrogen and total phosphorus loads from point-source dischargers in the southeastern United States. Nutrient load estimates for 2002 were used in the calibration and application of a regional nutrient model, referred to as the SPARROW (SPAtially Referenced Regression On Watershed attributes) watershed model. Loads from dischargers permitted under the National Pollutant Discharge Elimination System were calculated using data from the U.S. Environmental Protection Agency Permit Compliance System database and individual state databases. Site information from both state and U.S. Environmental Protection Agency databases, including latitude and longitude and monitored effluent data, was compiled into a project database. For sites with a complete effluent-monitoring record, effluent-flow and nutrient-concentration data were used to develop estimates of annual point-source nitrogen and phosphorus loads. When flow data were available but nutrient-concentration data were missing or incomplete, typical pollutant-concentration values of total nitrogen and total phosphorus were used to estimate load. In developing typical pollutant-concentration values, the major factors assumed to influence wastewater nutrient-concentration variability were the size of the discharger (the amount of flow), the season during which discharge occurred, and the Standard Industrial Classification code of the discharger. One insight gained from this study is that in order to gain access to flow, concentration, and location data, close communication and collaboration are required with the agencies that collect and manage the data. In addition, the accuracy and usefulness of the load estimates depend on the willingness of the states and the U.S. Environmental Protection Agency to provide guidance and review for at least a subset of the load estimates that may be problematic.
A WebGIS system on the base of satellite data processing system for marine application
NASA Astrophysics Data System (ADS)
Gong, Fang; Wang, Difeng; Huang, Haiqing; Chen, Jianyu
2007-10-01
From 2002 to 2004, a satellite data processing system for marine application had been built up in State Key Laboratory of Satellite Ocean Environment Dynamics (Second Institute of Oceanography, State Oceanic Administration). The system received satellite data from TERRA, AQUA, NOAA-12/15/16/17/18, FY-1D and automatically generated Level3 products and Level4 products(products of single orbit and merged multi-orbits products) deriving from Level0 data, which is controlled by an operational control sub-system. Currently, the products created by this system play an important role in the marine environment monitoring, disaster monitoring and researches. Now a distribution platform has been developed on this foundation, namely WebGIS system for querying and browsing of oceanic remote sensing data. This system is based upon large database system-Oracle. We made use of the space database engine of ArcSDE and other middleware to perform database operation in addition. J2EE frame was adopted as development model, and Oracle 9.2 DBMS as database background and server. Simply using standard browsers(such as IE6.0), users can visit and browse the public service information that provided by system, including browsing for oceanic remote sensing data, and enlarge, contract, move, renew, traveling, further data inquiry, attribution search and data download etc. The system is still under test now. Founding of such a system will become an important distribution platform of Chinese satellite oceanic environment products of special topic and category (including Sea surface temperature, Concentration of chlorophyll, and so on), for the exaltation of satellite products' utilization and promoting the data share and the research of the oceanic remote sensing platform.
Ben-Ari Fuchs, Shani; Lieder, Iris; Stelzer, Gil; Mazor, Yaron; Buzhor, Ella; Kaplan, Sergey; Bogoch, Yoel; Plaschkes, Inbar; Shitrit, Alina; Rappaport, Noa; Kohn, Asher; Edgar, Ron; Shenhav, Liraz; Safran, Marilyn; Lancet, Doron; Guan-Golan, Yaron; Warshawsky, David; Shtrichman, Ronit
2016-03-01
Postgenomics data are produced in large volumes by life sciences and clinical applications of novel omics diagnostics and therapeutics for precision medicine. To move from "data-to-knowledge-to-innovation," a crucial missing step in the current era is, however, our limited understanding of biological and clinical contexts associated with data. Prominent among the emerging remedies to this challenge are the gene set enrichment tools. This study reports on GeneAnalytics™ ( geneanalytics.genecards.org ), a comprehensive and easy-to-apply gene set analysis tool for rapid contextualization of expression patterns and functional signatures embedded in the postgenomics Big Data domains, such as Next Generation Sequencing (NGS), RNAseq, and microarray experiments. GeneAnalytics' differentiating features include in-depth evidence-based scoring algorithms, an intuitive user interface and proprietary unified data. GeneAnalytics employs the LifeMap Science's GeneCards suite, including the GeneCards®--the human gene database; the MalaCards-the human diseases database; and the PathCards--the biological pathways database. Expression-based analysis in GeneAnalytics relies on the LifeMap Discovery®--the embryonic development and stem cells database, which includes manually curated expression data for normal and diseased tissues, enabling advanced matching algorithm for gene-tissue association. This assists in evaluating differentiation protocols and discovering biomarkers for tissues and cells. Results are directly linked to gene, disease, or cell "cards" in the GeneCards suite. Future developments aim to enhance the GeneAnalytics algorithm as well as visualizations, employing varied graphical display items. Such attributes make GeneAnalytics a broadly applicable postgenomics data analyses and interpretation tool for translation of data to knowledge-based innovation in various Big Data fields such as precision medicine, ecogenomics, nutrigenomics, pharmacogenomics, vaccinomics, and others yet to emerge on the postgenomics horizon.
Devleesschauwer, Brecht; Haagsma, Juanita A.; Angulo, Frederick J.; Bellinger, David C.; Cole, Dana; Döpfer, Dörte; Fazil, Aamir; Fèvre, Eric M.; Gibb, Herman J.; Hald, Tine; Kirk, Martyn D.; Lake, Robin J.; Maertens de Noordhout, Charline; Mathers, Colin D.; McDonald, Scott A.; Pires, Sara M.; Speybroeck, Niko; Thomas, M. Kate; Torgerson, Paul R.; Wu, Felicia; Havelaar, Arie H.; Praet, Nicolas
2015-01-01
Background The Foodborne Disease Burden Epidemiology Reference Group (FERG) was established in 2007 by the World Health Organization to estimate the global burden of foodborne diseases (FBDs). This paper describes the methodological framework developed by FERG's Computational Task Force to transform epidemiological information into FBD burden estimates. Methods and Findings The global and regional burden of 31 FBDs was quantified, along with limited estimates for 5 other FBDs, using Disability-Adjusted Life Years in a hazard- and incidence-based approach. To accomplish this task, the following workflow was defined: outline of disease models and collection of epidemiological data; design and completion of a database template; development of an imputation model; identification of disability weights; probabilistic burden assessment; and estimating the proportion of the disease burden by each hazard that is attributable to exposure by food (i.e., source attribution). All computations were performed in R and the different functions were compiled in the R package 'FERG'. Traceability and transparency were ensured by sharing results and methods in an interactive way with all FERG members throughout the process. Conclusions We developed a comprehensive framework for estimating the global burden of FBDs, in which methodological simplicity and transparency were key elements. All the tools developed have been made available and can be translated into a user-friendly national toolkit for studying and monitoring food safety at the local level. PMID:26633883
Introduction to the DISRUPT postprandial database: subjects, studies and methodologies.
Jackson, Kim G; Clarke, Dave T; Murray, Peter; Lovegrove, Julie A; O'Malley, Brendan; Minihane, Anne M; Williams, Christine M
2010-03-01
Dysregulation of lipid and glucose metabolism in the postprandial state are recognised as important risk factors for the development of cardiovascular disease and type 2 diabetes. Our objective was to create a comprehensive, standardised database of postprandial studies to provide insights into the physiological factors that influence postprandial lipid and glucose responses. Data were collated from subjects (n = 467) taking part in single and sequential meal postprandial studies conducted by researchers at the University of Reading, to form the DISRUPT (DIetary Studies: Reading Unilever Postprandial Trials) database. Subject attributes including age, gender, genotype, menopausal status, body mass index, blood pressure and a fasting biochemical profile, together with postprandial measurements of triacylglycerol (TAG), non-esterified fatty acids, glucose, insulin and TAG-rich lipoprotein composition are recorded. A particular strength of the studies is the frequency of blood sampling, with on average 10-13 blood samples taken during each postprandial assessment, and the fact that identical test meal protocols were used in a number of studies, allowing pooling of data to increase statistical power. The DISRUPT database is the most comprehensive postprandial metabolism database that exists worldwide and preliminary analysis of the pooled sequential meal postprandial dataset has revealed both confirmatory and novel observations with respect to the impact of gender and age on the postprandial TAG response. Further analysis of the dataset using conventional statistical techniques along with integrated mathematical models and clustering analysis will provide a unique opportunity to greatly expand current knowledge of the aetiology of inter-individual variability in postprandial lipid and glucose responses.
Geologic Map Database of Texas
Stoeser, Douglas B.; Shock, Nancy; Green, Gregory N.; Dumonceaux, Gayle M.; Heran, William D.
2005-01-01
The purpose of this report is to release a digital geologic map database for the State of Texas. This database was compiled for the U.S. Geological Survey (USGS) Minerals Program, National Surveys and Analysis Project, whose goal is a nationwide assemblage of geologic, geochemical, geophysical, and other data. This release makes the geologic data from the Geologic Map of Texas available in digital format. Original clear film positives provided by the Texas Bureau of Economic Geology were photographically enlarged onto Mylar film. These films were scanned, georeferenced, digitized, and attributed by Geologic Data Systems (GDS), Inc., Denver, Colorado. Project oversight and quality control was the responsibility of the U.S. Geological Survey. ESRI ArcInfo coverages, AMLs, and shapefiles are provided.
Sensory overload: A concept analysis.
Scheydt, Stefan; Müller Staub, Maria; Frauenfelder, Fritz; Nielsen, Gunnar H; Behrens, Johann; Needham, Ian
2017-04-01
In the context of mental disorders sensory overload is a widely described phenomenon used in conjunction with psychiatric interventions such as removal from stimuli. However, the theoretical foundation of sensory overload as addressed in the literature can be described as insufficient and fragmentary. To date, the concept of sensory overload has not yet been sufficiently specified or analyzed. The aim of the study was to analyze the concept of sensory overload in mental health care. A literature search was undertaken using specific electronic databases, specific journals and websites, hand searches, specific library catalogues, and electronic publishing databases. Walker and Avant's method of concept analysis was used to analyze the sources included in the analysis. All aspects of the method of Walker and Avant were covered in this concept analysis. The conceptual understanding has become more focused, the defining attributes, influencing factors and consequences are described and empirical referents identified. The concept analysis is a first step in the development of a middle-range descriptive theory of sensory overload based on social scientific and stress-theoretical approaches. This specification may serve as a fundament for further research, for the development of a nursing diagnosis or for guidelines. © 2017 Australian College of Mental Health Nurses Inc.
Asquith, William H.
2014-01-01
A database containing more than 16,300 discharge values and ancillary hydraulic attributes was assembled from summaries of discharge measurement records for 391 USGS streamflow-gauging stations (streamgauges) in Texas. Each discharge is between the 40th- and 60th-percentile daily mean streamflow as determined by period-of-record, streamgauge-specific, flow-duration curves. Each discharge therefore is assumed to represent a discharge measurement made for near-median streamflow conditions, and such conditions are conceptualized as representative of midrange to baseflow conditions in much of the state. The hydraulic attributes of each discharge measurement included concomitant cross-section flow area, water-surface top width, and reported mean velocity. Two regression equations are presented: (1) an expression for discharge and (2) an expression for mean velocity, both as functions of selected hydraulic attributes and watershed characteristics. Specifically, the discharge equation uses cross-sectional area, water-surface top width, contributing drainage area of the watershed, and mean annual precipitation of the location; the equation has an adjusted R-squared of approximately 0.95 and residual standard error of approximately 0.23 base-10 logarithm (cubic meters per second). The mean velocity equation uses discharge, water-surface top width, contributing drainage area, and mean annual precipitation; the equation has an adjusted R-squared of approximately 0.50 and residual standard error of approximately 0.087 third root (meters per second). Residual plots from both equations indicate that reliable estimates of discharge and mean velocity at ungauged stream sites are possible. Further, the relation between contributing drainage area and main-channel slope (a measure of whole-watershed slope) is depicted to aid analyst judgment of equation applicability for ungauged sites. Example applications and computations are provided and discussed within a real-world, discharge-measurement scenario, and an illustration of the development of a preliminary stage-discharge relation using the discharge equation is given.
Rejecting a bad option feels like choosing a good one.
Perfecto, Hannah; Galak, Jeff; Simmons, Joseph P; Nelson, Leif D
2017-11-01
Across 4,151 participants, the authors demonstrate a novel framing effect, attribute matching, whereby matching a salient attribute of a decision frame with that of a decision's options facilitates decision-making. This attribute matching is shown to increase decision confidence and, ultimately, consensus estimates by increasing feelings of metacognitive ease. In Study 1, participants choosing the more attractive of two faces or rejecting the less attractive face reported greater confidence in and perceived consensus around their decision. Using positive and negative words, Study 2 showed that the attribute's extremity moderates the size of the effect. Study 3 found decision ease mediates these changes in confidence and consensus estimates. Consistent with a misattribution account, when participants were warned about this external source of ease in Study 4, the effect disappeared. Study 5 extended attribute matching beyond valence to objective judgments. The authors conclude by discussing related psychological constructs as well as downstream consequences. (PsycINFO Database Record (c) 2017 APA, all rights reserved).
Digital hand atlas and computer-aided bone age assessment via the Web
NASA Astrophysics Data System (ADS)
Cao, Fei; Huang, H. K.; Pietka, Ewa; Gilsanz, Vicente
1999-07-01
A frequently used assessment method of bone age is atlas matching by a radiological examination of a hand image against a reference set of atlas patterns of normal standards. We are in a process of developing a digital hand atlas with a large standard set of normal hand and wrist images that reflect the skeletal maturity, race and sex difference, and current child development. The digital hand atlas will be used for a computer-aided bone age assessment via Web. We have designed and partially implemented a computer-aided diagnostic (CAD) system for Web-based bone age assessment. The system consists of a digital hand atlas, a relational image database and a Web-based user interface. The digital atlas is based on a large standard set of normal hand an wrist images with extracted bone objects and quantitative features. The image database uses a content- based indexing to organize the hand images and their attributes and present to users in a structured way. The Web-based user interface allows users to interact with the hand image database from browsers. Users can use a Web browser to push a clinical hand image to the CAD server for a bone age assessment. Quantitative features on the examined image, which reflect the skeletal maturity, will be extracted and compared with patterns from the atlas database to assess the bone age. The relevant reference imags and the final assessment report will be sent back to the user's browser via Web. The digital atlas will remove the disadvantages of the currently out-of-date one and allow the bone age assessment to be computerized and done conveniently via Web. In this paper, we present the system design and Web-based client-server model for computer-assisted bone age assessment and our initial implementation of the digital atlas database.
Fine-grained Database Field Search Using Attribute-Based Encryption for E-Healthcare Clouds.
Guo, Cheng; Zhuang, Ruhan; Jie, Yingmo; Ren, Yizhi; Wu, Ting; Choo, Kim-Kwang Raymond
2016-11-01
An effectively designed e-healthcare system can significantly enhance the quality of access and experience of healthcare users, including facilitating medical and healthcare providers in ensuring a smooth delivery of services. Ensuring the security of patients' electronic health records (EHRs) in the e-healthcare system is an active research area. EHRs may be outsourced to a third-party, such as a community healthcare cloud service provider for storage due to cost-saving measures. Generally, encrypting the EHRs when they are stored in the system (i.e. data-at-rest) or prior to outsourcing the data is used to ensure data confidentiality. Searchable encryption (SE) scheme is a promising technique that can ensure the protection of private information without compromising on performance. In this paper, we propose a novel framework for controlling access to EHRs stored in semi-trusted cloud servers (e.g. a private cloud or a community cloud). To achieve fine-grained access control for EHRs, we leverage the ciphertext-policy attribute-based encryption (CP-ABE) technique to encrypt tables published by hospitals, including patients' EHRs, and the table is stored in the database with the primary key being the patient's unique identity. Our framework can enable different users with different privileges to search on different database fields. Differ from previous attempts to secure outsourcing of data, we emphasize the control of the searches of the fields within the database. We demonstrate the utility of the scheme by evaluating the scheme using datasets from the University of California, Irvine.
Digital release of the Alaska Quaternary fault and fold database
NASA Astrophysics Data System (ADS)
Koehler, R. D.; Farrell, R.; Burns, P.; Combellick, R. A.; Weakland, J. R.
2011-12-01
The Alaska Division of Geological & Geophysical Surveys (DGGS) has designed a Quaternary fault and fold database for Alaska in conformance with standards defined by the U.S. Geological Survey for the National Quaternary fault and fold database. Alaska is the most seismically active region of the United States, however little information exists on the location, style of deformation, and slip rates of Quaternary faults. Thus, to provide an accurate, user-friendly, reference-based fault inventory to the public, we are producing a digital GIS shapefile of Quaternary fault traces and compiling summary information on each fault. Here, we present relevant information pertaining to the digital GIS shape file and online access and availability of the Alaska database. This database will be useful for engineering geologic studies, geologic, geodetic, and seismic research, and policy planning. The data will also contribute to the fault source database being constructed by the Global Earthquake Model (GEM), Faulted Earth project, which is developing tools to better assess earthquake risk. We derived the initial list of Quaternary active structures from The Neotectonic Map of Alaska (Plafker et al., 1994) and supplemented it with more recent data where available. Due to the limited level of knowledge on Quaternary faults in Alaska, pre-Quaternary fault traces from the Plafker map are shown as a layer in our digital database so users may view a more accurate distribution of mapped faults and to suggest the possibility that some older traces may be active yet un-studied. The database will be updated as new information is developed. We selected each fault by reviewing the literature and georegistered the faults from 1:250,000-scale paper maps contained in 1970's vintage and earlier bedrock maps. However, paper map scales range from 1:20,000 to 1:500,000. Fault parameters in our GIS fault attribute tables include fault name, age, slip rate, slip sense, dip direction, fault line type (i.e., well constrained, moderately constrained, or inferred), and mapped scale. Each fault is assigned a three-integer CODE, based upon age, slip rate, and how well the fault is located. This CODE dictates the line-type for the GIS files. To host the database, we are developing an interactive web-map application with ArcGIS for Server and the ArcGIS API for JavaScript from Environmental Systems Research Institute, Inc. (Esri). The web-map application will present the database through a visible scale range with each fault displayed at the resolution of the original map. Application functionality includes: search by name or location, identification of fault by manual selection, and choice of base map. Base map options include topographic, satellite imagery, and digital elevation maps available from ArcGIS on-line. We anticipate that the database will be publically accessible from a portal embedded on the DGGS website by the end of 2011.
Pritoni, Marco; Ford, Rebecca; Karlin, Beth; Sanguinetti, Angela
2018-02-01
Policymakers worldwide are currently discussing whether to include home energy management (HEM) products in their portfolio of technologies to reduce carbon emissions and improve grid reliability. However, very little data is available about these products. Here we present the results of an extensive review including 308 HEM products available on the US market in 2015-2016. We gathered these data from publicly available sources such as vendor websites, online marketplaces and other vendor documents. A coding guide was developed iteratively during the data collection and utilized to classify the devices. Each product was coded based on 96 distinct attributes, grouped into 11 categories: Identifying information, Product components, Hardware, Communication, Software, Information - feedback, Information - feedforward, Control, Utility interaction, Additional benefits and Usability. The codes describe product features and functionalities, user interaction and interoperability with other devices. A mix of binary attributes and more descriptive codes allow to sort and group data without losing important qualitative information. The information is stored in a large spreadsheet included with this article, along with an explanatory coding guide. This dataset is analyzed and described in a research article entitled "Categories and functionality of smart home technology for energy management" (Ford et al., 2017) [1].
Modeling first impressions from highly variable facial images.
Vernon, Richard J W; Sutherland, Clare A M; Young, Andrew W; Hartley, Tom
2014-08-12
First impressions of social traits, such as trustworthiness or dominance, are reliably perceived in faces, and despite their questionable validity they can have considerable real-world consequences. We sought to uncover the information driving such judgments, using an attribute-based approach. Attributes (physical facial features) were objectively measured from feature positions and colors in a database of highly variable "ambient" face photographs, and then used as input for a neural network to model factor dimensions (approachability, youthful-attractiveness, and dominance) thought to underlie social attributions. A linear model based on this approach was able to account for 58% of the variance in raters' impressions of previously unseen faces, and factor-attribute correlations could be used to rank attributes by their importance to each factor. Reversing this process, neural networks were then used to predict facial attributes and corresponding image properties from specific combinations of factor scores. In this way, the factors driving social trait impressions could be visualized as a series of computer-generated cartoon face-like images, depicting how attributes change along each dimension. This study shows that despite enormous variation in ambient images of faces, a substantial proportion of the variance in first impressions can be accounted for through linear changes in objectively defined features.
The Cerrado (Brazil) plant cytogenetics database.
Roa, Fernando; Telles, Mariana Pires de Campos
2017-01-01
Cerrado is a biodiversity hotspot that has lost ca. 50% of its original vegetation cover and hosts ca. 11,000 species belonging to 1,423 genera of phanerogams. For a fraction of those species some cytogenetic characteristics like chromosome numbers and C-value were available in databases, while other valuable information such as karyotype formula and banding patterns are missing. In order to integrate and share all cytogenetic information published for Cerrado species, including frequency of cytogenetic attributes and scientometrics aspects, Cerrado plant species were searched in bibliographic sources, including the 50 richest genera (with more than 45 taxa) and 273 genera with only one species in Cerrado. Determination of frequencies and the database website (http://cyto.shinyapps.io/cerrado) were developed in R. Studies were pooled by employed technique and decade, showing a rise in non-conventional cytogenetics since 2000. However, C-value estimation, heterochromatin staining and molecular cytogenetics are still not common for any family. For the richest and best sampled families, the following modal 2n counts were observed: Oxalidaceae 2n = 12, Lythraceae 2n = 30, Sapindaceae 2n = 24, Solanaceae 2n = 24, Cyperaceae 2n = 10, Poaceae 2n = 20, Asteraceae 2n = 18 and Fabaceae 2n = 26. Chromosome number information is available for only 16.1% of species, while there are genome size data for only 1.25%, being lower than the global percentages. In general, genome sizes were small, ranging from 2C = ca. 1.5 to ca. 3.5 pg. Intra-specific 2n number variation and higher 2n counts were mainly related to polyploidy, which relates to the prevalence of even haploid numbers above the mode of 2n in most major plant clades. Several orphan genera with almost no cytogenetic studies for Cerrado were identified. This effort represents a complete diagnosis for cytogenetic attributes of plants of Cerrado.
The Cerrado (Brazil) plant cytogenetics database
Roa, Fernando; Telles, Mariana Pires de Campos
2017-01-01
Abstract Cerrado is a biodiversity hotspot that has lost ca. 50% of its original vegetation cover and hosts ca. 11,000 species belonging to 1,423 genera of phanerogams. For a fraction of those species some cytogenetic characteristics like chromosome numbers and C-value were available in databases, while other valuable information such as karyotype formula and banding patterns are missing. In order to integrate and share all cytogenetic information published for Cerrado species, including frequency of cytogenetic attributes and scientometrics aspects, Cerrado plant species were searched in bibliographic sources, including the 50 richest genera (with more than 45 taxa) and 273 genera with only one species in Cerrado. Determination of frequencies and the database website (http://cyto.shinyapps.io/cerrado) were developed in R. Studies were pooled by employed technique and decade, showing a rise in non-conventional cytogenetics since 2000. However, C-value estimation, heterochromatin staining and molecular cytogenetics are still not common for any family. For the richest and best sampled families, the following modal 2n counts were observed: Oxalidaceae 2n = 12, Lythraceae 2n = 30, Sapindaceae 2n = 24, Solanaceae 2n = 24, Cyperaceae 2n = 10, Poaceae 2n = 20, Asteraceae 2n = 18 and Fabaceae 2n = 26. Chromosome number information is available for only 16.1% of species, while there are genome size data for only 1.25%, being lower than the global percentages. In general, genome sizes were small, ranging from 2C = ca. 1.5 to ca. 3.5 pg. Intra-specific 2n number variation and higher 2n counts were mainly related to polyploidy, which relates to the prevalence of even haploid numbers above the mode of 2n in most major plant clades. Several orphan genera with almost no cytogenetic studies for Cerrado were identified. This effort represents a complete diagnosis for cytogenetic attributes of plants of Cerrado. PMID:28919965
Abugessaisa, Imad; Gomez-Cabrero, David; Snir, Omri; Lindblad, Staffan; Klareskog, Lars; Malmström, Vivianne; Tegnér, Jesper
2013-04-02
Sequencing of the human genome and the subsequent analyses have produced immense volumes of data. The technological advances have opened new windows into genomics beyond the DNA sequence. In parallel, clinical practice generate large amounts of data. This represents an underused data source that has much greater potential in translational research than is currently realized. This research aims at implementing a translational medicine informatics platform to integrate clinical data (disease diagnosis, diseases activity and treatment) of Rheumatoid Arthritis (RA) patients from Karolinska University Hospital and their research database (biobanks, genotype variants and serology) at the Center for Molecular Medicine, Karolinska Institutet. Requirements engineering methods were utilized to identify user requirements. Unified Modeling Language and data modeling methods were used to model the universe of discourse and data sources. Oracle11g were used as the database management system, and the clinical development center (CDC) was used as the application interface. Patient data were anonymized, and we employed authorization and security methods to protect the system. We developed a user requirement matrix, which provided a framework for evaluating three translation informatics systems. The implementation of the CDC successfully integrated biological research database (15172 DNA, serum and synovial samples, 1436 cell samples and 65 SNPs per patient) and clinical database (5652 clinical visit) for the cohort of 379 patients presents three profiles. Basic functionalities provided by the translational medicine platform are research data management, development of bioinformatics workflow and analysis, sub-cohort selection, and re-use of clinical data in research settings. Finally, the system allowed researchers to extract subsets of attributes from cohorts according to specific biological, clinical, or statistical features. Research and clinical database integration is a real challenge and a road-block in translational research. Through this research we addressed the challenges and demonstrated the usefulness of CDC. We adhered to ethical regulations pertaining to patient data, and we determined that the existing software solutions cannot meet the translational research needs at hand. We used RA as a test case since we have ample data on active and longitudinal cohort.
2013-01-01
Background Sequencing of the human genome and the subsequent analyses have produced immense volumes of data. The technological advances have opened new windows into genomics beyond the DNA sequence. In parallel, clinical practice generate large amounts of data. This represents an underused data source that has much greater potential in translational research than is currently realized. This research aims at implementing a translational medicine informatics platform to integrate clinical data (disease diagnosis, diseases activity and treatment) of Rheumatoid Arthritis (RA) patients from Karolinska University Hospital and their research database (biobanks, genotype variants and serology) at the Center for Molecular Medicine, Karolinska Institutet. Methods Requirements engineering methods were utilized to identify user requirements. Unified Modeling Language and data modeling methods were used to model the universe of discourse and data sources. Oracle11g were used as the database management system, and the clinical development center (CDC) was used as the application interface. Patient data were anonymized, and we employed authorization and security methods to protect the system. Results We developed a user requirement matrix, which provided a framework for evaluating three translation informatics systems. The implementation of the CDC successfully integrated biological research database (15172 DNA, serum and synovial samples, 1436 cell samples and 65 SNPs per patient) and clinical database (5652 clinical visit) for the cohort of 379 patients presents three profiles. Basic functionalities provided by the translational medicine platform are research data management, development of bioinformatics workflow and analysis, sub-cohort selection, and re-use of clinical data in research settings. Finally, the system allowed researchers to extract subsets of attributes from cohorts according to specific biological, clinical, or statistical features. Conclusions Research and clinical database integration is a real challenge and a road-block in translational research. Through this research we addressed the challenges and demonstrated the usefulness of CDC. We adhered to ethical regulations pertaining to patient data, and we determined that the existing software solutions cannot meet the translational research needs at hand. We used RA as a test case since we have ample data on active and longitudinal cohort. PMID:23548156
The Perfectly Organized Search Service.
ERIC Educational Resources Information Center
Leach, Sandra Sinsel; Spencer, Mary Ellen
1993-01-01
Describes the evolution and operation of the successful Database Search Service (DSS) at the John C. Hodges Library, University of Tennessee, with detailed information about equipment, policies, software, training, and physical layout. Success is attributed to careful administration, standardization of search equipment and interfaces, staff…
Material Property Database and Environmental Attribute Models for NM Science Research
2011-03-28
3 Goals 1. Provide place "to go " for initial information 2. Basic understanding of what types of information you might need to... MWCNT , SWCNT, Fullerene, Waste • Aluminum ...,. Explosive, propellant • Silver ...,. Coatings, textiles, polymers • Titanium dioxide
Applications of spatial statistical network models to stream data
Isaak, Daniel J.; Peterson, Erin E.; Ver Hoef, Jay M.; Wenger, Seth J.; Falke, Jeffrey A.; Torgersen, Christian E.; Sowder, Colin; Steel, E. Ashley; Fortin, Marie-Josée; Jordan, Chris E.; Ruesch, Aaron S.; Som, Nicholas; Monestiez, Pascal
2014-01-01
Streams and rivers host a significant portion of Earth's biodiversity and provide important ecosystem services for human populations. Accurate information regarding the status and trends of stream resources is vital for their effective conservation and management. Most statistical techniques applied to data measured on stream networks were developed for terrestrial applications and are not optimized for streams. A new class of spatial statistical model, based on valid covariance structures for stream networks, can be used with many common types of stream data (e.g., water quality attributes, habitat conditions, biological surveys) through application of appropriate distributions (e.g., Gaussian, binomial, Poisson). The spatial statistical network models account for spatial autocorrelation (i.e., nonindependence) among measurements, which allows their application to databases with clustered measurement locations. Large amounts of stream data exist in many areas where spatial statistical analyses could be used to develop novel insights, improve predictions at unsampled sites, and aid in the design of efficient monitoring strategies at relatively low cost. We review the topic of spatial autocorrelation and its effects on statistical inference, demonstrate the use of spatial statistics with stream datasets relevant to common research and management questions, and discuss additional applications and development potential for spatial statistics on stream networks. Free software for implementing the spatial statistical network models has been developed that enables custom applications with many stream databases.
Ronfard, Samuel; Harris, Paul L
2014-01-01
As children listen to a simple action-based narrative, they construct a dynamic representation of the protagonist's movements, visual perspective, and goal-directed thoughts. We examined children's representations of more complex narratives in which the protagonist will encounter an unexpected outcome upon reaching his or her goal. Three studies involving 105 children between 3 and 6 years of age showed that children shifted in the mental states they attributed depending on the distance of the protagonist from the unexpected outcome. Even though children consistently recognized that the protagonist did not know about the surprise at any point, they increasingly attributed feelings and thoughts consistent with the surprise. The studies highlight the degree to which children's mental state attributions are dynamic rather than fixed by their current theory of mind. PsycINFO Database Record (c) 2014 APA, all rights reserved.
Attributes of clinical leadership in contemporary nursing: an integrative review.
Mannix, Judy; Wilkes, Lesley; Daly, John
2013-08-01
Effective clinical leadership is offered as the key to healthy, functional and supportive work environments for nurses and other health professionals. However, as a concept it lacks a standard definition and is poorly understood. This paper reports on an integrative review undertaken to uncover current understandings of defining attributes of contemporary clinical leadership in nursing. Data collection involved a search of relevant electronic databases for a 10-year period. Keywords for the search were 'clinical leadership' and 'nursing'. Ten research papers met the inclusion criteria for the integrative review. Analysis of these studies indicated clinical leadership attributes had a clinical focus, a follower/team focus or a personal qualities focus; attributes necessary to sustain supportive workplaces and build the capacity and resilience of nursing workforces. The small number of research-based studies yielded for the review indicates the need for further research in the area of clinical leadership.
Kalaiselvan, Vivekanandan; Sharma, Surbhi; Singh, Gyanendra Nath
2014-09-01
Contrast media are used widely to improve medical imaging. Like all other pharmaceuticals, these agents are not completely devoid of risk, and continuous monitoring of adverse reactions with these agents is important. Spontaneous reporting is the simplest method for understanding the safety profile of pharmaceutical products after their approval. Our objective was to identify the pattern and characteristics of adverse reactions attributed to contrast media in the Indian population reported to the National Coordination Centre for the Pharmacovigilance Programme of India (NCC-PvPI). Individual case safety reports (ICSRs) attributed to contrast media submitted spontaneously to the NCC-PvPI were extracted from the database for July 2010 to September 2013. We analysed these reports for information related to reporter's professional category, patient's age and sex, reporter's diagnosis of the reaction, seriousness of the reaction, type of contrast media exposure, system organ class (SOC) affected (as described in World Health Organization Adverse Reaction Terminology [WHO-ART]) and outcome. Of the total 59,915 ICSRs in the database, 415 (0.7%) were suspected adverse reactions to contrast media; 44 reports were serious, including three fatal cases. The most affected SOCs were skin and appendage disorders, body as a whole-general disorders, gastrointestinal system disorders and respiratory system disorders. Hypersensitivity reactions were reported in the majority of ICSRs. The contrast media with the highest number of reports were iohexol (40.7%), iomeprol (17.8%), iopamidol (12%) and diatrizoate (12%). Most of the reactions to contrast media were allergic-like, and no previously unrecognised adverse reactions were observed in the Indian population. Further data and increased awareness among healthcare professionals is required to signal and prevent the consequences of adverse reactions attributed to contrast media.
Metagenomic Taxonomy-Guided Database-Searching Strategy for Improving Metaproteomic Analysis.
Xiao, Jinqiu; Tanca, Alessandro; Jia, Ben; Yang, Runqing; Wang, Bo; Zhang, Yu; Li, Jing
2018-04-06
Metaproteomics provides a direct measure of the functional information by investigating all proteins expressed by a microbiota. However, due to the complexity and heterogeneity of microbial communities, it is very hard to construct a sequence database suitable for a metaproteomic study. Using a public database, researchers might not be able to identify proteins from poorly characterized microbial species, while a sequencing-based metagenomic database may not provide adequate coverage for all potentially expressed protein sequences. To address this challenge, we propose a metagenomic taxonomy-guided database-search strategy (MT), in which a merged database is employed, consisting of both taxonomy-guided reference protein sequences from public databases and proteins from metagenome assembly. By applying our MT strategy to a mock microbial mixture, about two times as many peptides were detected as with the metagenomic database only. According to the evaluation of the reliability of taxonomic attribution, the rate of misassignments was comparable to that obtained using an a priori matched database. We also evaluated the MT strategy with a human gut microbial sample, and we found 1.7 times as many peptides as using a standard metagenomic database. In conclusion, our MT strategy allows the construction of databases able to provide high sensitivity and precision in peptide identification in metaproteomic studies, enabling the detection of proteins from poorly characterized species within the microbiota.
NASA Technical Reports Server (NTRS)
Orcutt, John M.; Brenton, James C.
2016-01-01
An accurate database of meteorological data is essential for designing any aerospace vehicle and for preparing launch commit criteria. Meteorological instrumentation were recently placed on the three Lightning Protection System (LPS) towers at Kennedy Space Center (KSC) launch complex 39B (LC-39B), which provide a unique meteorological dataset existing at the launch complex over an extensive altitude range. Data records of temperature, dew point, relative humidity, wind speed, and wind direction are produced at 40, 78, 116, and 139 m at each tower. The Marshall Space Flight Center Natural Environments Branch (EV44) received an archive that consists of one-minute averaged measurements for the period of record of January 2011 - April 2015. However, before the received database could be used EV44 needed to remove any erroneous data from within the database through a comprehensive quality control (QC) process. The QC process applied to the LPS towers' meteorological data is similar to other QC processes developed by EV44, which were used in the creation of meteorological databases for other towers at KSC. The QC process utilized in this study has been modified specifically for use with the LPS tower database. The QC process first includes a check of each individual sensor. This check includes removing any unrealistic data and checking the temporal consistency of each variable. Next, data from all three sensors at each height are checked against each other, checked against climatology, and checked for sensors that erroneously report a constant value. Then, a vertical consistency check of each variable at each tower is completed. Last, the upwind sensor at each level is selected to minimize the influence of the towers and other structures at LC-39B on the measurements. The selection process for the upwind sensor implemented a study of tower-induced turbulence. This paper describes in detail the QC process, QC results, and the attributes of the LPS towers meteorological database.
Evaluation of personal digital assistant drug information databases for the managed care pharmacist.
Lowry, Colleen M; Kostka-Rokosz, Maria D; McCloskey, William W
2003-01-01
Personal digital assistants (PDAs) are becoming a necessity for practicing pharmacists. They offer a time-saving and convenient way to obtain current drug information. Several software companies now offer general drug information databases for use on hand held computers. PDAs priced less than 200 US dollars often have limited memory capacity; therefore, the user must choose from a growing list of general drug information database options in order to maximize utility without exceeding memory capacity. This paper reviews the attributes of available general drug information software databases for the PDA. It provides information on the content, advantages, limitations, pricing, memory requirements, and accessibility of drug information software databases. Ten drug information databases were subjectively analyzed and evaluated based on information from the product.s Web site, vendor Web sites, and from our experience. Some of these databases have attractive auxiliary features such as kinetics calculators, disease references, drug-drug and drug-herb interaction tools, and clinical guidelines, which may make them more useful to the PDA user. Not all drug information databases are equal with regard to content, author credentials, frequency of updates, and memory requirements. The user must therefore evaluate databases for completeness, currency, and cost effectiveness before purchase. In addition, consideration should be given to the ease of use and flexibility of individual programs.
The Human Phenotype Ontology project: linking molecular biology and disease through phenotype data
Köhler, Sebastian; Doelken, Sandra C.; Mungall, Christopher J.; Bauer, Sebastian; Firth, Helen V.; Bailleul-Forestier, Isabelle; Black, Graeme C. M.; Brown, Danielle L.; Brudno, Michael; Campbell, Jennifer; FitzPatrick, David R.; Eppig, Janan T.; Jackson, Andrew P.; Freson, Kathleen; Girdea, Marta; Helbig, Ingo; Hurst, Jane A.; Jähn, Johanna; Jackson, Laird G.; Kelly, Anne M.; Ledbetter, David H.; Mansour, Sahar; Martin, Christa L.; Moss, Celia; Mumford, Andrew; Ouwehand, Willem H.; Park, Soo-Mi; Riggs, Erin Rooney; Scott, Richard H.; Sisodiya, Sanjay; Vooren, Steven Van; Wapner, Ronald J.; Wilkie, Andrew O. M.; Wright, Caroline F.; Vulto-van Silfhout, Anneke T.; de Leeuw, Nicole; de Vries, Bert B. A.; Washingthon, Nicole L.; Smith, Cynthia L.; Westerfield, Monte; Schofield, Paul; Ruef, Barbara J.; Gkoutos, Georgios V.; Haendel, Melissa; Smedley, Damian; Lewis, Suzanna E.; Robinson, Peter N.
2014-01-01
The Human Phenotype Ontology (HPO) project, available at http://www.human-phenotype-ontology.org, provides a structured, comprehensive and well-defined set of 10,088 classes (terms) describing human phenotypic abnormalities and 13,326 subclass relations between the HPO classes. In addition we have developed logical definitions for 46% of all HPO classes using terms from ontologies for anatomy, cell types, function, embryology, pathology and other domains. This allows interoperability with several resources, especially those containing phenotype information on model organisms such as mouse and zebrafish. Here we describe the updated HPO database, which provides annotations of 7,278 human hereditary syndromes listed in OMIM, Orphanet and DECIPHER to classes of the HPO. Various meta-attributes such as frequency, references and negations are associated with each annotation. Several large-scale projects worldwide utilize the HPO for describing phenotype information in their datasets. We have therefore generated equivalence mappings to other phenotype vocabularies such as LDDB, Orphanet, MedDRA, UMLS and phenoDB, allowing integration of existing datasets and interoperability with multiple biomedical resources. We have created various ways to access the HPO database content using flat files, a MySQL database, and Web-based tools. All data and documentation on the HPO project can be found online. PMID:24217912
Biological data integration: wrapping data and tools.
Lacroix, Zoé
2002-06-01
Nowadays scientific data is inevitably digital and stored in a wide variety of formats in heterogeneous systems. Scientists need to access an integrated view of remote or local heterogeneous data sources with advanced data accessing, analyzing, and visualization tools. Building a digital library for scientific data requires accessing and manipulating data extracted from flat files or databases, documents retrieved from the Web as well as data generated by software. We present an approach to wrapping web data sources, databases, flat files, or data generated by tools through a database view mechanism. Generally, a wrapper has two tasks: it first sends a query to the source to retrieve data and, second builds the expected output with respect to the virtual structure. Our wrappers are composed of a retrieval component based on an intermediate object view mechanism called search views mapping the source capabilities to attributes, and an eXtensible Markup Language (XML) engine, respectively, to perform these two tasks. The originality of the approach consists of: 1) a generic view mechanism to access seamlessly data sources with limited capabilities and 2) the ability to wrap data sources as well as the useful specific tools they may provide. Our approach has been developed and demonstrated as part of the multidatabase system supporting queries via uniform object protocol model (OPM) interfaces.
Study of style effects on OCR errors in the MEDLINE database
NASA Astrophysics Data System (ADS)
Garrison, Penny; Davis, Diane L.; Andersen, Tim L.; Barney Smith, Elisa H.
2005-01-01
The National Library of Medicine has developed a system for the automatic extraction of data from scanned journal articles to populate the MEDLINE database. Although the 5-engine OCR system used in this process exhibits good performance overall, it does make errors in character recognition that must be corrected in order for the process to achieve the requisite accuracy. The correction process works by feeding words that have characters with less than 100% confidence (as determined automatically by the OCR engine) to a human operator who then must manually verify the word or correct the error. The majority of these errors are contained in the affiliation information zone where the characters are in italics or small fonts. Therefore only affiliation information data is used in this research. This paper examines the correlation between OCR errors and various character attributes in the MEDLINE database, such as font size, italics, bold, etc. and OCR confidence levels. The motivation for this research is that if a correlation between the character style and types of errors exists it should be possible to use this information to improve operator productivity by increasing the probability that the correct word option is presented to the human editor. We have determined that this correlation exists, in particular for the case of characters with diacritics.
AR Based App for Tourist Attraction in ESKİ ÇARŞI (Safranbolu)
NASA Astrophysics Data System (ADS)
Polat, Merve; Rakıp Karaş, İsmail; Kahraman, İdris; Alizadehashrafi, Behnam
2016-10-01
This research is dealing with 3D modeling of historical and heritage landmarks of Safranbolu that are registered by UNESCO. This is an Augmented Reality (AR) based project in order to trigger virtual three-dimensional (3D) models, cultural music, historical photos, artistic features and animated text information. The aim is to propose a GIS-based approach with these features and add to the system as attribute data in a relational database. The database will be available in an AR-based application to provide information for the tourists.
De Jonckheere, Johan F; Gryseels, Sophie; Eddyani, Miriam
2012-08-01
We have isolated several free-living amoeba strains from the environment in Ghana, which have internal transcribed spacers, including the 5.8S rDNA, sequences similar to sequences attributed to Vahlkampfiidae (Heterolobosea) in databases. However, morphological examination shows that the isolates belong to the Hartmannellidae (Amoebozoa). We provide evidence that the sequences in the databases are wrongly classified as belonging to a genus or species of the Vahlkampfiidae, but rather belong to strains of the genus Hartmannella. Copyright © 2012 Elsevier GmbH. All rights reserved.
Costing Child Protective Services Staff Turnover.
ERIC Educational Resources Information Center
Graef, Michelle I.; Hill, Erick L.
2000-01-01
Details process of determining a child welfare agency's actual dollar costs directly attributed to protective services staff turnover, using the agency's human resources database and interviews with administrative personnel. Provides formulas and process for calculating specific cost elements due to employee separation, replacement, and training.…
NASA Astrophysics Data System (ADS)
Fontaine, Alain; Sauvage, Bastien; Pétetin, Hervé; Auby, Antoine; Boulanger, Damien; Thouret, Valerie
2016-04-01
Since 1994, the IAGOS program (In-Service Aircraft for a Global Observing System http://www.iagos.org) and its predecessor MOZAIC has produced in-situ measurements of the atmospheric composition during more than 46000 commercial aircraft flights. In order to help analyzing these observations and further understanding the processes driving their evolution, we developed a modelling tool SOFT-IO quantifying their source/receptor link. We improved the methodology used by Stohl et al. (2003), based on the FLEXPART plume dispersion model, to simulate the contributions of anthropogenic and biomass burning emissions from the ECCAD database (http://eccad.aeris-data.fr) to the measured carbon monoxide mixing ratio along each IAGOS flight. Thanks to automated processes, contributions are simulated for the last 20 days before observation, separating individual contributions from the different source regions. The main goal is to supply add-value products to the IAGOS database showing pollutants geographical origin and emission type. Using this information, it may be possible to link trends in the atmospheric composition to changes in the transport pathways and to the evolution of emissions. This tool could be used for statistical validation as well as for inter-comparisons of emission inventories using large amounts of data, as Lagrangian models are able to bring the global scale emissions down to a smaller scale, where they can be directly compared to the in-situ observations from the IAGOS database.
Soil organic carbon stocks in Alaska estimated with spatial and pedon data
Bliss, Norman B.; Maursetter, J.
2010-01-01
Temperatures in high-latitude ecosystems are increasing faster than the average rate of global warming, which may lead to a positive feedback for climate change by increasing the respiration rates of soil organic C. If a positive feedback is confirmed, soil C will represent a source of greenhouse gases that is not currently considered in international protocols to regulate C emissions. We present new estimates of the stocks of soil organic C in Alaska, calculated by linking spatial and field data developed by the USDA NRCS. The spatial data are from the State Soil Geographic database (STATSGO), and the field and laboratory data are from the National Soil Characterization Database, also known as the pedon database. The new estimates range from 32 to 53 Pg of soil organic C for Alaska, formed by linking the spatial and field data using the attributes of Soil Taxonomy. For modelers, we recommend an estimation method based on taxonomic subgroups with interpolation for missing areas, which yields an estimate of 48 Pg. This is a substantial increase over a magnitude of 13 Pg estimated from only the STATSGO data as originally distributed in 1994, but the increase reflects different estimation methods and is not a measure of the change in C on the landscape. Pedon samples were collected between 1952 and 2002, so the results do not represent a single point in time. The linked databases provide an improved basis for modeling the impacts of climate change on net ecosystem exchange.
A meta-analysis and statistical modelling of nitrates in groundwater at the African scale
NASA Astrophysics Data System (ADS)
Ouedraogo, Issoufou; Vanclooster, Marnik
2016-06-01
Contamination of groundwater with nitrate poses a major health risk to millions of people around Africa. Assessing the space-time distribution of this contamination, as well as understanding the factors that explain this contamination, is important for managing sustainable drinking water at the regional scale. This study aims to assess the variables that contribute to nitrate pollution in groundwater at the African scale by statistical modelling. We compiled a literature database of nitrate concentration in groundwater (around 250 studies) and combined it with digital maps of physical attributes such as soil, geology, climate, hydrogeology, and anthropogenic data for statistical model development. The maximum, medium, and minimum observed nitrate concentrations were analysed. In total, 13 explanatory variables were screened to explain observed nitrate pollution in groundwater. For the mean nitrate concentration, four variables are retained in the statistical explanatory model: (1) depth to groundwater (shallow groundwater, typically < 50 m); (2) recharge rate; (3) aquifer type; and (4) population density. The first three variables represent intrinsic vulnerability of groundwater systems to pollution, while the latter variable is a proxy for anthropogenic pollution pressure. The model explains 65 % of the variation of mean nitrate contamination in groundwater at the African scale. Using the same proxy information, we could develop a statistical model for the maximum nitrate concentrations that explains 42 % of the nitrate variation. For the maximum concentrations, other environmental attributes such as soil type, slope, rainfall, climate class, and region type improve the prediction of maximum nitrate concentrations at the African scale. As to minimal nitrate concentrations, in the absence of normal distribution assumptions of the data set, we do not develop a statistical model for these data. The data-based statistical model presented here represents an important step towards developing tools that will allow us to accurately predict nitrate distribution at the African scale and thus may support groundwater monitoring and water management that aims to protect groundwater systems. Yet they should be further refined and validated when more detailed and harmonized data become available and/or combined with more conceptual descriptions of the fate of nutrients in the hydrosystem.
Thomas, Roger E; Lorenzetti, Diane L; Spragins, Wendy; Jackson, Dave; Williamson, Tyler
2011-07-01
To assess the reporting rates of serious adverse events attributable to yellow fever vaccination with 17D and 17DD strains as reported in pharmacovigilance databases, and assess reasons for differences in reporting rates. We searched 9 electronic databases for peer reviewed and grey literature (government reports, conferences), in all languages. Reference lists of key studies were also reviewed to identify additional studies. We identified 2,415 abstracts, of which 472 were selected for full text review. We identified 15 pharmacovigilance databases which reported adverse events attributed to yellow fever vaccination, of which 10 contributed data to this review with about 107,600,000 patients (allowing for overlapping time periods for the studies of the US VAERS database), and the data are very heavily weighted (94%) by the Brazilian database. The estimates of serious adverse events form three groups. The estimates for Australia were low at 0/210,656 for "severe neurological disease" and 1/210,656 for YEL-AVD, and also low for Brazil with 9 hypersensitivity events, 0.23 anaphylactic shock events, 0.84 neurologic syndrome events and 0.19 viscerotropic events cases/million doses. The five analyses of partly overlapping periods for the US VAERS database provide an estimate of 3.6/cases per million YEL-AND in one analysis and 7.8 in another, and 3.1 YEL-AVD in one analysis and 3.9 in another. The estimates for the UK used only the inclusive term of "serious adverse events" not further classified into YEL-And or YEL-AND and reported 34 "serious adverse events." The Swiss database used the term "serious adverse events" and reported 7 such events (including 4 "neurologic reactions") for a reporting rate of 25 "serious adverse events"/million doses. Reporting rates for serious adverse events following yellow fever vaccination are low. Differences in reporting rates may be due to differences in definitions, surveillance system organisation, methods of reporting cases, administration of YFV with other vaccines, incomplete information about denominators, time intervals for reporting events, the degree of passive reporting, access to diagnostic resources, and differences in time periods of reporting.
Tagare, Hemant D.; Jaffe, C. Carl; Duncan, James
1997-01-01
Abstract Information contained in medical images differs considerably from that residing in alphanumeric format. The difference can be attributed to four characteristics: (1) the semantics of medical knowledge extractable from images is imprecise; (2) image information contains form and spatial data, which are not expressible in conventional language; (3) a large part of image information is geometric; (4) diagnostic inferences derived from images rest on an incomplete, continuously evolving model of normality. This paper explores the differentiating characteristics of text versus images and their impact on design of a medical image database intended to allow content-based indexing and retrieval. One strategy for implementing medical image databases is presented, which employs object-oriented iconic queries, semantics by association with prototypes, and a generic schema. PMID:9147338
Identification and characterization of high methane-emitting abandoned oil and gas wells
Kang, Mary; Christian, Shanna; Celia, Michael A.; Mauzerall, Denise L.; Bill, Markus; Miller, Alana R.; Chen, Yuheng; Conrad, Mark E.; Darrah, Thomas H.; Jackson, Robert B.
2016-01-01
Recent measurements of methane emissions from abandoned oil/gas wells show that these wells can be a substantial source of methane to the atmosphere, particularly from a small proportion of high-emitting wells. However, identifying high emitters remains a challenge. We couple 163 well measurements of methane flow rates; ethane, propane, and n-butane concentrations; isotopes of methane; and noble gas concentrations from 88 wells in Pennsylvania with synthesized data from historical documents, field investigations, and state databases. Using our databases, we (i) improve estimates of the number of abandoned wells in Pennsylvania; (ii) characterize key attributes that accompany high emitters, including depth, type, plugging status, and coal area designation; and (iii) estimate attribute-specific and overall methane emissions from abandoned wells. High emitters are best predicted as unplugged gas wells and plugged/vented gas wells in coal areas and appear to be unrelated to the presence of underground natural gas storage areas or unconventional oil/gas production. Repeat measurements over 2 years show that flow rates of high emitters are sustained through time. Our attribute-based methane emission data and our comprehensive estimate of 470,000–750,000 abandoned wells in Pennsylvania result in estimated state-wide emissions of 0.04–0.07 Mt (1012 g) CH4 per year. This estimate represents 5–8% of annual anthropogenic methane emissions in Pennsylvania. Our methodology combining new field measurements with data mining of previously unavailable well attributes and numbers of wells can be used to improve methane emission estimates and prioritize cost-effective mitigation strategies for Pennsylvania and beyond. PMID:27849603
Relevant, irredundant feature selection and noisy example elimination.
Lashkia, George V; Anthony, Laurence
2004-04-01
In many real-world situations, the method for computing the desired output from a set of inputs is unknown. One strategy for solving these types of problems is to learn the input-output functionality from examples in a training set. However, in many situations it is difficult to know what information is relevant to the task at hand. Subsequently, researchers have investigated ways to deal with the so-called problem of consistency of attributes, i.e., attributes that can distinguish examples from different classes. In this paper, we first prove that the notion of relevance of attributes is directly related to the consistency of attributes, and show how relevant, irredundant attributes can be selected. We then compare different relevant attribute selection algorithms, and show the superiority of algorithms that select irredundant attributes over those that select relevant attributes. We also show that searching for an "optimal" subset of attributes, which is considered to be the main purpose of attribute selection, is not the best way to improve the accuracy of classifiers. Employing sets of relevant, irredundant attributes improves classification accuracy in many more cases. Finally, we propose a new method for selecting relevant examples, which is based on filtering the so-called pattern frequency domain. By identifying examples that are nontypical in the determination of relevant, irredundant attributes, irrelevant examples can be eliminated prior to the learning process. Empirical results using artificial and real databases show the effectiveness of the proposed method in selecting relevant examples leading to improved performance even on greatly reduced training sets.
A GIS-Enabled, Michigan-Specific, Hierarchical Groundwater Modeling and Visualization System
NASA Astrophysics Data System (ADS)
Liu, Q.; Li, S.; Mandle, R.; Simard, A.; Fisher, B.; Brown, E.; Ross, S.
2005-12-01
Efficient management of groundwater resources relies on a comprehensive database that represents the characteristics of the natural groundwater system as well as analysis and modeling tools to describe the impacts of decision alternatives. Many agencies in Michigan have spent several years compiling expensive and comprehensive surface water and groundwater inventories and other related spatial data that describe their respective areas of responsibility. However, most often this wealth of descriptive data has only been utilized for basic mapping purposes. The benefits from analyzing these data, using GIS analysis functions or externally developed analysis models or programs, has yet to be systematically realized. In this talk, we present a comprehensive software environment that allows Michigan groundwater resources managers and frontline professionals to make more effective use of the available data and improve their ability to manage and protect groundwater resources, address potential conflicts, design cleanup schemes, and prioritize investigation activities. In particular, we take advantage of the Interactive Ground Water (IGW) modeling system and convert it to a customized software environment specifically for analyzing, modeling, and visualizing the Michigan statewide groundwater database. The resulting Michigan IGW modeling system (IGW-M) is completely window-based, fully interactive, and seamlessly integrated with a GIS mapping engine. The system operates in real-time (on the fly) providing dynamic, hierarchical mapping, modeling, spatial analysis, and visualization. Specifically, IGW-M allows water resources and environmental professionals in Michigan to: * Access and utilize the extensive data from the statewide groundwater database, interactively manipulate GIS objects, and display and query the associated data and attributes; * Analyze and model the statewide groundwater database, interactively convert GIS objects into numerical model features, automatically extract data and attributes, and simulate unsteady groundwater flow and contaminant transport in response to water and land management decisions; * Visualize and map model simulations and predictions with data from the statewide groundwater database in a seamless interactive environment. IGW-M has the potential to significantly improve the productivity of Michigan groundwater management investigations. It changes the role of engineers and scientists in modeling and analyzing the statewide groundwater database from heavily physical to cognitive problem-solving and decision-making tasks. The seamless real-time integration, real-time visual interaction, and real-time processing capability allows a user to focus on critical management issues, conflicts, and constraints, to quickly and iteratively examine conceptual approximations, management and planning scenarios, and site characterization assumptions, to identify dominant processes, to evaluate data worth and sensitivity, and to guide further data-collection activities. We illustrate the power and effectiveness of the M-IGW modeling and visualization system with a real case study and a real-time, live demonstration.
Corwin, John; Silberschatz, Avi; Miller, Perry L; Marenco, Luis
2007-01-01
Data sparsity and schema evolution issues affecting clinical informatics and bioinformatics communities have led to the adoption of vertical or object-attribute-value-based database schemas to overcome limitations posed when using conventional relational database technology. This paper explores these issues and discusses why biomedical data are difficult to model using conventional relational techniques. The authors propose a solution to these obstacles based on a relational database engine using a sparse, column-store architecture. The authors provide benchmarks comparing the performance of queries and schema-modification operations using three different strategies: (1) the standard conventional relational design; (2) past approaches used by biomedical informatics researchers; and (3) their sparse, column-store architecture. The performance results show that their architecture is a promising technique for storing and processing many types of data that are not handled well by the other two semantic data models.
Intake of energy and nutrients; harmonization of Food Composition Databases.
Martinez-Victoria, Emilio; Martinez de Victoria, Ignacio; Martinez-Burgos, M Alba
2015-02-26
Food composition databases (FCDBs) provide detailed information about the nutritional composition of foods. The conversion of food consumption into nutrient intake need a Food composition database (FCDB) which lists the mean nutritional values for a given food portion. The limitations of FCDBs are sometimes little known by the users. Multicentre studies have raised several methodology challenges which allow to standardize nutritional assessments in different populations and geographical areas for food composition and nutrient intake. Differences between FCDBs include those attributed to technical matters, such as description of foods, calculation of energy and definition of nutrients, analytical methods, and principles for recipe calculation. Such differences need to be identified and eliminated before comparing data from different studies, especially when dietary data is related to a health outcome. There are ongoing efforts since 1984 to standardize FCDBs over the world (INFOODS, EPIC, EuroFIR, etc.). Food composition data can be gathered from different sources like private company analysis, universities, government laboratories and food industry. They can also be borrowed from scientific literature or even from the food labelling. There are different proposals to evaluate the quality of food composition data. For the development of a FCDB it is fundamental document in the most detailed way, each of the data values of the different components and nutrients of a food. The objective of AECOSAN (Agencia Española de Consumo Seguridad Alimentaria y Nutrición) and BEDCA (Base de Datos Española de Composición de Alimentos) association was the development and support of a reference FCDB in Spain according to the standards to be defined in Europe. BEDCA is currently the only FCDB developed in Spain with compiled and documented data following EuroFIR standards. Copyright AULA MEDICA EDICIONES 2015. Published by AULA MEDICA. All rights reserved.
ESIM: Edge Similarity for Screen Content Image Quality Assessment.
Ni, Zhangkai; Ma, Lin; Zeng, Huanqiang; Chen, Jing; Cai, Canhui; Ma, Kai-Kuang
2017-10-01
In this paper, an accurate full-reference image quality assessment (IQA) model developed for assessing screen content images (SCIs), called the edge similarity (ESIM), is proposed. It is inspired by the fact that the human visual system (HVS) is highly sensitive to edges that are often encountered in SCIs; therefore, essential edge features are extracted and exploited for conducting IQA for the SCIs. The key novelty of the proposed ESIM lies in the extraction and use of three salient edge features-i.e., edge contrast, edge width, and edge direction. The first two attributes are simultaneously generated from the input SCI based on a parametric edge model, while the last one is derived directly from the input SCI. The extraction of these three features will be performed for the reference SCI and the distorted SCI, individually. The degree of similarity measured for each above-mentioned edge attribute is then computed independently, followed by combining them together using our proposed edge-width pooling strategy to generate the final ESIM score. To conduct the performance evaluation of our proposed ESIM model, a new and the largest SCI database (denoted as SCID) is established in our work and made to the public for download. Our database contains 1800 distorted SCIs that are generated from 40 reference SCIs. For each SCI, nine distortion types are investigated, and five degradation levels are produced for each distortion type. Extensive simulation results have clearly shown that the proposed ESIM model is more consistent with the perception of the HVS on the evaluation of distorted SCIs than the multiple state-of-the-art IQA methods.
Interactive Querying Techniques for an Office Filing Facility.
ERIC Educational Resources Information Center
Morrissey, J. M.; And Others
1986-01-01
Proposes a "Model of Querying" for users of office filing facilities and discusses its motivation, aspects, attributes, and advantages. A review of current information systems and attempts to combine information retrieval, artificial intelligence, and database management techniques leads to conclusion that no resultant system is adequate…
76 FR 71511 - Proposed Information Collection; Comment Request; 2012 Company Organization Survey
Federal Register 2010, 2011, 2012, 2013, 2014
2011-11-18
..., multipurpose Business Register (BR) database. In particular, the COS supplies critical information on the... to identify all known United States business establishments and their parent companies. Further, the BR must accurately record basic business attributes needed to control sampling and enumeration. These...
78 FR 19190 - Proposed Information Collection; Comment Request; 2013 Company Organization Survey
Federal Register 2010, 2011, 2012, 2013, 2014
2013-03-29
..., multipurpose Business Register (BR) database. In particular, the COS supplies critical information on the... to identify all known United States business establishments and their parent companies. Further, the BR must accurately record basic business attributes needed to control sampling and enumeration. These...
A Web-Based GIS for Reporting Water Usage in the High Plains Underground Water Conservation District
NASA Astrophysics Data System (ADS)
Jia, M.; Deeds, N.; Winckler, M.
2012-12-01
The High Plains Underground Water Conservation District (HPWD) is the largest and oldest of the Texas water conservation districts, and oversees approximately 1.7 million irrigated acres. Recent rule changes have motivated HPWD to develop a more automated system to allow owners and operators to report well locations, meter locations, meter readings, the association between meters and wells, and contiguous acres. INTERA, Inc. has developed a web-based interactive system for HPWD water users to report water usage and for the district to better manage its water resources. The HPWD web management system utilizes state-of-the-art GIS techniques, including cloud-based Amazon EC2 virtual machine, ArcGIS Server, ArcSDE and ArcGIS Viewer for Flex, to support web-based water use management. The system enables users to navigate to their area of interest using a well-established base-map and perform a variety of operations and inquiries against their spatial features. The application currently has six components: user privilege management, property management, water meter registration, area registration, meter-well association and water use report. The system is composed of two main databases: spatial database and non-spatial database. With the help of Adobe Flex application at the front end and ArcGIS Server as the middle-ware, the spatial feature geometry and attributes update will be reflected immediately in the back end. As a result, property owners, along with the HPWD staff, collaborate together to weave the fabric of the spatial database. Interactions between the spatial and non-spatial databases are established by Windows Communication Foundation (WCF) services to record water-use report, user-property associations, owner-area associations, as well as meter-well associations. Mobile capabilities will be enabled in the near future for field workers to collect data and synchronize them to the spatial database. The entire solution is built on a highly scalable cloud server to dynamically allocate the computational resources so as to reduce the cost on security and hardware maintenance. In addition to the default capabilities provided by ESRI, customizations include 1) enabling interactions between spatial and non-spatial databases, 2) providing role-based feature editing, 3) dynamically filtering spatial features on the map based on user accounts and 4) comprehensive data validation.
Ambiguous data association and entangled attribute estimation
NASA Astrophysics Data System (ADS)
Trawick, David J.; Du Toit, Philip C.; Paffenroth, Randy C.; Norgard, Gregory J.
2012-05-01
This paper presents an approach to attribute estimation incorporating data association ambiguity. In modern tracking systems, time pressures often leave all but the most likely data association alternatives unexplored, possibly producing track inaccuracies. Numerica's Bayesian Network Tracking Database, a key part of its Tracker Adjunct Processor, captures and manages the data association ambiguity for further analysis and possible ambiguity reduction/resolution using subsequent data. Attributes are non-kinematic discrete sample space sensor data. They may be as distinctive as aircraft ID, or as broad as friend or foe. Attribute data may provide improvements to data association by a process known as Attribute Aided Tracking (AAT). Indeed, certain uniquely identifying attributes (e.g. aircraft ID), when continually reported, can be used to define data association (tracks are the collections of observations with the same ID). However, attribute data arriving infrequently, combined with erroneous choices from ambiguous data associations, can produce incorrect attribute and kinematic state estimation. Ambiguous data associations define the tracks that are entangled with each other. Attribute data observed on an entangled track then modify the attribute estimates on all tracks entangled with it. For example, if a red track and a blue track pass through a region of data association ambiguity, these tracks become entangled. Later red observations on one entangled track make the other track more blue, and reduce the data association ambiguity. Methods for this analysis have been derived and implemented for efficient forward filtering and forensic analysis.
Patient safety in dentistry - state of play as revealed by a national database of errors.
Thusu, S; Panesar, S; Bedi, R
2012-08-01
Modern dentistry has become increasingly invasive and sophisticated. Consequently the risk to the patient has increased. The aim of this study is to investigate the types of patient safety incidents (PSIs) that occur in dentistry and the accuracy of the National Patient Safety Agency (NPSA) database in identifying those attributed to dentistry. The database was analysed for all incidents of iatrogenic harm in the speciality of dentistry. A snapshot view using the timeframe January to December 2009 was used. The free text elements from the database were analysed thematically and reclassified according to the nature of the PSI. Descriptive statistics were provided. Two thousand and twelve incident reports were analysed and organised into ten categories. The commonest was due to clerical errors - 36%. Five areas of PSI were further analysed: injury (10%), medical emergency (6%), inhalation/ingestion (4%), adverse reaction (4%) and wrong site extraction (2%). There is generally low reporting of PSIs within the dental specialities. This may be attributed to the voluntary nature of reporting and the reluctance of dental practitioners to disclose incidences for fear of loss of earnings. A significant amount of iatrogenic harm occurs not during treatment but through controllable pre- and post-procedural checks. Incidences of iatrogenic harm to dental patients do occur but their reporting is not widely used. The use of a dental specific reporting system would aid in minimising iatrogenic harm and adhere to the Care Quality Commission (CQC) compliance monitoring system on essential standards of quality and safety in dental practices.
Blumenfeld, Olga O
2002-04-01
Recent advances in molecular biology and technology have provided evidence, at a molecular level, for long-known observations that the human genome is not unique but is characterized by individual sequence variation. At the present time, documentation of genetic variation occurring in a large number of genes is increasing exponentially. The characterization of alleles that encode a variety of blood group antigens has been particularly fruitful for transfusion medicine. Phenotypic variation, as identified by the serologic study of blood group variants, is required to identify the presence of a variant allele. Many of the other alleles currently recorded have been selected and identified on the basis of inherited disease traits. New approaches document single nucleotide polymorphisms that occur throughout the genome and best show how the DNA sequence varies in the human population. The primary data dealing with variant alleles or more general genomic variation are scattered throughout the scientific literature and only within the last few years has information begun to be organized into databases. This article provides guidance on how to access those databases online as a source of information about genetic variation for purposes of molecular, clinical, and diagnostic medicine, research, and teaching. The attributes of the sites are described. A more detailed view of the database dealing specifically with alleles of genes encoding the blood group antigens includes a brief preliminary analysis of the molecular basis for observed polymorphisms. Other online sites that may be particularly useful to the transfusion medicine readership as well as a brief historical account are also presented. Copyright 2002, Elsevier Science (USA). All rights reserved.
Implicit measures: A normative analysis and review.
De Houwer, Jan; Teige-Mocigemba, Sarah; Spruyt, Adriaan; Moors, Agnes
2009-05-01
Implicit measures can be defined as outcomes of measurement procedures that are caused in an automatic manner by psychological attributes. To establish that a measurement outcome is an implicit measure, one should examine (a) whether the outcome is causally produced by the psychological attribute it was designed to measure, (b) the nature of the processes by which the attribute causes the outcome, and (c) whether these processes operate automatically. This normative analysis provides a heuristic framework for organizing past and future research on implicit measures. The authors illustrate the heuristic function of their framework by using it to review past research on the 2 implicit measures that are currently most popular: effects in implicit association tests and affective priming tasks. (PsycINFO Database Record (c) 2009 APA, all rights reserved).
Troselj, Mario; Fanton, Davor
2005-01-01
The possibilities of creating a health care resources registry and its operating in Croatia as well as the importance of information in health system are described. At the Croatian Institute of Public Health, monitoring of human resources is performed through the national Health Workers Registry. It also covers basic data on all health units, bed capacities of health facilities included. The initiated health care computerization has urged the idea of forming one more database on physical resources, i.e. on registered medical devices and equipment, more complete. Linking these databases on health resources would produce a single Health Care Resources Registry. The concept views Health Care Resources Registry as part of the overall health information system with centralized information on the health system. The planned development of segments of a single health information system is based on the implementation of the accepted international standards and common network services. Network services that are based on verified Internet technologies are used within a safe, reliable and closed health computer network, which makes up the health intranet (WAN--Wide Area Network). The resource registry is a software solution based on the relational database that monitors history, thus permitting the data collected over a longer period to be analyzed. Such a solution assumes the existence of a directory service, which would replace the current independent software for the Health Workers Registry. In the Health Care Resources Registry, the basic data set encompasses data objects and attributes from the directory service. The directory service is compatible with the LDAP protocol (Lightweight Directory Access Protocol), providing services uniformly to the current records on human and physical resources. Through the storage of attributes defined according to the HL7 (Health Level Seven) standard, directory service is accessible to all applications of the health information system. Directory service does not follow the history of attribute changes, and is optimized for a large number of authorizing inquiries. With it, one follows the following objects and attributes: persons, groups of people (patients, physicians, other personnel), roles (right of access and administrator permissions), organizational units, unit locations, devices and services (according to the list of services and procedures). One can add to the Health Care Resource Registry such attributes as are nonessential for inclusion in the directory service, but are of public health value. Authentication, authorization and digital signature are done by means of Smart Cards, which are used as protective elements against access to system functions, and simultaneously as a physical medium for the storage of the official certificate with which documents are signed digitally. As FINA (state financial control agency) has completed a system for certificate issuance and verification, the option of official digital signature is also available as a computer network service. Any changes taking place in the directory service are transferred by XML messages to a separate part of the Registry that reads them and automatically modifies records in the relational database. Because data input and data changes are made in health units, this makes the data updated and directly connected with health working operations. This avoids all one-time data collection campaigns using form filling about the devices and equipment in the future. As it is very difficult to monitor from a central standpoint how accurate and update the information is, it is necessary to delegate the permissions and duties associated with making changes to the directory service. By this organizational setup, the time needed to ensure data quality control is reduced. In the case described, the Health Care Resource Registry becomes an indicator of change, acquiring certain characteristics of an analytical system. An analysis of topical data renders possible proactive action and makes more effective the planning and utilization of available resources. Providing answers on the current data quickly could also be important to solution-seeking in emergencies. The present proposal to establish the Registry is intended to facilitate the future process of planning and striking a balance between investments in human and physical resources. For health expenditure control, having reliable information related to the use and purchase of new medical technology is particularly important. World Health Organization and European Union have also emphasized the need to develop new indicators in this area.
Modeling first impressions from highly variable facial images
Vernon, Richard J. W.; Sutherland, Clare A. M.; Young, Andrew W.; Hartley, Tom
2014-01-01
First impressions of social traits, such as trustworthiness or dominance, are reliably perceived in faces, and despite their questionable validity they can have considerable real-world consequences. We sought to uncover the information driving such judgments, using an attribute-based approach. Attributes (physical facial features) were objectively measured from feature positions and colors in a database of highly variable “ambient” face photographs, and then used as input for a neural network to model factor dimensions (approachability, youthful-attractiveness, and dominance) thought to underlie social attributions. A linear model based on this approach was able to account for 58% of the variance in raters’ impressions of previously unseen faces, and factor-attribute correlations could be used to rank attributes by their importance to each factor. Reversing this process, neural networks were then used to predict facial attributes and corresponding image properties from specific combinations of factor scores. In this way, the factors driving social trait impressions could be visualized as a series of computer-generated cartoon face-like images, depicting how attributes change along each dimension. This study shows that despite enormous variation in ambient images of faces, a substantial proportion of the variance in first impressions can be accounted for through linear changes in objectively defined features. PMID:25071197
The StarView intelligent query mechanism
NASA Technical Reports Server (NTRS)
Semmel, R. D.; Silberberg, D. P.
1993-01-01
The StarView interface is being developed to facilitate the retrieval of scientific and engineering data produced by the Hubble Space Telescope. While predefined screens in the interface can be used to specify many common requests, ad hoc requests require a dynamic query formulation capability. Unfortunately, logical level knowledge is too sparse to support this capability. In particular, essential formulation knowledge is lost when the domain of interest is mapped to a set of database relation schemas. Thus, a system known as QUICK has been developed that uses conceptual design knowledge to facilitate query formulation. By heuristically determining strongly associated objects at the conceptual level, QUICK is able to formulate semantically reasonable queries in response to high-level requests that specify only attributes of interest. Moreover, by exploiting constraint knowledge in the conceptual design, QUICK assures that queries are formulated quickly and will execute efficiently.
NASA Astrophysics Data System (ADS)
Stewart, R.; Piburn, J.; Sorokine, A.; Myers, A.; Moehl, J.; White, D.
2015-07-01
The application of spatiotemporal (ST) analytics to integrated data from major sources such as the World Bank, United Nations, and dozens of others holds tremendous potential for shedding new light on the evolution of cultural, health, economic, and geopolitical landscapes on a global level. Realizing this potential first requires an ST data model that addresses challenges in properly merging data from multiple authors, with evolving ontological perspectives, semantical differences, and changing attributes, as well as content that is textual, numeric, categorical, and hierarchical. Equally challenging is the development of analytical and visualization approaches that provide a serious exploration of this integrated data while remaining accessible to practitioners with varied backgrounds. The WSTAMP project at Oak Ridge National Laboratory has yielded two major results in addressing these challenges: 1) development of the WSTAMP database, a significant advance in ST data modeling that integrates 10,000+ attributes covering over 200 nation states spanning over 50 years from over 30 major sources and 2) a novel online ST exploratory and analysis tool providing an array of modern statistical and visualization techniques for analyzing these data temporally, spatially, and spatiotemporally under a standard analytic workflow. We discuss the status of this work and report on major findings.
Nature of Blame in Patient Safety Incident Reports: Mixed Methods Analysis of a National Database.
Cooper, Jennifer; Edwards, Adrian; Williams, Huw; Sheikh, Aziz; Parry, Gareth; Hibbert, Peter; Butlin, Amy; Donaldson, Liam; Carson-Stevens, Andrew
2017-09-01
A culture of blame and fear of retribution are recognized barriers to reporting patient safety incidents. The extent of blame attribution in safety incident reports, which may reflect the underlying safety culture of health care systems, is unknown. This study set out to explore the nature of blame in family practice safety incident reports. We characterized a random sample of family practice patient safety incident reports from the England and Wales National Reporting and Learning System. Reports were analyzed according to prespecified classification systems to describe the incident type, contributory factors, outcomes, and severity of harm. We developed a taxonomy of blame attribution, and we then used descriptive statistical analyses to identify the proportions of blame types and to explore associations between incident characteristics and one type of blame. Health care professionals making family practice incident reports attributed blame to a person in 45% of cases (n = 975 of 2,148; 95% CI, 43%-47%). In 36% of cases, those who reported the incidents attributed fault to another person, whereas 2% of those reporting acknowledged personal responsibility. Blame was commonly associated with incidents where a complaint was anticipated. The high frequency of blame in these safety, incident reports may reflect a health care culture that leads to blame and retribution, rather than to identifying areas for learning and improvement, and a failure to appreciate the contribution of system factors in others' behavior. Successful improvement in patient safety through the analysis of incident reports is unlikely without achieving a blame-free culture. © 2017 Annals of Family Medicine, Inc.
Nature of Blame in Patient Safety Incident Reports: Mixed Methods Analysis of a National Database
Cooper, Jennifer; Edwards, Adrian; Williams, Huw; Sheikh, Aziz; Parry, Gareth; Hibbert, Peter; Butlin, Amy; Donaldson, Liam; Carson-Stevens, Andrew
2017-01-01
PURPOSE A culture of blame and fear of retribution are recognized barriers to reporting patient safety incidents. The extent of blame attribution in safety incident reports, which may reflect the underlying safety culture of health care systems, is unknown. This study set out to explore the nature of blame in family practice safety incident reports. METHODS We characterized a random sample of family practice patient safety incident reports from the England and Wales National Reporting and Learning System. Reports were analyzed according to prespecified classification systems to describe the incident type, contributory factors, outcomes, and severity of harm. We developed a taxonomy of blame attribution, and we then used descriptive statistical analyses to identify the proportions of blame types and to explore associations between incident characteristics and one type of blame. RESULTS Health care professionals making family practice incident reports attributed blame to a person in 45% of cases (n = 975 of 2,148; 95% CI, 43%–47%). In 36% of cases, those who reported the incidents attributed fault to another person, whereas 2% of those reporting acknowledged personal responsibility. Blame was commonly associated with incidents where a complaint was anticipated. CONCLUSIONS The high frequency of blame in these safety, incident reports may reflect a health care culture that leads to blame and retribution, rather than to identifying areas for learning and improvement, and a failure to appreciate the contribution of system factors in others’ behavior. Successful improvement in patient safety through the analysis of incident reports is unlikely without achieving a blame-free culture. PMID:28893816
NASA Astrophysics Data System (ADS)
Artana, K. B.; Pitana, T.; Dinariyana, D. P.; Ariana, M.; Kristianto, D.; Pratiwi, E.
2018-06-01
The aim of this research is to develop an algorithm and application that can perform real-time monitoring of the safety operation of offshore platforms and subsea gas pipelines as well as determine the need for ship inspection using data obtained from automatic identification system (AIS). The research also focuses on the integration of shipping database, AIS data, and others to develop a prototype for designing a real-time monitoring system of offshore platforms and pipelines. A simple concept is used in the development of this prototype, which is achieved by using an overlaying map that outlines the coordinates of the offshore platform and subsea gas pipeline with the ship's coordinates (longitude/latitude) as detected by AIS. Using such information, we can then build an early warning system (EWS) relayed through short message service (SMS), email, or other means when the ship enters the restricted and exclusion zone of platforms and pipelines. The ship inspection system is developed by combining several attributes. Then, decision analysis software is employed to prioritize the vessel's four attributes, including ship age, ship type, classification, and flag state. Results show that the EWS can increase the safety level of offshore platforms and pipelines, as well as the efficient use of patrol boats in monitoring the safety of the facilities. Meanwhile, ship inspection enables the port to prioritize the ship to be inspected in accordance with the priority ranking inspection score.
Instructor Quality Affecting Emergency Medical Technician (EMT) Preparedness: A LEADS Project
ERIC Educational Resources Information Center
Russ-Eft, Darlene F.; Dickison, Philip D.; Levine, Roger
2005-01-01
This represents one of a series of studies of the Longitudinal Emergency Medical Technician Attributes and Demographics Study (LEADS) being undertaken by the National Registry of Emergency Medical Technicians and the National Highway Traffic Safety Administration (NHTSA). This secondary analysis of the LEADS database, which provides a…
1. Expands the database for pesticide toxicity on native freshwater mussels. 2. Aids in determining any potential differences in toxic sensitivity of gravid female mussel attributed to age and laboratory holding times. 3. Aids in determining potential differences in juvenile ...
Nursing Faculty Perceptions on Teaching Critical Thinking
ERIC Educational Resources Information Center
Clark, Doris A.
2010-01-01
The perceptions of nursing faculty teaching critical thinking (CT) affective attributes and cognitive skills are described in this quantitative, descriptive study. The study sample consisted of nurse educators from the National League of Nursing database. The purpose of the study was to gain nursing faculty perception of which teaching strategies…
Expert Systems the Old Fashioned Way: Person to Person.
ERIC Educational Resources Information Center
McCleary, Hunter; Mayer, William J.
1988-01-01
Describes the services of Teltech, Inc., which mimic the desirable attributes of artificial intelligence and expert systems via a "database" of 5,000 experts in technical areas and interactive literature searches executed by staff. Advantages and shortcomings of the network are exemplified by sample searches. Several sample menus and…
The New Southern FIA Data Compilation System
V. Clark Baldwin; Larry Royer
2001-01-01
In general, the major national Forest Inventory and Analysis annual inventory emphasis has been on data-base design and not on data processing and calculation of various new attributes. Two key programming techniques required for efficient data processing are indexing and modularization. The Southern Research Station Compilation System utilizes modular and indexing...
Nicholson, Suzanne W.; Dicken, Connie L.; Horton, John D.; Foose, Michael P.; Mueller, Julia A.L.; Hon, Rudi
2006-01-01
The rapid growth in the use of Geographic Information Systems (GIS) has highlighted the need for regional and national scale digital geologic maps that have standardized information about geologic age and lithology. Such maps can be conveniently used to generate derivative maps for manifold special purposes such as mineral-resource assessment, metallogenic studies, tectonic studies, and environmental research. Although two digital geologic maps (Schruben and others, 1994; Reed and Bush, 2004) of the United States currently exist, their scales (1:2,500,000 and 1:5,000,000) are too general for many regional applications. Most states have digital geologic maps at scales of about 1:500,000, but the databases are not comparably structured and, thus, it is difficult to use the digital database for more than one state at a time. This report describes the result for a seven state region of an effort by the U.S. Geological Survey to produce a series of integrated and standardized state geologic map databases that cover the entire United States. In 1997, the United States Geological Survey's Mineral Resources Program initiated the National Surveys and Analysis (NSA) Project to develop national digital databases. One primary activity of this project was to compile a national digital geologic map database, utilizing state geologic maps, to support studies in the range of 1:250,000- to 1:1,000,000-scale. To accomplish this, state databases were prepared using a common standard for the database structure, fields, attribution, and data dictionaries. For Alaska and Hawaii new state maps are being prepared and the preliminary work for Alaska is being released as a series of 1:250,000 scale quadrangle reports. This document provides background information and documentation for the integrated geologic map databases of this report. This report is one of a series of such reports releasing preliminary standardized geologic map databases for the United States. The data products of the project consist of two main parts, the spatial databases and a set of supplemental tables relating to geologic map units. The datasets serve as a data resource to generate a variety of stratigraphic, age, and lithologic maps. This documentation is divided into four main sections: (1) description of the set of data files provided in this report, (2) specifications of the spatial databases, (3) specifications of the supplemental tables, and (4) an appendix containing the data dictionaries used to populate some fields of the spatial database and supplemental tables.
Schema for the LANL infrasound analysis tool, infrapy
DOE Office of Scientific and Technical Information (OSTI.GOV)
Dannemann, Fransiska Kate; Marcillo, Omar Eduardo
2017-04-14
The purpose of this document is to define the schema used for the operation of the infrasound analysis tool, infrapy. The tables described by this document extend the CSS3.0 or KB core schema to include information required for the operation of infrapy. This document is divided into three sections, the first being this introduction. Section two defines eight new, infrasonic data processing-specific database tables. Both internal (ORACLE) and external formats for the attributes are defined, along with a short description of each attribute. Section three of the document shows the relationships between the different tables by using entity-relationship diagrams.
Overview of NASA MSFC IEC Federated Engineering Collaboration Capability
NASA Technical Reports Server (NTRS)
Moushon, Brian; McDuffee, Patrick
2005-01-01
The MSFC IEC federated engineering framework is currently developing a single collaborative engineering framework across independent NASA centers. The federated approach allows NASA centers the ability to maintain diversity and uniqueness, while providing interoperability. These systems are integrated together in a federated framework without compromising individual center capabilities. MSFC IEC's Federation Framework will have a direct affect on how engineering data is managed across the Agency. The approach is directly attributed in response to the Columbia Accident Investigation Board (CAB) finding F7.4-11 which states the Space Shuttle Program has a wealth of data sucked away in multiple databases without a convenient way to integrate and use the data for management, engineering, or safety decisions. IEC s federated capability is further supported by OneNASA recommendation 6 that identifies the need to enhance cross-Agency collaboration by putting in place common engineering and collaborative tools and databases, processes, and knowledge-sharing structures. MSFC's IEC Federated Framework is loosely connected to other engineering applications that can provide users with the integration needed to achieve an Agency view of the entire product definition and development process, while allowing work to be distributed across NASA Centers and contractors. The IEC DDMS federation framework eliminates the need to develop a single, enterprise-wide data model, where the goal of having a common data model shared between NASA centers and contractors is very difficult to achieve.
NASA Astrophysics Data System (ADS)
García-Mayordomo, Julián; Martín-Banda, Raquel; Insua-Arévalo, Juan M.; Álvarez-Gómez, José A.; Martínez-Díaz, José J.; Cabral, João
2017-08-01
Active fault databases are a very powerful and useful tool in seismic hazard assessment, particularly when singular faults are considered seismogenic sources. Active fault databases are also a very relevant source of information for earth scientists, earthquake engineers and even teachers or journalists. Hence, active fault databases should be updated and thoroughly reviewed on a regular basis in order to keep a standard quality and uniformed criteria. Desirably, active fault databases should somehow indicate the quality of the geological data and, particularly, the reliability attributed to crucial fault-seismic parameters, such as maximum magnitude and recurrence interval. In this paper we explain how we tackled these issues during the process of updating and reviewing the Quaternary Active Fault Database of Iberia (QAFI) to its current version 3. We devote particular attention to describing the scheme devised for classifying the quality and representativeness of the geological evidence of Quaternary activity and the accuracy of the slip rate estimation in the database. Subsequently, we use this information as input for a straightforward rating of the level of reliability of maximum magnitude and recurrence interval fault seismic parameters. We conclude that QAFI v.3 is a much better database than version 2 either for proper use in seismic hazard applications or as an informative source for non-specialized users. However, we already envision new improvements for a future update.
Panagos, Panos; Ballabio, Cristiano; Yigini, Yusuf; Dunbar, Martha B
2013-01-01
Under the European Union Thematic Strategy for Soil Protection, the European Commission Directorate-General for the Environment and the European Environmental Agency (EEA) identified a decline in soil organic carbon and soil losses by erosion as priorities for the collection of policy relevant soil data at European scale. Moreover, the estimation of soil organic carbon content is of crucial importance for soil protection and for climate change mitigation strategies. Soil organic carbon is one of the attributes of the recently developed LUCAS soil database. The request for data on soil organic carbon and other soil attributes arose from an on-going debate about efforts to establish harmonized datasets for all EU countries with data on soil threats in order to support modeling activities and display variations in these soil conditions across Europe. In 2009, the European Commission's Joint Research Centre conducted the LUCAS soil survey, sampling ca. 20,000 points across 23 EU member states. This article describes the results obtained from analyzing the soil organic carbon data in the LUCAS soil database. The collected data were compared with the modeled European topsoil organic carbon content data developed at the JRC. The best fitted comparison was performed at NUTS2 level and showed underestimation of modeled data in southern Europe and overestimation in the new central eastern member states. There is a good correlation in certain regions for countries such as the United Kingdom, Slovenia, Italy, Ireland, and France. Here we assess the feasibility of producing comparable estimates of the soil organic carbon content at NUTS2 regional level for the European Union (EU27) and draw a comparison with existing modeled data. In addition to the data analysis, we suggest how the modeled data can be improved in future updates with better calibration of the model. Copyright © 2012 Elsevier B.V. All rights reserved.
Estimating Economic Burden of Cancer Deaths Attributable to Smoking in Iran.
Rezaei, Satar; Akbari Sari, Ali; Arab, Mohammad; Majdzadeh, Reza; Mohammadpoorasl, Asghar
2015-01-01
There is a broad consensus among health policy-makers that smoking has a significant impact on both heath system and society. The purpose of this study was to estimate the economic burden of major cancer deaths caused by smoking in Iran in 2012. Number of major cancer deaths due to smoking by sex and age groups in 2012 was obtained from GLOBCAN database. The life expectancy and retirement age were used to estimate years of potential life lost (YPLL) and cost of productive lost attributable to smoking, respectively. Data on prevalence of smoking, relative risk of smoking, life expectancy table, annual wage and employment rate were extracted from the various resources such as previous studies, WHO database and Iranian statistic centers. The data analysis was conducted by Excel software. Smoking was responsible for 4,623 cancer deaths, 80808 YPLL and $US 83,019,583 cost of productivity lost. Lung cancer accounts for largest proportion of total cancer deaths, YPLL and cost of productivity lost attributable to smoking. Males account for 86.6% of cancer deaths, 82.6% of YPLL and 85.3% of cost of productivity lost caused by smoking. Smoking places a high economic burden on health system and society as a whole. In addition, if no one had been smokers in Iran, approximately two out of ten cancer deaths could be prevented.
ComVisMD - compact visualization of multidimensional data: experimenting with cricket players data
NASA Astrophysics Data System (ADS)
Dandin, Shridhar B.; Ducassé, Mireille
2018-03-01
Database information is multidimensional and often displayed in tabular format (row/column display). Presented in aggregated form, multidimensional data can be used to analyze the records or objects. Online Analytical database Processing (OLAP) proposes mechanisms to display multidimensional data in aggregated forms. A choropleth map is a thematic map in which areas are colored in proportion to the measurement of a statistical variable being displayed, such as population density. They are used mostly for compact graphical representation of geographical information. We propose a system, ComVisMD inspired by choropleth map and the OLAP cube to visualize multidimensional data in a compact way. ComVisMD displays multidimensional data like OLAP Cube, where we are mapping an attribute a (first dimension, e.g. year started playing cricket) in vertical direction, object coloring based on b (second dimension, e.g. batting average), mapping varying-size circles based on attribute c (third dimension, e.g. highest score), mapping numbers based on attribute d (fourth dimension, e.g. matches played). We illustrate our approach on cricket players data, namely on two tables Country and Player. They have a large number of rows and columns: 246 rows and 17 columns for players of one country. ComVisMD’s visualization reduces the size of the tabular display by a factor of about 4, allowing users to grasp more information at a time than the bare table display.
Magee, Glenn; Strauss, Marcie E; Thomas, Sheila M; Brown, Harold; Baumer, Dorothy; Broderick, Kelly C
2015-11-01
The recent epidemiologic changes of Clostridium difficile-associated diarrhea (CDAD) have resulted in substantial economic burden to U.S. acute care hospitals. Past studies evaluating CDAD-attributable costs have been geographically and demographically limited. Here, we describe CDAD-attributable burden in inpatients, overall, and in vulnerable subpopulations from the Premier hospital database, a large, diverse cohort with a wide range of high-risk subgroups. Discharges from the Premier database were retrospectively analyzed to assess length of stay (LOS), total inpatient costs, readmission, and inpatient mortality. Patients with CDAD had significantly worse outcomes than matched controls in terms of total LOS, rates of intensive care unit (ICU) admission, and inpatient mortality. After adjustment for risk factors, patients with CDAD had increased odds of inpatient mortality, total and ICU LOS, costs, and odds of 30-, 60- and 90-day all-cause readmission versus non-CDAD patients. CDAD-attributable costs were higher in all studied vulnerable subpopulations, which also had increased odds of 30-, 60- and 90-day all-cause readmission than those without CDAD. Given the significant economic impact CDAD has on hospitals, prevention of initial episodes and targeted therapy to prevent recurrences in vulnerable patients are essential to decrease the overall burden to hospitals. Copyright © 2015 Association for Professionals in Infection Control and Epidemiology, Inc. Published by Elsevier Inc. All rights reserved.
Nanocubes for real-time exploration of spatiotemporal datasets.
Lins, Lauro; Klosowski, James T; Scheidegger, Carlos
2013-12-01
Consider real-time exploration of large multidimensional spatiotemporal datasets with billions of entries, each defined by a location, a time, and other attributes. Are certain attributes correlated spatially or temporally? Are there trends or outliers in the data? Answering these questions requires aggregation over arbitrary regions of the domain and attributes of the data. Many relational databases implement the well-known data cube aggregation operation, which in a sense precomputes every possible aggregate query over the database. Data cubes are sometimes assumed to take a prohibitively large amount of space, and to consequently require disk storage. In contrast, we show how to construct a data cube that fits in a modern laptop's main memory, even for billions of entries; we call this data structure a nanocube. We present algorithms to compute and query a nanocube, and show how it can be used to generate well-known visual encodings such as heatmaps, histograms, and parallel coordinate plots. When compared to exact visualizations created by scanning an entire dataset, nanocube plots have bounded screen error across a variety of scales, thanks to a hierarchical structure in space and time. We demonstrate the effectiveness of our technique on a variety of real-world datasets, and present memory, timing, and network bandwidth measurements. We find that the timings for the queries in our examples are dominated by network and user-interaction latencies.
Chen, You-Shyang; Cheng, Ching-Hsue; Lai, Chien-Jung; Hsu, Cheng-Yi; Syu, Han-Jhou
2012-02-01
Identifying patients in a Target Customer Segment (TCS) is important to determine the demand for, and to appropriately allocate resources for, health care services. The purpose of this study is to propose a two-stage clustering-classification model through (1) initially integrating the RFM attribute and K-means algorithm for clustering the TCS patients and (2) then integrating the global discretization method and the rough set theory for classifying hospitalized departments and optimizing health care services. To assess the performance of the proposed model, a dataset was used from a representative hospital (termed Hospital-A) that was extracted from a database from an empirical study in Taiwan comprised of 183,947 samples that were characterized by 44 attributes during 2008. The proposed model was compared with three techniques, Decision Tree, Naive Bayes, and Multilayer Perceptron, and the empirical results showed significant promise of its accuracy. The generated knowledge-based rules provide useful information to maximize resource utilization and support the development of a strategy for decision-making in hospitals. From the findings, 75 patients in the TCS, three hospital departments, and specific diagnostic items were discovered in the data for Hospital-A. A potential determinant for gender differences was found, and the age attribute was not significant to the hospital departments. Copyright © 2011 Elsevier Ltd. All rights reserved.
Connectedness in the context of patient-provider relationships: a concept analysis.
Phillips-Salimi, Celeste R; Haase, Joan E; Kooken, Wendy Carter
2012-01-01
This paper is a report of an analysis of the concept of connectedness. Previous attempts to conceptualize patient-provider relationships were limited in explaining how such relationships are fostered and maintained, and how they influence patient outcomes. Connectedness is a concept that may provide insights into the advantages of patient-provider relationships; however, the usefulness of this concept in health care is limited by its conceptual ambiguity. Although connectedness is widely used to describe other social relationships, little consistency exists among its definitions and measures. Sources identified through CINAHL, OVID, PubMed and PsychINFO databases and references lists of selected articles between 1983 and 2010. A hybrid concept analysis approach was used, involving a combination of traditional concept analysis strategies that included: describing historical conceptualizations, identifying attributes, critiquing existing definitions, examining boundaries and identifying antecedents and consequences. Using five distinct historical perspectives, seven attributes of connectedness were identified: intimacy, sense of belonging, caring, empathy, respect, trust and reciprocity. A broad definition of connectedness, which can be used in the context of patient-provider relationships, was developed. A preliminary theoretical framework of connectedness was derived from the identified antecedents, attributes and consequences. Research efforts to advance the concept of connectedness in patient-provider relationships have been hampered by a lack of conceptual clarity. This concept analysis offers a clearer understanding of connectedness, provides recommendations for future research and suggests practice implications. © 2011 Blackwell Publishing Ltd.
A proposal of fuzzy connective with learning function and its application to fuzzy retrieval system
NASA Technical Reports Server (NTRS)
Hayashi, Isao; Naito, Eiichi; Ozawa, Jun; Wakami, Noboru
1993-01-01
A new fuzzy connective and a structure of network constructed by fuzzy connectives are proposed to overcome a drawback of conventional fuzzy retrieval systems. This network represents a retrieval query and the fuzzy connectives in networks have a learning function to adjust its parameters by data from a database and outputs of a user. The fuzzy retrieval systems employing this network are also constructed. Users can retrieve results even with a query whose attributes do not exist in a database schema and can get satisfactory results for variety of thinkings by learning function.
Hydroacoustic forcing function modeling using DNS database
NASA Technical Reports Server (NTRS)
Zawadzki, I.; Gershfield, J. L.; Na, Y.; Wang, M.
1996-01-01
A wall pressure frequency spectrum model (Blake 1971 ) has been evaluated using databases from Direct Numerical Simulations (DNS) of a turbulent boundary layer (Na & Moin 1996). Good agreement is found for moderate to strong adverse pressure gradient flows in the absence of separation. In the separated flow region, the model underpredicts the directly calculated spectra by an order of magnitude. The discrepancy is attributed to the violation of the model assumptions in that part of the flow domain. DNS computed coherence length scales and the normalized wall pressure cross-spectra are compared with experimental data. The DNS results are consistent with experimental observations.
Identity and the body: Trajectories of body esteem from adolescence to emerging adulthood.
Nelson, Sarah C; Kling, Johanna; Wängqvist, Maria; Frisén, Ann; Syed, Moin
2018-06-01
Although Erikson (1968) originally conceptualized identity development as a process of becoming at home in one's body, little work has been done linking identity development and research on the body. This study examines how trajectories of the development of body esteem over time are related to young people's sense of identity and psychological functioning in a longitudinal sample from age 10 to 24 (N = 967). Using group-based trajectory modeling, three cubic subgroups were determined for each of the three types of body esteem: appearance, weight, and attribution. These groups demonstrated significant variations in the ways in which body esteem changes over time. These trajectory groups importantly differed in relationship to gender, identity coherence, identity confusion, and psychological functioning. Results are discussed in terms of the need to use a sociocultural perspective to explore the body's relation to identity development and the importance of disaggregating mean-level findings using person-centered approaches to determine high-risk groups. (PsycINFO Database Record (c) 2018 APA, all rights reserved).
Geographic information system as country-level development and monitoring tool, Senegal example
Moore, Donald G.; Howard, Stephen M.; ,
1990-01-01
Geographic information systems (GIS) allow an investigator the capability to merge and analyze numerous types of country-level resource data. Hypothetical resource analysis applications in Senegal were conducted to illustrate the utility of a GIS for development planning and resource monitoring. Map and attribute data for soils, vegetation, population, infrastructure, and administrative units were merged to form a database within a GIS. Several models were implemented using a GIS to: analyze development potential for sustainable dryland agriculture; prioritize where agricultural development should occur based upon a regional food budget; and monitor dynamic events with remote sensing. The steps for implementing a GIS analysis are described and illustrated, and the use of a GIS for conducting an economic analysis is outlined. Using a GIS for analysis and display of results opens new methods of communication between resource scientists and decision makers. Analyses yielding country-wide map output and detailed statistical data for each level of administration provide the advantage of a single system that can serve a variety of users.
Requirements for benchmarking personal image retrieval systems
NASA Astrophysics Data System (ADS)
Bouguet, Jean-Yves; Dulong, Carole; Kozintsev, Igor; Wu, Yi
2006-01-01
It is now common to have accumulated tens of thousands of personal ictures. Efficient access to that many pictures can only be done with a robust image retrieval system. This application is of high interest to Intel processor architects. It is highly compute intensive, and could motivate end users to upgrade their personal computers to the next generations of processors. A key question is how to assess the robustness of a personal image retrieval system. Personal image databases are very different from digital libraries that have been used by many Content Based Image Retrieval Systems.1 For example a personal image database has a lot of pictures of people, but a small set of different people typically family, relatives, and friends. Pictures are taken in a limited set of places like home, work, school, and vacation destination. The most frequent queries are searched for people, and for places. These attributes, and many others affect how a personal image retrieval system should be benchmarked, and benchmarks need to be different from existing ones based on art images, or medical images for examples. The attributes of the data set do not change the list of components needed for the benchmarking of such systems as specified in2: - data sets - query tasks - ground truth - evaluation measures - benchmarking events. This paper proposed a way to build these components to be representative of personal image databases, and of the corresponding usage models.
Shantakumar, Sumitra; Nordstrom, Beth L; Hall, Susan A; Djousse, Luc; van Herk-Sukel, Myrthe P P; Fraeman, Kathy H; Gagnon, David R; Chagin, Karen; Nelson, Jeanenne J
2017-04-20
Pazopanib received US Food and Drug Administration approval in 2009 for advanced renal cell carcinoma. During clinical development, liver chemistry abnormalities and adverse hepatic events were observed, leading to a boxed warning for hepatotoxicity and detailed label prescriber guidelines for liver monitoring. As part of postapproval regulatory commitments, a cohort study was conducted to assess prescriber compliance with liver monitoring guidelines. Over a 4-year period, a distributed network approach was used across 3 databases: US Veterans Affairs Healthcare System, a US outpatient oncology community practice database, and the Dutch PHARMO Database Network. Measures of prescriber compliance were designed using the original pazopanib label guidelines for liver monitoring. Results from the VA (n = 288) and oncology databases (n = 283) indicate that prescriber liver chemistry monitoring was less than 100%: 73% to 74% compliance with baseline testing and 37% to 39% compliance with testing every 4 weeks. Compliance was highest near drug initiation and decreased over time. Among patients who should have had weekly testing, the compliance was 56% in both databases. The more serious elevations examined, including combinations of liver enzyme elevations meeting the laboratory definition of Hy's law were infrequent but always led to appropriate discontinuation of pazopanib. Only 4 patients were identified for analysis in the Dutch database; none had recorded baseline testing. In this population-based study, prescriber compliance was reasonable near pazopanib initiation but low during subsequent weeks of treatment. This study provides information from real-world community practice settings and offers feedback to regulators on the effectiveness of label monitoring guidelines.This is an open-access article distributed under the terms of the Creative Commons Attribution-Non Commercial-No Derivatives License 4.0 (CCBY-NC-ND), where it is permissible to download and share the work provided it is properly cited. The work cannot be changed in any way or used commercially without permission from the journal.
Developmental precursors of young school-age children's hostile attribution bias.
Choe, Daniel Ewon; Lane, Jonathan D; Grabell, Adam S; Olson, Sheryl L
2013-12-01
This prospective longitudinal study provides evidence of preschool-age precursors of hostile attribution bias in young school-age children, a topic that has received little empirical attention. We examined multiple risk domains, including laboratory and observational assessments of children's social-cognition, general cognitive functioning, effortful control, and peer aggression. Preschoolers (N = 231) with a more advanced theory-of-mind, better emotion understanding, and higher IQ made fewer hostile attributions of intent in the early school years. Further exploration of these significant predictors revealed that only certain components of these capacities (i.e., nonstereotypical emotion understanding, false-belief explanation, and verbal IQ) were robust predictors of a hostile attribution bias in young school-age children and were especially strong predictors among children with more advanced effortful control. These relations were prospective in nature-the effects of preschool variables persisted after accounting for similar variables at school age. We conclude by discussing the implications of our findings for future research and prevention. PsycINFO Database Record (c) 2013 APA, all rights reserved.
Porter, P Steven; Rao, S Trivikrama; Zurbenko, Igor G; Dunker, Alan M; Wolff, George T
2001-02-01
Assessment of regulatory programs aimed at improving ambient O 3 air quality is of considerable interest to the scientific community and to policymakers. Trend detection, the identification of statistically significant long-term changes, and attribution, linking change to specific clima-tological and anthropogenic forcings, are instrumental to this assessment. Detection and attribution are difficult because changes in pollutant concentrations of interest to policymakers may be much smaller than natural variations due to weather and climate. In addition, there are considerable differences in reported trends seemingly based on similar statistical methods and databases. Differences arise from the variety of techniques used to reduce nontrend variation in time series, including mitigating the effects of meteorology and the variety of metrics used to track changes. In this paper, we review the trend assessment techniques being used in the air pollution field and discuss their strengths and limitations in discerning and attributing changes in O 3 to emission control policies.
Engaging and supporting fathers to promote breast feeding: a concept analysis.
Sherriff, Nigel; Hall, Valerie; Panton, Christina
2014-06-01
Empirical evidence demonstrates that fathers have a strong influence on a mother's decision to initiate and continue breast feeding. However, no clear delineation of what behaviours and attributes constitute father support or differentiate it from other kinds of support is provided in the current literature. The purpose of this study was to analyse the concept of 'father support' in relation to maternity services and broader health settings, thereby clarifying meaning to enable comprehension and application in practice, education, and research. A concept analysis combining the evolutionary model of concept development with the inter-related theoretical, fieldwork and analytical phases of the hybrid model of concept development. Children's Centres in East and West Sussex in Southern England. Repeated qualitative research over two phases with 16 parents of breast fed infants through seven focus groups and five telephone interviews. CINAHL, PsycINFO, AMED, MEDLINE, OVID and EMBASE databases were searched for articles published in English between 1999 and 2013 using the keywords breast feeding, father, and support. Seven same-sex focus groups and five individual interviews were also conducted over two research phases with the parents of breast fed infants to expand and exemplify, and then validate the analysis of the literature search. Five main attributes of father support in relation to breast feeding were identified: (1) knowledge about breast feeding; (2) positive attitude to breast feeding; (3) involvement in the decision-making process; (4) practical support; and (5) emotional support. Multiple antecedents and consequences to these attributes were also identified. This study has contributed to clarifying the meaning of father support in relation to breast feeding and provides an important starting point for the development of a theoretical and practical model of optimal breast feeding that takes into account father support. Identification of attributes, antecedents, and consequences of father support may assist practitioners to reflect on current working practices and service delivery models, and offer important educational opportunities for the training of student midwives and other health professionals. Copyright © 2013 The Authors. Published by Elsevier Ltd.. All rights reserved.
Using Crowdsourced Trajectories for Automated OSM Data Entry Approach
Basiri, Anahid; Amirian, Pouria; Mooney, Peter
2016-01-01
The concept of crowdsourcing is nowadays extensively used to refer to the collection of data and the generation of information by large groups of users/contributors. OpenStreetMap (OSM) is a very successful example of a crowd-sourced geospatial data project. Unfortunately, it is often the case that OSM contributor inputs (including geometry and attribute data inserts, deletions and updates) have been found to be inaccurate, incomplete, inconsistent or vague. This is due to several reasons which include: (1) many contributors with little experience or training in mapping and Geographic Information Systems (GIS); (2) not enough contributors familiar with the areas being mapped; (3) contributors having different interpretations of the attributes (tags) for specific features; (4) different levels of enthusiasm between mappers resulting in different number of tags for similar features and (5) the user-friendliness of the online user-interface where the underlying map can be viewed and edited. This paper suggests an automatic mechanism, which uses raw spatial data (trajectories of movements contributed by contributors to OSM) to minimise the uncertainty and impact of the above-mentioned issues. This approach takes the raw trajectory datasets as input and analyses them using data mining techniques. In addition, we extract some patterns and rules about the geometry and attributes of the recognised features for the purpose of insertion or editing of features in the OSM database. The underlying idea is that certain characteristics of user trajectories are directly linked to the geometry and the attributes of geographic features. Using these rules successfully results in the generation of new features with higher spatial quality which are subsequently automatically inserted into the OSM database. PMID:27649192
Metadata mapping and reuse in caBIG.
Kunz, Isaac; Lin, Ming-Chin; Frey, Lewis
2009-02-05
This paper proposes that interoperability across biomedical databases can be improved by utilizing a repository of Common Data Elements (CDEs), UML model class-attributes and simple lexical algorithms to facilitate the building domain models. This is examined in the context of an existing system, the National Cancer Institute (NCI)'s cancer Biomedical Informatics Grid (caBIG). The goal is to demonstrate the deployment of open source tools that can be used to effectively map models and enable the reuse of existing information objects and CDEs in the development of new models for translational research applications. This effort is intended to help developers reuse appropriate CDEs to enable interoperability of their systems when developing within the caBIG framework or other frameworks that use metadata repositories. The Dice (di-grams) and Dynamic algorithms are compared and both algorithms have similar performance matching UML model class-attributes to CDE class object-property pairs. With algorithms used, the baselines for automatically finding the matches are reasonable for the data models examined. It suggests that automatic mapping of UML models and CDEs is feasible within the caBIG framework and potentially any framework that uses a metadata repository. This work opens up the possibility of using mapping algorithms to reduce cost and time required to map local data models to a reference data model such as those used within caBIG. This effort contributes to facilitating the development of interoperable systems within caBIG as well as other metadata frameworks. Such efforts are critical to address the need to develop systems to handle enormous amounts of diverse data that can be leveraged from new biomedical methodologies.
[Design of computerised database for clinical and basic management of uveal melanoma].
Bande Rodríguez, M F; Santiago Varela, M; Blanco Teijeiro, M J; Mera Yañez, P; Pardo Perez, M; Capeans Tome, C; Piñeiro Ces, A
2012-09-01
The uveal melanoma is the most common primary intraocular tumour in adults. The objective of this work is to show how a computerised database has been formed with specific applications, for clinical and research use, to an extensive group of patients diagnosed with uveal melanoma. For the design of the database a selection of categories, attributes and values was created based on the classifications and parameters given by various authors of articles which have had great relevance in the field of uveal melanoma in recent years. The database has over 250 patient entries with specific information on their clinical history, diagnosis, treatment and progress. It enables us to search any parameter of the entry and make quick and simple statistical studies of them. The database models have been transformed into a basic tool for clinical practice, as they are an efficient way of storing, compiling and selective searching of information. When creating a database it is very important to define a common strategy and the use of a standard language. Copyright © 2011 Sociedad Española de Oftalmología. Published by Elsevier Espana. All rights reserved.
Real Time Monitor of Grid job executions
NASA Astrophysics Data System (ADS)
Colling, D. J.; Martyniak, J.; McGough, A. S.; Křenek, A.; Sitera, J.; Mulač, M.; Dvořák, F.
2010-04-01
In this paper we describe the architecture and operation of the Real Time Monitor (RTM), developed by the Grid team in the HEP group at Imperial College London. This is arguably the most popular dissemination tool within the EGEE [1] Grid. Having been used, on many occasions including GridFest and LHC inauguration events held at CERN in October 2008. The RTM gathers information from EGEE sites hosting Logging and Bookkeeping (LB) services. Information is cached locally at a dedicated server at Imperial College London and made available for clients to use in near real time. The system consists of three main components: the RTM server, enquirer and an apache Web Server which is queried by clients. The RTM server queries the LB servers at fixed time intervals, collecting job related information and storing this in a local database. Job related data includes not only job state (i.e. Scheduled, Waiting, Running or Done) along with timing information but also other attributes such as Virtual Organization and Computing Element (CE) queue - if known. The job data stored in the RTM database is read by the enquirer every minute and converted to an XML format which is stored on a Web Server. This decouples the RTM server database from the client removing the bottleneck problem caused by many clients simultaneously accessing the database. This information can be visualized through either a 2D or 3D Java based client with live job data either being overlaid on to a 2 dimensional map of the world or rendered in 3 dimensions over a globe map using OpenGL.
DB Dehydrogenase: an online integrated structural database on enzyme dehydrogenase.
Nandy, Suman Kumar; Bhuyan, Rajabrata; Seal, Alpana
2012-01-01
Dehydrogenase enzymes are almost inevitable for metabolic processes. Shortage or malfunctioning of dehydrogenases often leads to several acute diseases like cancers, retinal diseases, diabetes mellitus, Alzheimer, hepatitis B & C etc. With advancement in modern-day research, huge amount of sequential, structural and functional data are generated everyday and widens the gap between structural attributes and its functional understanding. DB Dehydrogenase is an effort to relate the functionalities of dehydrogenase with its structures. It is a completely web-based structural database, covering almost all dehydrogenases [~150 enzyme classes, ~1200 entries from ~160 organisms] whose structures are known. It is created by extracting and integrating various online resources to provide the true and reliable data and implemented by MySQL relational database through user friendly web interfaces using CGI Perl. Flexible search options are there for data extraction and exploration. To summarize, sequence, structure, function of all dehydrogenases in one place along with the necessary option of cross-referencing; this database will be utile for researchers to carry out further work in this field. The database is available for free at http://www.bifku.in/DBD/
Measuring "Equity" and "Equitability" in School Effectiveness Research
ERIC Educational Resources Information Center
Kelly, Anthony
2012-01-01
This paper introduces a Gini-type index for measuring "attainment equity" in schools; that is to say, how far a school (or group of schools) is from having a "fair" proportion of its examination success attributable to a fair proportion of its student population. Using data from the National Pupil Database, the Index is applied…
An Empirical Review of Internet Addiction Outcome Studies in China
ERIC Educational Resources Information Center
Liu, Chennan; Liao, Minli; Smith, Douglas C.
2012-01-01
Objectives: The authors systematically reviewed the outcomes and methodological quality of 24 Internet addiction (IA) treatment outcome studies in China. Method: The authors used 15 attributes from the quality of evidence scores to evaluate 24 outcome studies. These studies came from both English and Chinese academic databases from 2000 to 2010.…
In a previously described study, only 15% of the bacterial strains isolated from a water distribution system (WDS) grown on R2A agar were identifiable using fatty acid methyl esthers (FAME) profiling. The lack of success was attributed to the use of fatty acid databases of bacter...
Nuclear Forensic Inferences Using Iterative Multidimensional Statistics
DOE Office of Scientific and Technical Information (OSTI.GOV)
Robel, M; Kristo, M J; Heller, M A
2009-06-09
Nuclear forensics involves the analysis of interdicted nuclear material for specific material characteristics (referred to as 'signatures') that imply specific geographical locations, production processes, culprit intentions, etc. Predictive signatures rely on expert knowledge of physics, chemistry, and engineering to develop inferences from these material characteristics. Comparative signatures, on the other hand, rely on comparison of the material characteristics of the interdicted sample (the 'questioned sample' in FBI parlance) with those of a set of known samples. In the ideal case, the set of known samples would be a comprehensive nuclear forensics database, a database which does not currently exist. Inmore » fact, our ability to analyze interdicted samples and produce an extensive list of precise materials characteristics far exceeds our ability to interpret the results. Therefore, as we seek to develop the extensive databases necessary for nuclear forensics, we must also develop the methods necessary to produce the necessary inferences from comparison of our analytical results with these large, multidimensional sets of data. In the work reported here, we used a large, multidimensional dataset of results from quality control analyses of uranium ore concentrate (UOC, sometimes called 'yellowcake'). We have found that traditional multidimensional techniques, such as principal components analysis (PCA), are especially useful for understanding such datasets and drawing relevant conclusions. In particular, we have developed an iterative partial least squares-discriminant analysis (PLS-DA) procedure that has proven especially adept at identifying the production location of unknown UOC samples. By removing classes which fell far outside the initial decision boundary, and then rebuilding the PLS-DA model, we have consistently produced better and more definitive attributions than with a single pass classification approach. Performance of the iterative PLS-DA method compared favorably to that of classification and regression tree (CART) and k nearest neighbor (KNN) algorithms, with the best combination of accuracy and robustness, as tested by classifying samples measured independently in our laboratories against the vendor QC based reference set.« less
A proposed group management scheme for XTP multicast
NASA Technical Reports Server (NTRS)
Dempsey, Bert J.; Weaver, Alfred C.
1990-01-01
The purpose of a group management scheme is to enable its associated transfer layer protocol to be responsive to user determined reliability requirements for multicasting. Group management (GM) must assist the client process in coordinating multicast group membership, allow the user to express the subset of the multicast group that a particular multicast distribution must reach in order to be successful (reliable), and provide the transfer layer protocol with the group membership information necessary to guarantee delivery to this subset. GM provides services and mechanisms that respond to the need of the client process or process level management protocols to coordinate, modify, and determine attributes of the multicast group, especially membership. XTP GM provides a link between process groups and their multicast groups by maintaining a group membership database that identifies members in a name space understood by the underlying transfer layer protocol. Other attributes of the multicast group useful to both the client process and the data transfer protocol may be stored in the database. Examples include the relative dispersion, most recent update, and default delivery parameters of a group.
Zhou, Zhiqing E; Yan, Yu; Che, Xin Xuan; Meier, Laurenz L
2015-01-01
Although previous studies have linked workplace incivility with various negative outcomes, they mainly focused on the long-term effects of chronic exposure to workplace incivility, whereas targets' short-term reactions to incivility episodes have been largely neglected. Using a daily diary design, the current study examined effects of daily workplace incivility on end-of-work negative affect and explored potential individual and organizational moderators. Data collected from 76 full-time employees across 10 consecutive working days revealed that daily workplace incivility positively predicted end-of-work negative affect while controlling for before-work negative affect. Further, the relationship was stronger for people with low emotional stability, high hostile attribution bias, external locus of control, and people experiencing low chronic workload and more chronic organizational constraints, as compared with people with high emotional stability, low hostile attribution bias, internal locus of control, and people experiencing high chronic workload and fewer chronic organizational constraints, respectively. (PsycINFO Database Record (c) 2014 APA, all rights reserved). PsycINFO Database Record (c) 2014 APA, all rights reserved.
ProteoLens: a visual analytic tool for multi-scale database-driven biological network data mining.
Huan, Tianxiao; Sivachenko, Andrey Y; Harrison, Scott H; Chen, Jake Y
2008-08-12
New systems biology studies require researchers to understand how interplay among myriads of biomolecular entities is orchestrated in order to achieve high-level cellular and physiological functions. Many software tools have been developed in the past decade to help researchers visually navigate large networks of biomolecular interactions with built-in template-based query capabilities. To further advance researchers' ability to interrogate global physiological states of cells through multi-scale visual network explorations, new visualization software tools still need to be developed to empower the analysis. A robust visual data analysis platform driven by database management systems to perform bi-directional data processing-to-visualizations with declarative querying capabilities is needed. We developed ProteoLens as a JAVA-based visual analytic software tool for creating, annotating and exploring multi-scale biological networks. It supports direct database connectivity to either Oracle or PostgreSQL database tables/views, on which SQL statements using both Data Definition Languages (DDL) and Data Manipulation languages (DML) may be specified. The robust query languages embedded directly within the visualization software help users to bring their network data into a visualization context for annotation and exploration. ProteoLens supports graph/network represented data in standard Graph Modeling Language (GML) formats, and this enables interoperation with a wide range of other visual layout tools. The architectural design of ProteoLens enables the de-coupling of complex network data visualization tasks into two distinct phases: 1) creating network data association rules, which are mapping rules between network node IDs or edge IDs and data attributes such as functional annotations, expression levels, scores, synonyms, descriptions etc; 2) applying network data association rules to build the network and perform the visual annotation of graph nodes and edges according to associated data values. We demonstrated the advantages of these new capabilities through three biological network visualization case studies: human disease association network, drug-target interaction network and protein-peptide mapping network. The architectural design of ProteoLens makes it suitable for bioinformatics expert data analysts who are experienced with relational database management to perform large-scale integrated network visual explorations. ProteoLens is a promising visual analytic platform that will facilitate knowledge discoveries in future network and systems biology studies.
Development of the USGS national land-cover database over two decades
Xian, George Z.; Homer, Collin G.; Yang, Limin; Weng, Qihao
2011-01-01
Land-cover composition and change have profound impacts on terrestrial ecosystems. Land-cover and land-use (LCLU) conditions and their changes can affect social and physical environments by altering ecosystem conditions and services. Information about LCLU change is often used to produce landscape-based metrics and evaluate landscape conditions to monitor LCLU status and trends over a specific time interval (Loveland et al. 2002; Coppin et al. 2004; Lunetta et al. 2006). Continuous, accurate, and up-to-date land-cover data are important for natural resource and ecosystem management and are needed to support consistent monitoring of landscape attributes over time. Large-area land-cover information at regional, national, and global scales is critical for monitoring landscape variations over large areas.
NASA Astrophysics Data System (ADS)
Bi, Jiantao; Luo, Guilin; Wang, Xingxing; Zhu, Zuojia
2014-03-01
As the bridge over the Chinese and Western civilization, the ancient Silk Road has made a huge contribution to cultural, economic, political exchanges between China and western countries. In this paper, we treated the historical period of Western Han Dynasty, Eastern Han Dynasty and Tang Dynasty as the research time domain, and the Western Regions' countries that were existed along the Silk Road at the mean time as the research spatial domain. Then we imported these data into the SQL Server database we constructed, from which we could either query the attribute information such as population, military force, the era of the Central Plains empire, the significant events taking place in the country and some related attribute information of these events like the happened calendar year in addition to some related spatial information such as the present location, the coordinates of the capital and the territory by inputting the name of the Western countries. At the same time we could query the significant events, government institution in Central Plains and the existent Western countries at the mean time by inputting the calendar year. Based on the database, associated with GIS, RS, Flex, C# and other related information technology and network technology, we could not only browsing, searching and editing the information of the ancient Silk Road in Xinjiang Province during the Han and Tang Dynasties, but preliminary analysing as well. This is the combination of archaeology and modern information technology, and the database could also be a reference to further study, research and practice in the related fields in the future.
NASA Astrophysics Data System (ADS)
Erickson, M.; Olaguer, J.; Wijesinghe, A.; Colvin, J.; Neish, B.; Williams, J.
2014-12-01
It is becoming increasingly important to understand the emissions and health effects of industrial facilities. Many areas have no or limited sustained monitoring capabilities, making it difficult to quantify the major pollution sources affecting human health, especially in fence line communities. Developments in real-time monitoring and micro-scale modeling offer unique ways to tackle these complex issues. This presentation will demonstrate the capability of coupling real-time observations with micro-scale modeling to provide real-time information and near real-time source attribution. The Houston Advanced Research Center constructed the Mobile Acquisition of Real-time Concentrations (MARC) laboratory. MARC consists of a Ford E-350 passenger van outfitted with a Proton Transfer Reaction Mass Spectrometer (PTR-MS) and meteorological equipment. This allows for the fast measurement of various VOCs important to air quality. The data recorded from the van is uploaded to an off-site database and the information is broadcast to a website in real-time. This provides for off-site monitoring of MARC's observations, which allows off-site personnel to provide immediate input to the MARC operators on how to best achieve project objectives. The information stored in the database can also be used to provide near real-time source attribution. An inverse model has been used to ascertain the amount, location, and timing of emissions based on MARC measurements in the vicinity of industrial sites. The inverse model is based on a 3D micro-scale Eulerian forward and adjoint air quality model known as the HARC model. The HARC model uses output from the Quick Urban and Industrial Complex (QUIC) wind model and requires a 3D digital model of the monitored facility based on lidar or industrial permit data. MARC is one of the instrument platforms deployed during the 2014 Benzene and other Toxics Exposure Study (BEE-TEX) in Houston, TX. The main goal of the study is to quantify and explain the origin of ambient exposure to hazardous air pollutants in an industrial fence line community near the Houston Ship Channel. Preliminary results derived from analysis of MARC observations during the BEE-TEX experiment will be presented.
Le Vu, Stéphane; Ratmann, Oliver; Delpech, Valerie; Brown, Alison E; Gill, O Noel; Tostevin, Anna; Fraser, Christophe; Volz, Erik M
2018-06-01
Phylogenetic clustering of HIV sequences from a random sample of patients can reveal epidemiological transmission patterns, but interpretation is hampered by limited theoretical support and statistical properties of clustering analysis remain poorly understood. Alternatively, source attribution methods allow fitting of HIV transmission models and thereby quantify aspects of disease transmission. A simulation study was conducted to assess error rates of clustering methods for detecting transmission risk factors. We modeled HIV epidemics among men having sex with men and generated phylogenies comparable to those that can be obtained from HIV surveillance data in the UK. Clustering and source attribution approaches were applied to evaluate their ability to identify patient attributes as transmission risk factors. We find that commonly used methods show a misleading association between cluster size or odds of clustering and covariates that are correlated with time since infection, regardless of their influence on transmission. Clustering methods usually have higher error rates and lower sensitivity than source attribution method for identifying transmission risk factors. But neither methods provide robust estimates of transmission risk ratios. Source attribution method can alleviate drawbacks from phylogenetic clustering but formal population genetic modeling may be required to estimate quantitative transmission risk factors. Copyright © 2017 The Authors. Published by Elsevier B.V. All rights reserved.
Chewing and Attention: A Positive Effect on Sustained Attention
Onozuka, Minoru
2015-01-01
Chewing is crushing food not only to aid swallowing and digestion, but also to help stress relief and regulate cognitive function, especially in attention. It is well known that chewing gum is used for sleepiness prevention during work, learning, and driving, suggesting a link between chewing and sustained attention. We hypothesized that chewing elevates attention and/or alertness, leading to improvements in cognitive performance. We carried out a systematic review of the PubMed database. We inspected the attributes of effects on attention in studies investigating the effects of chewing on attention or alertness conducted with pre-post design in healthy subjects, except elderly. We identified 151 references, 22 of which were included: 14 (64%) showed positive attributes of effects on attention, 1 (5%) showed negative attributes of effects on attention, 5 (23%) showed both positive and negative attributes of effects on attention, and 2 (9%) showed no significant attributes of effects on attention. Thus, positive attributes of effects of chewing on attention, especially on sustained attention, were shown in over half of the reports. These effects also appeared with improvement in mood and stress relief and were influenced by time-on-task effect. Further studies are needed, but chewing could be useful for modifying cognitive function. PMID:26075234
An object-oriented approach to deploying highly configurable Web interfaces for the ATLAS experiment
NASA Astrophysics Data System (ADS)
Lange, Bruno; Maidantchik, Carmen; Pommes, Kathy; Pavani, Varlen; Arosa, Breno; Abreu, Igor
2015-12-01
The ATLAS Technical Coordination disposes of 17 Web systems to support its operation. These applications, whilst ranging from managing the process of publishing scientific papers to monitoring radiation levels in the equipment in the experimental cavern, are constantly prone to changes in requirements due to the collaborative nature of the experiment and its management. In this context, a Web framework is proposed to unify the generation of the supporting interfaces. FENCE assembles classes to build applications by making extensive use of JSON configuration files. It relies heavily on Glance, a technology that was set forth in 2003 to create an abstraction layer on top of the heterogeneous sources that store the technical coordination data. Once Glance maps out the database modeling, records can be referenced in the configuration files by wrapping unique identifiers around double enclosing brackets. The deployed content can be individually secured by attaching clearance attributes to their description thus ensuring that view/edit privileges are granted to eligible users only. The framework also provides tools for securely writing into a database. Fully HTML5-compliant multi-step forms can be generated from their JSON description to assure that the submitted data comply with a series of constraints. Input validation is carried out primarily on the server- side but, following progressive enhancement guidelines, verification might also be performed on the client-side by enabling specific markup data attributes which are then handed over to the jQuery validation plug-in. User monitoring is accomplished by thoroughly logging user requests along with any POST data. Documentation is built from the source code using the phpDocumentor tool and made readily available for developers online. Fence, therefore, speeds up the implementation of Web interfaces and reduces the response time to requirement changes by minimizing maintenance overhead.
A fuzzy hill-climbing algorithm for the development of a compact associative classifier
NASA Astrophysics Data System (ADS)
Mitra, Soumyaroop; Lam, Sarah S.
2012-02-01
Classification, a data mining technique, has widespread applications including medical diagnosis, targeted marketing, and others. Knowledge discovery from databases in the form of association rules is one of the important data mining tasks. An integrated approach, classification based on association rules, has drawn the attention of the data mining community over the last decade. While attention has been mainly focused on increasing classifier accuracies, not much efforts have been devoted towards building interpretable and less complex models. This paper discusses the development of a compact associative classification model using a hill-climbing approach and fuzzy sets. The proposed methodology builds the rule-base by selecting rules which contribute towards increasing training accuracy, thus balancing classification accuracy with the number of classification association rules. The results indicated that the proposed associative classification model can achieve competitive accuracies on benchmark datasets with continuous attributes and lend better interpretability, when compared with other rule-based systems.
An analysis of IGBP global land-cover characterization process
Loveland, Thomas R.; Zhu, Zhiliang; Ohlen, Donald O.; Brown, Jesslyn F.; Reed, Bradley C.; Yang, Limin
1999-01-01
The international Geosphere Biosphere Programme (IGBP) has called for the development of improved global land-cover data for use in increasingly sophisticated global environmental models. To meet this need, the staff of the U.S. Geological Survey and the University of Nebraska-Lincoln developed and applied a global land-cover characterization methodology using 1992-1993 1-km resolution Advanced Very High Resolution Radiometer (AVHRR) and other spatial data. The methodology, based on unsupervised classification with extensive postclassification refinement, yielded a multi-layer database consisting of eight land-cover data sets, descriptive attributes, and source data. An independent IGBP accuracy assessment reports a global accuracy of 73.5 percent, and continental results vary from 63 percent to 83 percent. Although data quality, methodology, interpreter performance, and logistics affected the results, significant problems were associated with the relationship between AVHRR data and fine-scale, spectrally similar land-cover patterns in complex natural or disturbed landscapes.
NASA Astrophysics Data System (ADS)
Ilieva, Tamara; Gekov, Svetoslav
2017-04-01
The Precise Point Positioning (PPP) method gives the users the opportunity to determine point locations using a single GNSS receiver. The accuracy of the determined by PPP point locations is better in comparison to the standard point positioning, due to the precise satellite orbit and clock corrections that are developed and maintained by the International GNSS Service (IGS). The aim of our current research is the accuracy assessment of the PPP method applied for surveys and tracking moving objects in GIS environment. The PPP data is collected by using preliminary developed by us software application that allows different sets of attribute data for the measurements and their accuracy to be used. The results from the PPP measurements are directly compared within the geospatial database to different other sets of terrestrial data - measurements obtained by total stations, real time kinematic and static GNSS.
A watermarking algorithm for polysomnography data.
Jamasebi, R; Johnson, N L; Kaffashi, F; Redline, S; Loparo, K A
2008-01-01
A blind watermarking algorithm for polysomnography (PSG) data in European Data Format (EDF) has been developed for the identification and attribution of shared data. This is accomplished by hiding a unique identifier in the phase spectrum of each PSG epoch using an undisclosed key so that a third party cannot retrieve the watermark without knowledge of the key. A pattern discovery algorithm is developed to find the watermark pattern even though the data may have been altered. The method is evaluated using 25 PSG studies from the Sleep Heart Health Study database. The integrity of the signal data was determined using time series measures of both the original and watermarked signals, and by determining its effect on scoring sleep stages from the PSG data. The results of the analysis indicate that the proposed watermarking method for PSG data is an effective and efficient way to identify shared data without compromising its intended use.
Developing Privacy Solutions for Sharing and Analyzing Healthcare Data
Motiwalla, Luvai; Li, Xiao-Bai
2013-01-01
The extensive use of electronic health data has increased privacy concerns. While most healthcare organizations are conscientious in protecting their data in their databases, very few organizations take enough precautions to protect data that is shared with third party organizations. Recently the regulatory environment has tightened the laws to enforce privacy protection. The goal of this research is to explore the application of data masking solutions for protecting patient privacy when data is shared with external organizations for research, analysis and other similar purposes. Specifically, this research project develops a system that protects data without removing sensitive attributes. Our application allows high quality data analysis with the masked data. Dataset-level properties and statistics remain approximately the same after data masking; however, individual record-level values are altered to prevent privacy disclosure. A pilot evaluation study on large real-world healthcare data shows the effectiveness of our solution in privacy protection. PMID:24285983
Occupational burden of asbestos-related cancer in Argentina, Brazil, Colombia, and Mexico.
Pasetto, Roberto; Terracini, Benedetto; Marsili, Daniela; Comba, Pietro
2014-01-01
An estimate at the national level of the occupational cancer burden brought about by the industrial use of asbestos requires detailed routine information on such uses as well as on vital statistics of good quality. A causal association with asbestos exposure has been established for mesothelioma and cancers of the lung, larynx, and ovary. The aim of this study was to provide estimates of the occupational burden of asbestos-related cancer for the Latin American countries that are or have been the highest asbestos consumers in the region: Argentina, Brazil, Colombia, and Mexico. The burden of multifactorial cancers has been estimated through the approach suggested for the World Health Organization using the population attributable fraction. The following data were used: Proportion of workforce employed in each economic sector. Proportion of workers exposed to asbestos in each sector. Occupational turnover. Levels of exposure. Proportion of the population in the workforce. Relative risk for each considered disease for 1 or more levels of exposure. Data on the proportion of workers exposed to asbestos in each sector are not available for Latin American countries; therefore, data from the European CAREX database (carcinogen exposure database) were used. Using mortality data of the World Health Organization Health Statistics database for the year 2009 and applying the estimated values for population attributable fractions, the number of estimated deaths in 5 years for mesothelioma and for lung, larynx, and ovary cancers attributable to occupational asbestos exposures, were respectively 735, 233, 29, and 14 for Argentina; 340, 611, 68, and 43 for Brazil; 255, 97, 14, and 9 for Colombia, and 1075, 219, 18, and 22 for Mexico. The limitations in compiling the estimates highlight the need for improvement in the quality of asbestos-related environmental and health data. Nevertheless, the figures are already usable to promote a ban on asbestos use. Copyright © 2014 Icahn School of Medicine at Mount Sinai. Published by Elsevier Inc. All rights reserved.
Advances in computational metabolomics and databases deepen the understanding of metabolisms.
Tsugawa, Hiroshi
2018-01-29
Mass spectrometry (MS)-based metabolomics is the popular platform for metabolome analyses. Computational techniques for the processing of MS raw data, for example, feature detection, peak alignment, and the exclusion of false-positive peaks, have been established. The next stage of untargeted metabolomics would be to decipher the mass fragmentation of small molecules for the global identification of human-, animal-, plant-, and microbiota metabolomes, resulting in a deeper understanding of metabolisms. This review is an update on the latest computational metabolomics including known/expected structure databases, chemical ontology classifications, and mass spectrometry cheminformatics for the interpretation of mass fragmentations and for the elucidation of unknown metabolites. The importance of metabolome 'databases' and 'repositories' is also discussed because novel biological discoveries are often attributable to the accumulation of data, to relational databases, and to their statistics. Lastly, a practical guide for metabolite annotations is presented as the summary of this review. Copyright © 2018 Elsevier Ltd. All rights reserved.
Evolution of the NASA/IPAC Extragalactic Database (NED) into a Data Mining Discovery Engine
NASA Astrophysics Data System (ADS)
Mazzarella, Joseph M.; NED Team
2017-06-01
We review recent advances and ongoing work in evolving the NASA/IPAC Extragalactic Database (NED) beyond an object reference database into a data mining discovery engine. Updates to the infrastructure and data integration techniques are enabling more than a 10-fold expansion; NED will soon contain over a billion objects with their fundamental attributes fused across the spectrum via cross-identifications among the largest sky surveys (e.g., GALEX, SDSS, 2MASS, AllWISE, EMU), and over 100,000 smaller but scientifically important catalogs and journal articles. The recent discovery of super-luminous spiral galaxies exemplifies the opportunities for data mining and science discovery directly from NED's rich data synthesis. Enhancements to the user interface, including new APIs, VO protocols, and queries involving derived physical quantities, are opening new pathways for panchromatic studies of large galaxy samples. Examples are shown of graphics characterizing the content of NED, as well as initial steps in exploring the database via interactive statistical visualizations.
MAGIC database and interfaces: an integrated package for gene discovery and expression.
Cordonnier-Pratt, Marie-Michèle; Liang, Chun; Wang, Haiming; Kolychev, Dmitri S; Sun, Feng; Freeman, Robert; Sullivan, Robert; Pratt, Lee H
2004-01-01
The rapidly increasing rate at which biological data is being produced requires a corresponding growth in relational databases and associated tools that can help laboratories contend with that data. With this need in mind, we describe here a Modular Approach to a Genomic, Integrated and Comprehensive (MAGIC) Database. This Oracle 9i database derives from an initial focus in our laboratory on gene discovery via production and analysis of expressed sequence tags (ESTs), and subsequently on gene expression as assessed by both EST clustering and microarrays. The MAGIC Gene Discovery portion of the database focuses on information derived from DNA sequences and on its biological relevance. In addition to MAGIC SEQ-LIMS, which is designed to support activities in the laboratory, it contains several additional subschemas. The latter include MAGIC Admin for database administration, MAGIC Sequence for sequence processing as well as sequence and clone attributes, MAGIC Cluster for the results of EST clustering, MAGIC Polymorphism in support of microsatellite and single-nucleotide-polymorphism discovery, and MAGIC Annotation for electronic annotation by BLAST and BLAT. The MAGIC Microarray portion is a MIAME-compliant database with two components at present. These are MAGIC Array-LIMS, which makes possible remote entry of all information into the database, and MAGIC Array Analysis, which provides data mining and visualization. Because all aspects of interaction with the MAGIC Database are via a web browser, it is ideally suited not only for individual research laboratories but also for core facilities that serve clients at any distance.
Organization of Heterogeneous Scientific Data Using the EAV/CR Representation
Nadkarni, Prakash M.; Marenco, Luis; Chen, Roland; Skoufos, Emmanouil; Shepherd, Gordon; Miller, Perry
1999-01-01
Entity-attribute-value (EAV) representation is a means of organizing highly heterogeneous data using a relatively simple physical database schema. EAV representation is widely used in the medical domain, most notably in the storage of data related to clinical patient records. Its potential strengths suggest its use in other biomedical areas, in particular research databases whose schemas are complex as well as constantly changing to reflect evolving knowledge in rapidly advancing scientific domains. When deployed for such purposes, the basic EAV representation needs to be augmented significantly to handle the modeling of complex objects (classes) as well as to manage interobject relationships. The authors refer to their modification of the basic EAV paradigm as EAV/CR (EAV with classes and relationships). They describe EAV/CR representation with examples from two biomedical databases that use it. PMID:10579606
LARCRIM user's guide, version 1.0
NASA Technical Reports Server (NTRS)
Davis, John S.; Heaphy, William J.
1993-01-01
LARCRIM is a relational database management system (RDBMS) which performs the conventional duties of an RDBMS with the added feature that it can store attributes which consist of arrays or matrices. This makes it particularly valuable for scientific data management. It is accessible as a stand-alone system and through an application program interface. The stand-alone system may be executed in two modes: menu or command. The menu mode prompts the user for the input required to create, update, and/or query the database. The command mode requires the direct input of LARCRIM commands. Although LARCRIM is an update of an old database family, its performance on modern computers is quite satisfactory. LARCRIM is written in FORTRAN 77 and runs under the UNIX operating system. Versions have been released for the following computers: SUN (3 & 4), Convex, IRIS, Hewlett-Packard, CRAY 2 & Y-MP.
Social justice: a concept analysis.
Buettner-Schmidt, Kelly; Lobo, Marie L
2012-04-01
This article is a report of an analysis of the concept of social justice. Nursing's involvement in social justice has waned in the recent past. A resurgence of interest in nurses' roles about social justice requires a clear understanding of the concept. Literature for this concept analysis included English language articles from CINAHL, PubMed, and broad multidisciplinary literature databases, within and outside of health-related literature, for the years 1968-2010. Two books and appropriate websites were also reviewed. The reference lists of the identified sources were reviewed for additional sources. The authors used Wilsonian methods of concept analysis as a guide. An efficient, synthesized definition of social justice was developed, based on the identification of its attributes, antecedents and consequences that provides clarification of the concept. Social justice was defined as full participation in society and the balancing of benefits and burdens by all citizens, resulting in equitable living and a just ordering of society. Its attributes included: (1) fairness; (2) equity in the distribution of power, resources, and processes that affect the sufficiency of the social determinants of health; (3) just institutions, systems, structures, policies, and processes; (4) equity in human development, rights, and sustainability; and (5) sufficiency of well-being. Nurses can have an important influence on the health of people globally by reinvesting in social justice. Implications for research, education, practice and policy, such as development of a social justice framework and educational competencies are presented. © 2011 The Authors. Journal of Advanced Nursing © 2011 Blackwell Publishing Ltd.
Ranjbar Ezatabadi, Mohammad; Rashidian, Arash; Shariati, Mohammad; Rahimi Foroushani, Abbas; Akbari Sari, Ali
2016-01-01
Background Family physician plans in Iran face several challenges, one of which is developing attractive and efficient contracts that motivate physicians to participate in the plan. Objectives This study aimed to elicit GPs’ preferences for family physician contracts. Patients and Methods In a cross-sectional study using the conjoint analysis technique, 580 GPs selected from the family physician database in Iran in 2014. Through qualitative and quantitative methods, 18 contract scenarios were developed via orthogonal design i.e., the impact of each attribute is measured independently from changes in other attributes and a questionnaire was developed. Data were collected through this questionnaire and analyzed using the ordered logistic regression (OLR) model. Results The results show that “quotas for admission to specialized courses” is the strongest preference of GPs (β = 1.123). In order of importance, the other preferences are having the right to provide services outside of the specified package (β = 0.962), increased number of covered population (β = 0.814), capitation payment + 15% bonus (β = 0.644), increased catchment area to 5 km (β = 0.349), and increased length of contract to five years (β = 0.345). Conclusions The conjoint analysis results show that GPs concerned about various factors of family physician contracts. These results can be helpful for policy-makers as they complete the process of creating family physician plans, which can help increase the motivation of GPs to participate in the plan. PMID:28191339
Setting the Alarm: Word Emotional Attributes Require Consolidation to be Operational.
Dumay, Nicolas; Sharma, Dinkar; Kellen, Nora; Abdelrahim, Sarah
2018-01-25
Demonstrations of emotional Stroop in conditioned made-up words are flawed because of the lack of task ensuring similar word encoding across conditions. Here, participants were trained on associations between made-up words (e.g., 'drott') and pictures with an alarming or neutral content (e.g., 'a dead sheep' vs. 'a munching cow') in a situation that required attention to both ends of each association. To test whether word emotional attributes need to consolidate before they can hijack attention, one set of associations was learned seven days before the test, whereas the other set was learned either six hrs or immediately before the test. The novel words' ability to evoke their emotional attributes was assessed by using both Stroop and an auditory analogue called pause detection. Matching words and pictures was harder for alarming associations. However, similar learning rate and forgetting at seven days were observed for both types of associations. Pause detection revealed no emotion effect for same-day (i.e., unconsolidated) associations, but robust interference for seven-day-old (i.e., consolidated) alarming associations. Attention capture was found in the emotional Stroop as well, though only when trial n-1 referred to a same-day association. This task also showed stronger response repetition priming (independently of emotion) when trials n and n-1 both tapped into seven-day-old associations. Word emotional attributes hence take between six hrs and seven days to be operational. Moreover, age interactions between consecutive trials can be used to gauge implicitly the indirect (relational) episodic associations that develop in the meantime between the memories of individual items. (PsycINFO Database Record (c) 2018 APA, all rights reserved).
Evolutionary conceptual analysis: faith community nursing.
Ziebarth, Deborah
2014-12-01
The aim of the study was to report an evolutionary concept analysis of faith community nursing (FCN). FCN is a source of healthcare delivery in the USA which has grown in comprehensiveness and complexity. With increasing healthcare cost and a focus on access and prevention, FCN has extended beyond the physical walls of the faith community building. Faith communities and healthcare organizations invest in FCN and standardized training programs exist. Using Rodgers' evolutionary analysis, the literature was examined for antecedents, attributes, and consequences of the concept. This design allows for understanding the historical and social nature of the concept and how it changes over time. A search of databases using the keywords FCN, faith community nurse, parish nursing, and parish nurse was done. The concept of FCN was explored using research and theoretical literature. A theoretical definition and model were developed with relevant implications. The search results netted a sample of 124 reports of research and theoretical articles from multiple disciplines: medicine, education, religion and philosophy, international health, and nursing. Theoretical definition: FCN is a method of healthcare delivery that is centered in a relationship between the nurse and client (client as person, family, group, or community). The relationship occurs in an iterative motion over time when the client seeks or is targeted for wholistic health care with the goal of optimal wholistic health functioning. Faith integrating is a continuous occurring attribute. Health promoting, disease managing, coordinating, empowering and accessing health care are other essential attributes. All essential attributes occur with intentionality in a faith community, home, health institution and other community settings with fluidity as part of a community, national, or global health initiative. A new theoretical definition and corresponding conceptual model of FCN provides a basis for future nursing knowledge and model-based applications for evidence-based practice and research.
2018-01-01
Introduction The aims of this study were to estimate all-cause and cause-specific mortality and years of life lost, investigated by disability-adjusted life-years (DALYs), due to colorectal cancer attributable to physical inactivity in Brazil and in the states; to analyze the temporal trend of these estimates over 25 years (1990–2015) compared with global estimates and according to the socioeconomic status of states of Brazil. Methods Databases from the Global Burden of Disease Study (GBD) for Brazil, Brazilian states and global information were used. It was estimated the total number and the age-standardized rates of deaths and DALYs for colorectal cancer attributable to physical inactivity in the years 1990 and 2015. We used the Socioeconomic Development Index (SDI). Results Physical inactivity was responsible for a substantial number of deaths (1990: 1,302; 2015: 119,351) and DALYs (1990: 31,121; 2015: 87,116) due to colorectal cancer in Brazil. From 1990 to 2015, the mortality and DALYs due to colorectal cancer attributable to physical inactivity increased in Brazil (0.6% and 0.6%, respectively) and decreased around the world (-0.8% and -1.1%, respectively). The Brazilian states with better socioeconomic indicators had higher rates of mortality and morbidity by colorectal cancer due to physical inactivity (p<0.01). Physical inactivity was responsible for deaths and DALYs due to colorectal cancer in Brazil. Conclusions Over 25 years, the Brazilian population showed more worrisome results than around the world. Actions to combat physical inactivity and greater cancer screening and treatment are urgent in the Brazilian states. PMID:29390002
Towards Usable E-Health. A Systematic Review of Usability Questionnaires.
Sousa, Vanessa E C; Dunn Lopez, Karen
2017-05-10
The use of e-health can lead to several positive outcomes. However, the potential for e-health to improve healthcare is partially dependent on its ease of use. In order to determine the usability for any technology, rigorously developed and appropriate measures must be chosen. To identify psychometrically tested questionnaires that measure usability of e-health tools, and to appraise their generalizability, attributes coverage, and quality. We conducted a systematic review of studies that measured usability of e-health tools using four databases (Scopus, PubMed, CINAHL, and HAPI). Non-primary research, studies that did not report measures, studies with children or people with cognitive limitations, and studies about assistive devices or medical equipment were systematically excluded. Two authors independently extracted information including: questionnaire name, number of questions, scoring method, item generation, and psychometrics using a data extraction tool with pre-established categories and a quality appraisal scoring table. Using a broad search strategy, 5,558 potentially relevant papers were identified. After removing duplicates and applying exclusion criteria, 35 articles remained that used 15 unique questionnaires. From the 15 questionnaires, only 5 were general enough to be used across studies. Usability attributes covered by the questionnaires were: learnability (15), efficiency (12), and satisfaction (11). Memorability (1) was the least covered attribute. Quality appraisal showed that face/content (14) and construct (7) validity were the most frequent types of validity assessed. All questionnaires reported reliability measurement. Some questionnaires scored low in the quality appraisal for the following reasons: limited validity testing (7), small sample size (3), no reporting of user centeredness (9) or feasibility estimates of time, effort, and expense (7). Existing questionnaires provide a foundation for research on e-health usability. However, future research is needed to broaden the coverage of the usability attributes and psychometric properties of the available questionnaires.
Silva, Diego Augusto Santos; Tremblay, Mark Stephen; Souza, Maria de Fatima Marinho de; Mooney, Meghan; Naghavi, Mohsen; Malta, Deborah Carvalho
2018-01-01
The aims of this study were to estimate all-cause and cause-specific mortality and years of life lost, investigated by disability-adjusted life-years (DALYs), due to colorectal cancer attributable to physical inactivity in Brazil and in the states; to analyze the temporal trend of these estimates over 25 years (1990-2015) compared with global estimates and according to the socioeconomic status of states of Brazil. Databases from the Global Burden of Disease Study (GBD) for Brazil, Brazilian states and global information were used. It was estimated the total number and the age-standardized rates of deaths and DALYs for colorectal cancer attributable to physical inactivity in the years 1990 and 2015. We used the Socioeconomic Development Index (SDI). Physical inactivity was responsible for a substantial number of deaths (1990: 1,302; 2015: 119,351) and DALYs (1990: 31,121; 2015: 87,116) due to colorectal cancer in Brazil. From 1990 to 2015, the mortality and DALYs due to colorectal cancer attributable to physical inactivity increased in Brazil (0.6% and 0.6%, respectively) and decreased around the world (-0.8% and -1.1%, respectively). The Brazilian states with better socioeconomic indicators had higher rates of mortality and morbidity by colorectal cancer due to physical inactivity (p<0.01). Physical inactivity was responsible for deaths and DALYs due to colorectal cancer in Brazil. Over 25 years, the Brazilian population showed more worrisome results than around the world. Actions to combat physical inactivity and greater cancer screening and treatment are urgent in the Brazilian states.
Mihelčić, Matej; Šimić, Goran; Babić Leko, Mirjana; Lavrač, Nada; Džeroski, Sašo; Šmuc, Tomislav
2017-01-01
Based on a set of subjects and a collection of attributes obtained from the Alzheimer's Disease Neuroimaging Initiative database, we used redescription mining to find interpretable rules revealing associations between those determinants that provide insights about the Alzheimer's disease (AD). We extended the CLUS-RM redescription mining algorithm to a constraint-based redescription mining (CBRM) setting, which enables several modes of targeted exploration of specific, user-constrained associations. Redescription mining enabled finding specific constructs of clinical and biological attributes that describe many groups of subjects of different size, homogeneity and levels of cognitive impairment. We confirmed some previously known findings. However, in some instances, as with the attributes: testosterone, ciliary neurotrophic factor, brain natriuretic peptide, Fas ligand, the imaging attribute Spatial Pattern of Abnormalities for Recognition of Early AD, as well as the levels of leptin and angiopoietin-2 in plasma, we corroborated previously debatable findings or provided additional information about these variables and their association with AD pathogenesis. Moreover, applying redescription mining on ADNI data resulted with the discovery of one largely unknown attribute: the Pregnancy-Associated Protein-A (PAPP-A), which we found highly associated with cognitive impairment in AD. Statistically significant correlations (p ≤ 0.01) were found between PAPP-A and clinical tests: Alzheimer's Disease Assessment Scale, Clinical Dementia Rating Sum of Boxes, Mini Mental State Examination, etc. The high importance of this finding lies in the fact that PAPP-A is a metalloproteinase, known to cleave insulin-like growth factor binding proteins. Since it also shares similar substrates with A Disintegrin and the Metalloproteinase family of enzymes that act as α-secretase to physiologically cleave amyloid precursor protein (APP) in the non-amyloidogenic pathway, it could be directly involved in the metabolism of APP very early during the disease course. Therefore, further studies should investigate the role of PAPP-A in the development of AD more thoroughly.
Mihelčić, Matej; Šimić, Goran; Babić Leko, Mirjana; Lavrač, Nada; Džeroski, Sašo; Šmuc, Tomislav
2017-01-01
Based on a set of subjects and a collection of attributes obtained from the Alzheimer’s Disease Neuroimaging Initiative database, we used redescription mining to find interpretable rules revealing associations between those determinants that provide insights about the Alzheimer’s disease (AD). We extended the CLUS-RM redescription mining algorithm to a constraint-based redescription mining (CBRM) setting, which enables several modes of targeted exploration of specific, user-constrained associations. Redescription mining enabled finding specific constructs of clinical and biological attributes that describe many groups of subjects of different size, homogeneity and levels of cognitive impairment. We confirmed some previously known findings. However, in some instances, as with the attributes: testosterone, ciliary neurotrophic factor, brain natriuretic peptide, Fas ligand, the imaging attribute Spatial Pattern of Abnormalities for Recognition of Early AD, as well as the levels of leptin and angiopoietin-2 in plasma, we corroborated previously debatable findings or provided additional information about these variables and their association with AD pathogenesis. Moreover, applying redescription mining on ADNI data resulted with the discovery of one largely unknown attribute: the Pregnancy-Associated Protein-A (PAPP-A), which we found highly associated with cognitive impairment in AD. Statistically significant correlations (p ≤ 0.01) were found between PAPP-A and clinical tests: Alzheimer’s Disease Assessment Scale, Clinical Dementia Rating Sum of Boxes, Mini Mental State Examination, etc. The high importance of this finding lies in the fact that PAPP-A is a metalloproteinase, known to cleave insulin-like growth factor binding proteins. Since it also shares similar substrates with A Disintegrin and the Metalloproteinase family of enzymes that act as α-secretase to physiologically cleave amyloid precursor protein (APP) in the non-amyloidogenic pathway, it could be directly involved in the metabolism of APP very early during the disease course. Therefore, further studies should investigate the role of PAPP-A in the development of AD more thoroughly. PMID:29088293
Joint seismic-infrasonic processing of recordings from a repeating source of atmospheric explosions.
Gibbons, Steven J; Ringdal, Frode; Kvaerna, Tormod
2007-11-01
A database has been established of seismic and infrasonic recordings from more than 100 well-constrained surface explosions, conducted by the Finnish military to destroy old ammunition. The recorded seismic signals are essentially identical and indicate that the variation in source location and magnitude is negligible. In contrast, the infrasonic arrivals on both seismic and infrasound sensors exhibit significant variation both with regard to the number of detected phases, phase travel times, and phase amplitudes, which would be attributable to atmospheric factors. This data set provides an excellent database for studies in sound propagation, infrasound array detection, and direction estimation.
Integrated cluster management at Manchester
NASA Astrophysics Data System (ADS)
McNab, Andrew; Forti, Alessandra
2012-12-01
We describe an integrated management system using third-party, open source components used in operating a large Tier-2 site for particle physics. This system tracks individual assets and records their attributes such as MAC and IP addresses; derives DNS and DHCP configurations from this database; creates each host's installation and re-configuration scripts; monitors the services on each host according to the records of what should be running; and cross references tickets with asset records and per-asset monitoring pages. In addition, scripts which detect problems and automatically remove hosts record these new states in the database which are available to operators immediately through the same interface as tickets and monitoring.
Fautrel, Bruno; Cukierman, Gabrielle; Joubert, Jean-Michel; Laurendeau, Caroline; Gourmelen, Julie; Fagnani, Francis
2016-01-01
To estimate healthcare service utilisation costs of patients with rheumatoid arthritis in France and to estimate the fraction of these costs attributable to RA. The "Échantillon généraliste des bénéficiaires" (EGB) is a 1/97 random sample of the main national claims database covering the French population. A cohort of patients with rheumatoid arthritis was constituted of all adults benefiting from full coverage for rheumatoid arthritis (ICD-10 M05-06) on 1st january 2009. A control group matched for age and gender was identified. Health expenditures were assessed from the payer's perspective for the year 2010. The annual per capita reimbursed total health expenditure was €6,404 in 2010, an amount around two times higher than in the control group €3,095 (P<0.0001). The main contributors to this extra cost were outpatient care (+€2,407; 72.7%), including medication (+€1,686; 50.0%), and inpatient care (+€903; 27.3%). Patients treated by biological agents generated an age-adjusted per capita annual expenditure about three times higher than untreated patients (€15,757 versus €4,640). Only half of medical expenditure by patients with rheumatoid arthritis is attributable to their disease and use of biological agents has become a major driver of cost. Copyright © 2015 Société française de rhumatologie. Published by Elsevier SAS. All rights reserved.
This dataset represents the dam density and storage volumes within individual, local NHDPlusV2 catchments and upstream, contributing watersheds based on National Inventory of Dams (NID) data. Attributes were calculated for every local NHDPlusV2 catchment and accumulated to provide watershed-level metrics.(See Supplementary Info for Glossary of Terms) The NID database contains information about the dam??s location, size, purpose, type, last inspection, regulatory facts, and other technical data. Structures on streams reduce the longitudinal and lateral hydrologic connectivity of the system. For example, impoundments above dams slow stream flow, cause deposition of sediment and reduce peak flows. Dams change both the discharge and sediment supply of streams, causing channel incision and bed coarsening downstream. Downstream areas are often sediment deprived, resulting in degradation, i.e., erosion of the stream bed and stream banks. This database was improved upon by locations verified by work from the USGS National Map (Jeff Simley Group). It was observed that some dams, some of them major and which do exist, were not part of the 2009 NID, but were represented in the USGS National Map dataset, and had been in the 2006 NID. Approximately 1,100 such dams were added, based on the USGS National Map lat/long and the 2006 NID attributes (dam height, storage, etc.) Finally, as clean-up, a) about 600 records with duplicate NIDID were removed, and b) about 300 reco
NASA Astrophysics Data System (ADS)
Zhu, Z.; Bi, J.; Wang, X.; Zhu, W.
2014-02-01
As an important sub-topic of the natural process of carbon emission data public information platform construction, coalfield spontaneous combustion of carbon emission WebGIS system has become an important study object. In connection with data features of coalfield spontaneous combustion carbon emissions (i.e. a wide range of data, which is rich and complex) and the geospatial characteristics, data is divided into attribute data and spatial data. Based on full analysis of the data, completed the detailed design of the Oracle database and stored on the Oracle database. Through Silverlight rich client technology and the expansion of WCF services, achieved the attribute data of web dynamic query, retrieval, statistical, analysis and other functions. For spatial data, we take advantage of ArcGIS Server and Silverlight-based API to invoke GIS server background published map services, GP services, Image services and other services, implemented coalfield spontaneous combustion of remote sensing image data and web map data display, data analysis, thematic map production. The study found that the Silverlight technology, based on rich client and object-oriented framework for WCF service, can efficiently constructed a WebGIS system. And then, combined with ArcGIS Silverlight API to achieve interactive query attribute data and spatial data of coalfield spontaneous emmission, can greatly improve the performance of WebGIS system. At the same time, it provided a strong guarantee for the construction of public information on China's carbon emission data.
Protoptype integrated design (Pride) system reference manual. Volume 2: Schema definition
NASA Technical Reports Server (NTRS)
Fishwick, P. A.; Sutter, T. R.; Blackburn, C. L.
1983-01-01
An initial description of an evolving relational database schema is presented for the management of finite element model design and analysis data. The report presents a description of each relation including attribute names, data types, and definitions. The format of this report is such that future modifications and enhancements may be easily incorporated.
National Wilderness Preservation System database: key attributes and trends, 1964 through 1999
Peter Landres; Shannon Meyer
2000-01-01
The Wilderness Act of 1964 established a National Wilderness Preservation System, and this publication is a compilation of selected information about every wilderness within this System. For each wilderness, the following information is given: legally correct wilderness name; public law that established the wilderness; date the enabling law was signed by the President...
Chen, Ming; Henry, Nathan; Almsaeed, Abdullah; Zhou, Xiao; Wegrzyn, Jill; Ficklin, Stephen
2017-01-01
Abstract Tripal is an open source software package for developing biological databases with a focus on genetic and genomic data. It consists of a set of core modules that deliver essential functions for loading and displaying data records and associated attributes including organisms, sequence features and genetic markers. Beyond the core modules, community members are encouraged to contribute extension modules to build on the Tripal core and to customize Tripal for individual community needs. To expand the utility of the Tripal software system, particularly for RNASeq data, we developed two new extension modules. Tripal Elasticsearch enables fast, scalable searching of the entire content of a Tripal site as well as the construction of customized advanced searches of specific data types. We demonstrate the use of this module for searching assembled transcripts by functional annotation. A second module, Tripal Analysis Expression, houses and displays records from gene expression assays such as RNA sequencing. This includes biological source materials (biomaterials), gene expression values and protocols used to generate the data. In the case of an RNASeq experiment, this would reflect the individual organisms and tissues used to produce sequencing libraries, the normalized gene expression values derived from the RNASeq data analysis and a description of the software or code used to generate the expression values. The module will load data from common flat file formats including standard NCBI Biosample XML. Data loading, display options and other configurations can be controlled by authorized users in the Drupal administrative backend. Both modules are open source, include usage documentation, and can be found in the Tripal organization’s GitHub repository. Database URL: Tripal Elasticsearch module: https://github.com/tripal/tripal_elasticsearch Tripal Analysis Expression module: https://github.com/tripal/tripal_analysis_expression PMID:29220446
Nims, Raymond W; Sykes, Greg; Cottrill, Karin; Ikonomi, Pranvera; Elmore, Eugene
2010-12-01
The role of cell authentication in biomedical science has received considerable attention, especially within the past decade. This quality control attribute is now beginning to be given the emphasis it deserves by granting agencies and by scientific journals. Short tandem repeat (STR) profiling, one of a few DNA profiling technologies now available, is being proposed for routine identification (authentication) of human cell lines, stem cells, and tissues. The advantage of this technique over methods such as isoenzyme analysis, karyotyping, human leukocyte antigen typing, etc., is that STR profiling can establish identity to the individual level, provided that the appropriate number and types of loci are evaluated. To best employ this technology, a standardized protocol and a data-driven, quality-controlled, and publically searchable database will be necessary. This public STR database (currently under development) will enable investigators to rapidly authenticate human-based cultures to the individual from whom the cells were sourced. Use of similar approaches for non-human animal cells will require developing other suitable loci sets. While implementing STR analysis on a more routine basis should significantly reduce the frequency of cell misidentification, additional technologies may be needed as part of an overall authentication paradigm. For instance, isoenzyme analysis, PCR-based DNA amplification, and sequence-based barcoding methods enable rapid confirmation of a cell line's species of origin while screening against cross-contaminations, especially when the cells present are not recognized by the species-specific STR method. Karyotyping may also be needed as a supporting tool during establishment of an STR database. Finally, good cell culture practices must always remain a major component of any effort to reduce the frequency of cell misidentification.
Family conference in palliative care: concept analysis.
Silva, Rudval Souza da; Trindade, Géssica Sodré Sampaio; Paixão, Gilvânia Patrícia do Nascimento; Silva, Maria Júlia Paes da
2018-01-01
to analyze the attributes, antecedents and consequents of the family conference concept. Walker and Avante's method for concept analysis and the stages of the integrative review process, with a selection of publications in the PubMed, Cinahl and Lilacs databases focusing on the family conference theme in the context of palliative care. the most cited antecedents were the presence of doubts and the need to define a care plan. Family reunion and working instrument were evidenced as attributes. With respect to consequents, to promote the effective communication and to establish a plan of consensual action were the most remarkable elements. the scarcity of publications on the subject was observed, as well as and the limitation of the empirical studies to the space of intensive therapy. Thus, by analyzing the attributes, antecedents and consequents of the concept it was possible to follow their evolution and to show their efficacy and effectiveness as a therapeutic intervention.
NASA Astrophysics Data System (ADS)
Hardebol, N. J.; Bertotti, G.
2013-04-01
This paper presents the development and use of our new DigiFract software designed for acquiring fracture data from outcrops more efficiently and more completely than done with other methods. Fracture surveys often aim at measuring spatial information (such as spacing) directly in the field. Instead, DigiFract focuses on collecting geometries and attributes and derives spatial information through subsequent analyses. Our primary development goal was to support field acquisition in a systematic digital format and optimized for a varied range of (spatial) analyses. DigiFract is developed using the programming interface of the Quantum Geographic Information System (GIS) with versatile functionality for spatial raster and vector data handling. Among other features, this includes spatial referencing of outcrop photos, and tools for digitizing geometries and assigning attribute information through a graphical user interface. While a GIS typically operates in map-view, DigiFract collects features on a surface of arbitrary orientation in 3D space. This surface is overlain with an outcrop photo and serves as reference frame for digitizing geologic features. Data is managed through a data model and stored in shapefiles or in a spatial database system. Fracture attributes, such as spacing or length, is intrinsic information of the digitized geometry and becomes explicit through follow-up data processing. Orientation statistics, scan-line or scan-window analyses can be performed from the graphical user interface or can be obtained through flexible Python scripts that directly access the fractdatamodel and analysisLib core modules of DigiFract. This workflow has been applied in various studies and enabled a faster collection of larger and more accurate fracture datasets. The studies delivered a better characterization of fractured reservoirs analogues in terms of fracture orientation and intensity distributions. Furthermore, the data organisation and analyses provided more independent constraints on the bed-confined or through-going nature of fractures relative to the stratigraphic layering.
NASA Astrophysics Data System (ADS)
Morton, A.; Stewart, R.; Held, E.; Piburn, J.; Allen, M. R.; McManamay, R.; Sanyal, J.; Sorokine, A.; Bhaduri, B. L.
2017-12-01
Spatiotemporal (ST) analytics applied to major spatio-temporal data sources from major vendors such as USGS, NOAA, World Bank and World Health Organization have tremendous value in shedding light on the evolution of physical, cultural, and geopolitical landscapes on a local and global level. Especially powerful is the integration of these physical and cultural datasets across multiple and disparate formats, facilitating new interdisciplinary analytics and insights. Realizing this potential first requires an ST data model that addresses challenges in properly merging data from multiple authors, with evolving ontological perspectives, semantical differences, changing attributes, and content that is textual, numeric, categorical, and hierarchical. Equally challenging is the development of analytical and visualization approaches that provide a serious exploration of this integrated data while remaining accessible to practitioners with varied backgrounds. The WSTAMP project at the Oak Ridge National Laboratory has yielded two major results in addressing these challenges: 1) development of the WSTAMP database, a significant advance in ST data modeling that integrates 16000+ attributes covering 200+ countries for over 50 years from over 30 major sources and 2) a novel online ST exploratory and analysis tool providing an array of modern statistical and visualization techniques for analyzing these data temporally, spatially, and spatiotemporally under a standard analytic workflow. We report on these advances, provide an illustrative case study, and inform how others may freely access the tool.
De la Torre, Fernando; Chu, Wen-Sheng; Xiong, Xuehan; Vicente, Francisco; Ding, Xiaoyu; Cohn, Jeffrey
2016-01-01
Within the last 20 years, there has been an increasing interest in the computer vision community in automated facial image analysis algorithms. This has been driven by applications in animation, market research, autonomous-driving, surveillance, and facial editing among others. To date, there exist several commercial packages for specific facial image analysis tasks such as facial expression recognition, facial attribute analysis or face tracking. However, free and easy-to-use software that incorporates all these functionalities is unavailable. This paper presents IntraFace (IF), a publicly-available software package for automated facial feature tracking, head pose estimation, facial attribute recognition, and facial expression analysis from video. In addition, IFincludes a newly develop technique for unsupervised synchrony detection to discover correlated facial behavior between two or more persons, a relatively unexplored problem in facial image analysis. In tests, IF achieved state-of-the-art results for emotion expression and action unit detection in three databases, FERA, CK+ and RU-FACS; measured audience reaction to a talk given by one of the authors; and discovered synchrony for smiling in videos of parent-infant interaction. IF is free of charge for academic use at http://www.humansensing.cs.cmu.edu/intraface/. PMID:27346987
Persky, Susan; Bouhlal, Sofia; Goldring, Megan R; McBride, Colleen M
2017-08-01
The development of precision approaches for customized health interventions is a promising application of genomic discovery. To optimize such weight management interventions, target audiences will need to be engaged in research and implementation efforts. Investigation into approaches that engage these audiences will be required to ensure that genomic information, particularly with respect to genomic influences on endophenotypes like eating behavior, is understood and accepted, and not associated with unintended adverse outcomes. We took steps to characterize healthy individuals' beliefs about genetic influences on eating behavior. Data were collected via online survey from 261 participants selected at random from a database. Respondents infrequently spontaneously identified eating behavior-related factors as running in families. However, those who perceived themselves as overweight and perceived a family history of overweight were more likely to attribute eating behavior to genetics on closed-ended assessments, β=0.252, p=0.039. Genetic attributions for eating behaviors were associated with lower confidence in ability to control eating and weight, β=-0.119, p=0.035. These exploratory findings shed light on beliefs about genetic influences on eating, a behavioral trait (rather than a disease). This investigation can inform future health intervention efforts. Published by Elsevier Ltd.
De la Torre, Fernando; Chu, Wen-Sheng; Xiong, Xuehan; Vicente, Francisco; Ding, Xiaoyu; Cohn, Jeffrey
2015-05-01
Within the last 20 years, there has been an increasing interest in the computer vision community in automated facial image analysis algorithms. This has been driven by applications in animation, market research, autonomous-driving, surveillance, and facial editing among others. To date, there exist several commercial packages for specific facial image analysis tasks such as facial expression recognition, facial attribute analysis or face tracking. However, free and easy-to-use software that incorporates all these functionalities is unavailable. This paper presents IntraFace (IF), a publicly-available software package for automated facial feature tracking, head pose estimation, facial attribute recognition, and facial expression analysis from video. In addition, IFincludes a newly develop technique for unsupervised synchrony detection to discover correlated facial behavior between two or more persons, a relatively unexplored problem in facial image analysis. In tests, IF achieved state-of-the-art results for emotion expression and action unit detection in three databases, FERA, CK+ and RU-FACS; measured audience reaction to a talk given by one of the authors; and discovered synchrony for smiling in videos of parent-infant interaction. IF is free of charge for academic use at http://www.humansensing.cs.cmu.edu/intraface/.
McMillan, Kimberly
2014-04-01
This paper aimed to explore the evolution of the concept of sustainability to facilitate further knowledge development in the discipline of nursing. The concept of 'sustainability' emerged in the 1950s as a result of the environmental movement. The concept has been adapted by the discipline of management and is increasingly discussed in the context of health care. The concept remains ambiguous in the discipline of nursing, resulting in a struggle to articulate the role of nursing in the sustainability movement. Rodgers evolutionary method of concept analysis was used. Literature was searched from 1987-2011, including English, peer reviewed texts in the databases CINAHL and ABI/INFORM global. Two book chapters and grey literature were also included. References were read and analysed according to antecedents, attributes, consequences, surrogate terms and related terms. Defining antecedents, attributes and consequences highlight the complexity and diversity of the concept. Attributes include: sustainability as a condition of change, as process, as outcome, as dependent of multiple stakeholders, and as social consciousness. 'Sustainability' is a fragile concept highly dependent on the processes and stakeholders involved in its fruition. There is a distinct difference in the level of concept clarity between the disciplines of management and nursing. The complexities associated with the concept of 'Sustainability' have led to its ambiguity. Nursing must, however, work to further clarify the concept to fully understand nursing's potential role in the sustainability movement. © 2013 John Wiley & Sons Ltd.
Mining moving object trajectories in location-based services for spatio-temporal database update
NASA Astrophysics Data System (ADS)
Guo, Danhuai; Cui, Weihong
2008-10-01
Advances in wireless transmission and mobile technology applied to LBS (Location-based Services) flood us with amounts of moving objects data. Vast amounts of gathered data from position sensors of mobile phones, PDAs, or vehicles hide interesting and valuable knowledge and describe the behavior of moving objects. The correlation between temporal moving patterns of moving objects and geo-feature spatio-temporal attribute was ignored, and the value of spatio-temporal trajectory data was not fully exploited too. Urban expanding or frequent town plan change bring about a large amount of outdated or imprecise data in spatial database of LBS, and they cannot be updated timely and efficiently by manual processing. In this paper we introduce a data mining approach to movement pattern extraction of moving objects, build a model to describe the relationship between movement patterns of LBS mobile objects and their environment, and put up with a spatio-temporal database update strategy in LBS database based on trajectories spatiotemporal mining. Experimental evaluation reveals excellent performance of the proposed model and strategy. Our original contribution include formulation of model of interaction between trajectory and its environment, design of spatio-temporal database update strategy based on moving objects data mining, and the experimental application of spatio-temporal database update by mining moving objects trajectories.
PlantRGDB: A Database of Plant Retrocopied Genes.
Wang, Yi
2017-01-01
RNA-based gene duplication, known as retrocopy, plays important roles in gene origination and genome evolution. The genomes of many plants have been sequenced, offering an opportunity to annotate and mine the retrocopies in plant genomes. However, comprehensive and unified annotation of retrocopies in these plants is still lacking. In this study I constructed the PlantRGDB (Plant Retrocopied Gene DataBase), the first database of plant retrocopies, to provide a putatively complete centralized list of retrocopies in plant genomes. The database is freely accessible at http://probes.pw.usda.gov/plantrgdb or http://aegilops.wheat.ucdavis.edu/plantrgdb. It currently integrates 49 plant species and 38,997 retrocopies along with characterization information. PlantRGDB provides a user-friendly web interface for searching, browsing and downloading the retrocopies in the database. PlantRGDB also offers graphical viewer-integrated sequence information for displaying the structure of each retrocopy. The attributes of the retrocopies of each species are reported using a browse function. In addition, useful tools, such as an advanced search and BLAST, are available to search the database more conveniently. In conclusion, the database will provide a web platform for obtaining valuable insight into the generation of retrocopies and will supplement research on gene duplication and genome evolution in plants. © The Author 2017. Published by Oxford University Press on behalf of Japanese Society of Plant Physiologists. All rights reserved. For permissions, please email: journals.permissions@oup.com.
The Universal Protein Resource (UniProt): an expanding universe of protein information.
Wu, Cathy H; Apweiler, Rolf; Bairoch, Amos; Natale, Darren A; Barker, Winona C; Boeckmann, Brigitte; Ferro, Serenella; Gasteiger, Elisabeth; Huang, Hongzhan; Lopez, Rodrigo; Magrane, Michele; Martin, Maria J; Mazumder, Raja; O'Donovan, Claire; Redaschi, Nicole; Suzek, Baris
2006-01-01
The Universal Protein Resource (UniProt) provides a central resource on protein sequences and functional annotation with three database components, each addressing a key need in protein bioinformatics. The UniProt Knowledgebase (UniProtKB), comprising the manually annotated UniProtKB/Swiss-Prot section and the automatically annotated UniProtKB/TrEMBL section, is the preeminent storehouse of protein annotation. The extensive cross-references, functional and feature annotations and literature-based evidence attribution enable scientists to analyse proteins and query across databases. The UniProt Reference Clusters (UniRef) speed similarity searches via sequence space compression by merging sequences that are 100% (UniRef100), 90% (UniRef90) or 50% (UniRef50) identical. Finally, the UniProt Archive (UniParc) stores all publicly available protein sequences, containing the history of sequence data with links to the source databases. UniProt databases continue to grow in size and in availability of information. Recent and upcoming changes to database contents, formats, controlled vocabularies and services are described. New download availability includes all major releases of UniProtKB, sequence collections by taxonomic division and complete proteomes. A bibliography mapping service has been added, and an ID mapping service will be available soon. UniProt databases can be accessed online at http://www.uniprot.org or downloaded at ftp://ftp.uniprot.org/pub/databases/.
The Gamma-Ray Burst ToolSHED is Open for Business
NASA Astrophysics Data System (ADS)
Giblin, Timothy W.; Hakkila, Jon; Haglin, David J.; Roiger, Richard J.
2004-09-01
The GRB ToolSHED, a Gamma-Ray Burst SHell for Expeditions in Data-Mining, is now online and available via a web browser to all in the scientific community. The ToolSHED is an online web utility that contains pre-processed burst attributes of the BATSE catalog and a suite of induction-based machine learning and statistical tools for classification and cluster analysis. Users create their own login account and study burst properties within user-defined multi-dimensional parameter spaces. Although new GRB attributes are periodically added to the database for user selection, the ToolSHED has a feature that allows users to upload their own burst attributes (e.g. spectral parameters, etc.) so that additional parameter spaces can be explored. A data visualization feature using GNUplot and web-based IDL has also been implemented to provide interactive plotting of user-selected session output. In an era in which GRB observations and attributes are becoming increasingly more complex, a utility such as the GRB ToolSHED may play an important role in deciphering GRB classes and understanding intrinsic burst properties.
NASA Astrophysics Data System (ADS)
Fussi, Fabio; Bonomi, Tullia; Fava, Francesco; Hamidou, Barry; Hamidou Khane, Cheikh; Faye, Gayane; Wade, Souleye; Colombo, Roberto
2014-05-01
Background In order to increase access to drinking water in Africa there is more and more interest in the promotion of manual drilling techniques, without need of expensive drilling equipment, but they can be applied only in those areas with suitable hydrogeological conditions: thick layers of unconsolidated sediments and shallow groundwater level. Mapping of suitable zones for manual drilling at national level in Africa is a crucial activity and local institutions and UNICEF are implementing specific programs for its promotion, but the limitation in available data concerning shallow hydrogeological aquifers are limited. The research has been developed in the project "Use of remote sensing and terrain modeling to identify suitable zones for manual drilling in Africa and support low cost water supply", within the scientific cooperation between the University of Milano-Bicocca, Universite' Cheick Anta Diop (Dakar Senegal) , SNAPE - Service Nationale de Points d'Eau (Conakry Guinea), UNICEF Senegal and UNICEF Guinea. The project is funded by NERC (National Environmental Research Council, UK). Objective of the research: The presented work is only the starting point of the project aiming to elaborate an automatic procedures to manage and improve the existing database of borehole logs in Senegal and Guinea for the interpretation of shallow hydrogeological conditions and identification of suitable zones for manual drilling, in two pilot areas: Louga (Northwestern Senegal) and Faranah/Kankan (Eastern Guinea). Within the objective of the project is also considered the integration of Remote Sensing to support hydrogeological interpretation, especially where borehole logs are not present. Methodology Focus is to create a hydrogeological database, TANGAFRIC, to organize, codify and elaborate hydrogeological data. The metodology derives from the software TANGRAM (www.tangram.samit.unimib.it) produced by the University of Milano Bicocca, with innovative aspect of stratigraphic data codification, quantification and processing, connected to a hydraulic conductivity value associated to each primary lithology. Results Starting from the database of borehole logs available at national level in Senegal and Guinea (about 1400 borehole logs in Senegal and 800 in Guinea, with 20000 definitions), their structure and information have been compared and a new common database has been set up; it has a consistent structure with the structure of existing national database and data can be easily imported and exported. From this joint, the new software TANGAFRIC has been created with different purposes: -to organize in the same way wells data, since the two countries have different administrative divisions (ID code, name of village, district, regions, coordinates); -to add new wells data, not existing in the previous databases; -to codify the stratigraphic layer of each well logs with a 5-digit alphanumeric codes, using a list of categories describing texture, status and color or each layers, identified from the most recurrent lithological classes and attributes; -to attribute a specific value of hydraulic conductivity to each texture, from well data, field pumping test, bibliographic review. TANGAFRIC includes one module for data input and a second module to process the data, and extract specific parameters concerning mean texture, hydraulic conductivity and transmissivity in selected depth ranges. This is made possible by attributing a weight to the digits of the code for textures. The program calculates the percentage of the chosen lithology, as related to each individual layer, and also a weighted average of hydraulic conductivity. It has been possible to produce maps showing the distribution of main texture classes, thickness of saturated unconsolidated sediments and expected transmissivity. Furthermore, these parameters have been used to estimate the suitability for manual drilling under the hydrogeological coniditions described in each borehole logs.
Knowlton, Michelle N; Li, Tongbin; Ren, Yongliang; Bill, Brent R; Ellis, Lynda Bm; Ekker, Stephen C
2008-01-07
The zebrafish is a powerful model vertebrate amenable to high throughput in vivo genetic analyses. Examples include reverse genetic screens using morpholino knockdown, expression-based screening using enhancer trapping and forward genetic screening using transposon insertional mutagenesis. We have created a database to facilitate web-based distribution of data from such genetic studies. The MOrpholino DataBase is a MySQL relational database with an online, PHP interface. Multiple quality control levels allow differential access to data in raw and finished formats. MODBv1 includes sequence information relating to almost 800 morpholinos and their targets and phenotypic data regarding the dose effect of each morpholino (mortality, toxicity and defects). To improve the searchability of this database, we have incorporated a fixed-vocabulary defect ontology that allows for the organization of morpholino affects based on anatomical structure affected and defect produced. This also allows comparison between species utilizing Phenotypic Attribute Trait Ontology (PATO) designated terminology. MODB is also cross-linked with ZFIN, allowing full searches between the two databases. MODB offers users the ability to retrieve morpholino data by sequence of morpholino or target, name of target, anatomical structure affected and defect produced. MODB data can be used for functional genomic analysis of morpholino design to maximize efficacy and minimize toxicity. MODB also serves as a template for future sequence-based functional genetic screen databases, and it is currently being used as a model for the creation of a mutagenic insertional transposon database.
Metadata mapping and reuse in caBIG™
Kunz, Isaac; Lin, Ming-Chin; Frey, Lewis
2009-01-01
Background This paper proposes that interoperability across biomedical databases can be improved by utilizing a repository of Common Data Elements (CDEs), UML model class-attributes and simple lexical algorithms to facilitate the building domain models. This is examined in the context of an existing system, the National Cancer Institute (NCI)'s cancer Biomedical Informatics Grid (caBIG™). The goal is to demonstrate the deployment of open source tools that can be used to effectively map models and enable the reuse of existing information objects and CDEs in the development of new models for translational research applications. This effort is intended to help developers reuse appropriate CDEs to enable interoperability of their systems when developing within the caBIG™ framework or other frameworks that use metadata repositories. Results The Dice (di-grams) and Dynamic algorithms are compared and both algorithms have similar performance matching UML model class-attributes to CDE class object-property pairs. With algorithms used, the baselines for automatically finding the matches are reasonable for the data models examined. It suggests that automatic mapping of UML models and CDEs is feasible within the caBIG™ framework and potentially any framework that uses a metadata repository. Conclusion This work opens up the possibility of using mapping algorithms to reduce cost and time required to map local data models to a reference data model such as those used within caBIG™. This effort contributes to facilitating the development of interoperable systems within caBIG™ as well as other metadata frameworks. Such efforts are critical to address the need to develop systems to handle enormous amounts of diverse data that can be leveraged from new biomedical methodologies. PMID:19208192
NASA Astrophysics Data System (ADS)
Kornhuber, Kai; Rybski, Diego; Costa, Luis; Reusser, Dominik E.; Kropp, Jürgen P.
2014-05-01
The Environmental Kuznets Curves (EKC) postulates that pollution increases with the income per capita up to a maximum, above which it decreases with the further increase in income per capita, i.e. following an inverse U-shape in the pollution vs. income per capita. It is commonly believed that EKC occurs for "local" pollutants such as nitrogen oxide and sulfur dioxide, but does not hold for CO2 emissions. This is attributed to the fact that while "local" pollutants cause a visible environmental damage on the local/regional scale (which authorities/governments seek to avoid), the consequences of CO2 emission have no immediate attributable local/regional consequences. We review EKC for CO2 exploring its relation between CO2 per capita and the Human Development Index (HDI) between 1990 and 2010 obtained from the World Bank database. We find evidence for a reduction in CO2 emissions per capita in highly developed countries. We propose a model according to which the emissions per capita of a country are composed of a component related to the actual state of development and a component related to the change of development. The model leads to four distinct cases of which two have EKC shape and two imply saturation. This outcome is in line with previously suggested qualitative relations. Our analysis indicates that the EKC shaped cases better describes the empirical values. We explore the less extreme version corresponding to the so-called conventional EKC and study the maximum of the fitted curve, providing a threshold-value for the HDI and a typical maximum value for the emissions per capita. We find that approx. 5 countries have crossed the CO2-HDI maximum, corresponding to approx. 1.5% of the world population.
NASA Astrophysics Data System (ADS)
Susanto, Arif; Mulyono, Nur Budi
2018-02-01
The changes of environmental management system standards into the latest version, i.e. ISO 14001:2015, may cause a change on a data and information need in decision making and achieving the objectives in the organization coverage. Information management is the organization's responsibility to ensure that effectiveness and efficiency start from its creating, storing, processing and distribution processes to support operations and effective decision making activity in environmental performance management. The objective of this research was to set up an information management program and to adopt the technology as the supporting component of the program which was done by PTFI Concentrating Division so that it could be in line with the desirable organization objective in environmental management based on ISO 14001:2015 environmental management system standards. Materials and methods used covered technical aspects in information management, i.e. with web-based application development by using usage centered design. The result of this research showed that the use of Single Sign On gave ease to its user to interact further on the use of the environmental management system. Developing a web-based through creating entity relationship diagram (ERD) and information extraction by conducting information extraction which focuses on attributes, keys, determination of constraints. While creating ERD is obtained from relational database scheme from a number of database from environmental performances in Concentrating Division.
Introducing the GRACEnet/REAP Data Contribution, Discovery, and Retrieval System.
Del Grosso, S J; White, J W; Wilson, G; Vandenberg, B; Karlen, D L; Follett, R F; Johnson, J M F; Franzluebbers, A J; Archer, D W; Gollany, H T; Liebig, M A; Ascough, J; Reyes-Fox, M; Pellack, L; Starr, J; Barbour, N; Polumsky, R W; Gutwein, M; James, D
2013-07-01
Difficulties in accessing high-quality data on trace gas fluxes and performance of bioenergy/bioproduct feedstocks limit the ability of researchers and others to address environmental impacts of agriculture and the potential to produce feedstocks. To address those needs, the GRACEnet (Greenhouse gas Reduction through Agricultural Carbon Enhancement network) and REAP (Renewable Energy Assessment Project) research programs were initiated by the USDA Agricultural Research Service (ARS). A major product of these programs is the creation of a database with greenhouse gas fluxes, soil carbon stocks, biomass yield, nutrient, and energy characteristics, and input data for modeling cropped and grazed systems. The data include site descriptors (e.g., weather, soil class, spatial attributes), experimental design (e.g., factors manipulated, measurements performed, plot layouts), management information (e.g., planting and harvesting schedules, fertilizer types and amounts, biomass harvested, grazing intensity), and measurements (e.g., soil C and N stocks, plant biomass amount and chemical composition). To promote standardization of data and ensure that experiments were fully described, sampling protocols and a spreadsheet-based data-entry template were developed. Data were first uploaded to a temporary database for checking and then were uploaded to the central database. A Web-accessible application allows for registered users to query and download data including measurement protocols. Separate portals have been provided for each project (GRACEnet and REAP) at nrrc.ars.usda.gov/slgracenet/#/Home and nrrc.ars.usda.gov/slreap/#/Home. The database architecture and data entry template have proven flexible and robust for describing a wide range of field experiments and thus appear suitable for other natural resource research projects. Copyright © by the American Society of Agronomy, Crop Science Society of America, and Soil Science Society of America, Inc.
Buczkowski, Brian J.; Reid, Jane A.; Jenkins, Chris J.; Reid, Jamey M.; Williams, S. Jeffress; Flocks, James G.
2006-01-01
Over the past 50 years there has been an explosion in scientific interest, research effort and information gathered on the geologic sedimentary character of the United States continental margin. Data and information from thousands of publications have greatly increased our scientific understanding of the geologic origins of the shelf surface but rarely have those data been combined and integrated. This publication is the first release of the Gulf of Mexico and Caribbean (Puerto Rico and U.S. Virgin Islands) coastal and offshore data from the usSEABED database. The report contains a compilation of published and previously unpublished sediment texture and other geologic data about the sea floor from diverse sources. usSEABED is an innovative database system developed to bring assorted data together in a unified database. The dbSEABED system is used to process the data. Examples of maps displaying attributes such as grain size and sediment color are included. This database contains information that is a scientific foundation for the USGS Marine Aggregate Resources and Processes Assessment and Benthic Habitats projects, and will be useful to the marine science community for other studies of the Gulf of Mexico and Caribbean continental margins. This publication is divided into ten sections: Home, Introduction, Content, usSEABED (data), dbSEABED (processing), Data Catalog, References, Contacts, Acknowledgments and Frequently Asked Questions. Use the navigation bar on the left to navigate to specific sections of this report. Underlined topics throughout the publication are links to more information. Links to specific and detailed information on processing and those to pages outside this report will open in a new browser window.
Inequality of obesity and socioeconomic factors in Iran: a systematic review and meta- analyses
Djalalinia, Shirin; Peykari, Niloofar; Qorbani, Mostafa; Larijani, Bagher; Farzadfar, Farshad
2015-01-01
Background: Socioeconomic status and demographic factors, such as education, occupation, place of residence, gender, age, and marital status have been reported to be associated with obesity. We conducted a systematic review to summarize evidences on associations between socioeconomic factors and obesity/overweight in Iranian population. Methods: We systematically searched international databases; ISI, PubMed/Medline, Scopus, and national databases Iran-medex, Irandoc, and Scientific Information Database (SID). We refined data for associations between socioeconomic factors and obesity/overweight by sex, age, province, and year. There were no limitations for time and languages. Results: Based on our search strategy we found 151 records; of them 139 were from international databases and the remaining 12 were obtained from national databases. After removing duplicates, via the refining steps, only 119 articles were found related to our study domains. Extracted results were attributed to 146596 person/data from included studies. Increased ages, low educational levels, being married, residence in urban area, as well as female sex were clearly associated with obesity. Conclusion: Results could be useful for better health policy and more planned studies in this field. These also could be used for future complementary analyses. PMID:26793632
Inequality of obesity and socioeconomic factors in Iran: a systematic review and meta- analyses.
Djalalinia, Shirin; Peykari, Niloofar; Qorbani, Mostafa; Larijani, Bagher; Farzadfar, Farshad
2015-01-01
Socioeconomic status and demographic factors, such as education, occupation, place of residence, gender, age, and marital status have been reported to be associated with obesity. We conducted a systematic review to summarize evidences on associations between socioeconomic factors and obesity/overweight in Iranian population. We systematically searched international databases; ISI, PubMed/Medline, Scopus, and national databases Iran-medex, Irandoc, and Scientific Information Database (SID). We refined data for associations between socioeconomic factors and obesity/overweight by sex, age, province, and year. There were no limitations for time and languages. Based on our search strategy we found 151 records; of them 139 were from international databases and the remaining 12 were obtained from national databases. After removing duplicates, via the refining steps, only 119 articles were found related to our study domains. Extracted results were attributed to 146596 person/data from included studies. Increased ages, low educational levels, being married, residence in urban area, as well as female sex were clearly associated with obesity. RESULTS could be useful for better health policy and more planned studies in this field. These also could be used for future complementary analyses.
Damage to offshore infrastructure in the Gulf of Mexico by hurricanes Katrina and Rita
NASA Astrophysics Data System (ADS)
Cruz, A. M.; Krausmann, E.
2009-04-01
The damage inflicted by hurricanes Katrina and Rita to the Gulf-of-Mexico's (GoM) oil and gas production, both onshore and offshore, has shown the proneness of industry to Natech accidents (natural hazard-triggered hazardous-materials releases). In order to contribute towards a better understanding of Natech events, we assessed the damage to and hazardous-materials releases from offshore oil and natural-gas platforms and pipelines induced by hurricanes Katrina and Rita. Data was obtained through a review of published literature and interviews with government officials and industry representatives from the affected region. We also reviewed over 60,000 records of reported hazardous-materials releases from the National Response Center's (NRC) database to identify and analyze the hazardous-materials releases directly attributed to offshore oil and gas platforms and pipelines affected by the two hurricanes. Our results show that hurricanes Katrina and Rita destroyed at least 113 platforms, and severely damaged at least 53 others. Sixty percent of the facilities destroyed were built 30 years ago or more prior to the adoption of the more stringent design standards that went into effect after 1977. The storms also destroyed 5 drilling rigs and severely damaged 19 mobile offshore drilling units (MODUs). Some 19 MODUs lost their moorings and became adrift during the storms which not only posed a danger to existing facilities but the dragging anchors also damaged pipelines and other infrastructure. Structural damage to platforms included toppling of sections, and tilting or leaning of platforms. Possible causes for failure of structural and non-structural components of platforms included loading caused by wave inundation of the deck. Failure of rigs attached to platforms was also observed resulting in significant damage to the platform or adjacent infrastructure, as well as damage to equipment, living quarters and helipads. The failures are attributable to tie-down components and occurred on both fixed and floating platforms. The total number of pipelines damaged by Hurricanes Katrina and Rita as of May 1, 2006, was 457. Pipeline damage was mostly caused by damage or failure of the host platform or its development and production piping, the impact of dragging and displaced objects, and pipeline interaction at a crossing. Damage to pipelines was a major contributing factor in delaying start up of offshore oil and gas production. During our analysis of the NRC database we identified 611 reported hazardous-materials releases directly attributed to offshore platforms and pipelines affected by the two hurricanes. There were twice as many releases during Hurricane Katrina than during Rita; 80% or more of the releases reported in the NRC database occurred from platforms. Our analysis suggests that the majority of releases were petroleum products, such as crude oil and condensate, followed by natural gas. In both Katrina and Rita, releases were more likely in the front, right quadrant of the storm. Storm-surge values were highest closer to the coastline. This may help explain the higher number of releases in shallow waters. The higher number of hazardous-materials releases from platforms during Katrina may partly be attributed to the higher wind speeds for this storm as it approached land.
García-Jiménez, Beatriz; Pons, Tirso; Sanchis, Araceli; Valencia, Alfonso
2014-01-01
Biological pathways are important elements of systems biology and in the past decade, an increasing number of pathway databases have been set up to document the growing understanding of complex cellular processes. Although more genome-sequence data are becoming available, a large fraction of it remains functionally uncharacterized. Thus, it is important to be able to predict the mapping of poorly annotated proteins to original pathway models. We have developed a Relational Learning-based Extension (RLE) system to investigate pathway membership through a function prediction approach that mainly relies on combinations of simple properties attributed to each protein. RLE searches for proteins with molecular similarities to specific pathway components. Using RLE, we associated 383 uncharacterized proteins to 28 pre-defined human Reactome pathways, demonstrating relative confidence after proper evaluation. Indeed, in specific cases manual inspection of the database annotations and the related literature supported the proposed classifications. Examples of possible additional components of the Electron transport system, Telomere maintenance and Integrin cell surface interactions pathways are discussed in detail. All the human predicted proteins in the 2009 and 2012 releases 30 and 40 of Reactome are available at http://rle.bioinfo.cnio.es.
An Ontology-Based GIS for Genomic Data Management of Rumen Microbes
Jelokhani-Niaraki, Saber; Minuchehr, Zarrin; Nassiri, Mohammad Reza
2015-01-01
During recent years, there has been exponential growth in biological information. With the emergence of large datasets in biology, life scientists are encountering bottlenecks in handling the biological data. This study presents an integrated geographic information system (GIS)-ontology application for handling microbial genome data. The application uses a linear referencing technique as one of the GIS functionalities to represent genes as linear events on the genome layer, where users can define/change the attributes of genes in an event table and interactively see the gene events on a genome layer. Our application adopted ontology to portray and store genomic data in a semantic framework, which facilitates data-sharing among biology domains, applications, and experts. The application was developed in two steps. In the first step, the genome annotated data were prepared and stored in a MySQL database. The second step involved the connection of the database to both ArcGIS and Protégé as the GIS engine and ontology platform, respectively. We have designed this application specifically to manage the genome-annotated data of rumen microbial populations. Such a GIS-ontology application offers powerful capabilities for visualizing, managing, reusing, sharing, and querying genome-related data. PMID:25873847
An Ontology-Based GIS for Genomic Data Management of Rumen Microbes.
Jelokhani-Niaraki, Saber; Tahmoorespur, Mojtaba; Minuchehr, Zarrin; Nassiri, Mohammad Reza
2015-03-01
During recent years, there has been exponential growth in biological information. With the emergence of large datasets in biology, life scientists are encountering bottlenecks in handling the biological data. This study presents an integrated geographic information system (GIS)-ontology application for handling microbial genome data. The application uses a linear referencing technique as one of the GIS functionalities to represent genes as linear events on the genome layer, where users can define/change the attributes of genes in an event table and interactively see the gene events on a genome layer. Our application adopted ontology to portray and store genomic data in a semantic framework, which facilitates data-sharing among biology domains, applications, and experts. The application was developed in two steps. In the first step, the genome annotated data were prepared and stored in a MySQL database. The second step involved the connection of the database to both ArcGIS and Protégé as the GIS engine and ontology platform, respectively. We have designed this application specifically to manage the genome-annotated data of rumen microbial populations. Such a GIS-ontology application offers powerful capabilities for visualizing, managing, reusing, sharing, and querying genome-related data.
Learning concepts of cinenurducation: an integrative review.
Oh, Jina; Kang, Jeongae; De Gagne, Jennie C
2012-11-01
Cinenurducation is the use of films in both didactic and clinical nursing education. Although films are already used as instructional aids in nursing education, few studies have been made that demonstrate the learning concepts that can be attributed to this particular teaching strategy. The purpose of this paper is to describe the learning concepts of cinenurducation and its conceptual metaphor based on a review of literature. The databases CINAHL, MEDLINE, PsychINFO, ERIC, EBSCO, ProQuest Library Journal, and Scopus databases were searched for articles. Fifteen peer-reviewed articles were selected through title and abstract screening from "films in nursing" related articles found in internationally published articles in English from the past 20 years. Four common concepts emerged that relate to cinenurducation: (a) student-centered, (b) experiential, (c) reflective, and (d) problem-solving learning. Current literature corroborates cinenurducation as an effective teaching strategy with its learning activities in nursing education. Future studies may include instructional guides of sample films that could be practically used in various domains to teach nursing competencies, as well as in the development of evaluation criteria and standards to assess students' learning outcomes. Copyright © 2012 Elsevier Ltd. All rights reserved.
Abrahams, Kristen; Harty, Michal; St Louis, Kenneth O; Thabane, Lehana; Kathard, Harsha
2016-07-27
As teachers form an important part of the intervention process with childrenwho stutter in primary school, the primary aim was to describe primary school teachers'attitudes in South Africa. The secondary aim was to compare teachers' attitudes towardsstuttering in South Africa with those from a pooled group of respondents in the Public OpinionSurvey of Human Attributes-Stuttering (POSHA-S) database from different countries collectedin 2009-2014. A quantitative, cross-sectional survey research design was used. Primary schools intwo education districts in Western Cape, South Africa, were sampled. The POSHA-S, a selfadministeredquestionnaire, was completed by a cluster sample of 469 participants. Overall positive attitudes towards stuttering were found, specifically related to thepotential of people who stutter, although the result should be interpreted with caution as thesample was not homogenously positive. Teachers still had misconceptions about personalitystereotypes and the cause of stuttering. The attitudes of the South African sample were slightlymore positive compared with the samples in the current POSHA-S database. When developing stuttering intervention strategies, there are a number of keyconsiderations to take into account. The study provides a basis for speech-language therapiststo think about intervention with teachers and which areas of stuttering to consider.
The burden of disease attributable to cannabis use in Canada in 2012.
Imtiaz, Sameer; Shield, Kevin D; Roerecke, Michael; Cheng, Joyce; Popova, Svetlana; Kurdyak, Paul; Fischer, Benedikt; Rehm, Jürgen
2016-04-01
Cannabis use is associated with several adverse health effects. However, little is known about the cannabis-attributable burden of disease. This study quantified the age-, sex- and adverse health effect-specific cannabis-attributable (1) mortality, (2) years of life lost due to premature mortality (YLLs), (3) years of life lost due to disability (YLDs) and (4) disability-adjusted life years (DALYs) in Canada in 2012. Epidemiological modeling. Canada. Canadians aged ≥ 15 years in 2012. Using comparative risk assessment methodology, cannabis-attributable fractions were computed using Canadian exposure data and risk relations from large studies or meta-analyses. Outcome data were obtained from Canadian databases and the World Health Organization. The 95% confidence intervals (CIs) were computed using Monte Carlo methodology. Cannabis use was estimated to have caused 287 deaths (95% CI = 108, 609), 10,533 YLLs (95% CI = 4760, 20,833), 55,813 YLDs (95% CI = 38,175, 74,094) and 66,346 DALYs (95% CI = 47,785, 87,207), based on causal impacts on cannabis use disorders, schizophrenia, lung cancer and road traffic injuries. Cannabis-attributable burden of disease was highest among young people, and males accounted for twice the burden than females. Cannabis use disorders were the most important single cause of the cannabis-attributable burden of disease. The cannabis-attributable burden of disease in Canada in 2012 included 55,813 years of life lost due to disability, caused mainly by cannabis use disorders. Although the cannabis-attributable burden of disease was substantial, it was much lower compared with other commonly used legal and illegal substances. Moreover, the evidence base for cannabis-attributable harms was smaller. © 2015 Society for the Study of Addiction.
Sepeda, Adrian L
2006-03-17
Learning from the experiences of others has long been recognized as a valued and relatively painless process. In the world of process safety, this learning method is an essential tool since industry has neither the time and resources nor the willingness to experience an incident before taking corrective or preventative steps. This paper examines the need for and value of process safety incident databases that collect incidents of high learning value and structure them so that needed information can be easily and quickly extracted. It also explores how they might be used to prevent incidents by increasing awareness and by being a tool for conducting PHAs and incident investigations. The paper then discusses how the CCPS PSID meets those requirements, how PSID is structured and managed, and its attributes and features.
The role of authority power in explaining procedural fairness effects.
van Dijke, Marius; De Cremer, David; Mayer, David M
2010-05-01
Building on fairness heuristic theory, fairness theory, and trust development models, we argue that unfairly enacted procedures decrease followers' trust in the authority particularly when authorities have high power over their followers. Moreover, we expected trust to mediate procedural fairness effects on followers' attitudes (authorities' legitimacy and charisma attributed to authorities) and organizational citizenship behavior. Procedural fairness effects on these variables, as mediated by trust, should therefore also be stronger when authority power is high. The results of a single- and multisource field study and a laboratory experiment supported these predictions. These studies support the role of authority power as a theoretically and practically relevant moderator of procedural fairness effects and show that its effectiveness is explained through trust in authorities. PsycINFO Database Record (c) 2010 APA, all rights reserved.
Exploring vegetation in the fourth dimension.
Mitchell, Fraser J G
2011-01-01
Much ecological research focuses on changes in vegetation on spatial scales from stands to landscapes; however, capturing data on vegetation change over relevant timescales remains a challenge. Pollen analysis offers unrivalled access to data with global coverage over long timescales. Robust techniques have now been developed that enable pollen data to be converted into vegetation data in terms of individual taxa, plant communities or biomes, with the possibility of deriving from those data a range of plant attributes and ecological indicators. In this review, I discuss how coupling pollen with macrofossil, charcoal and genetic data opens up the extensive pollen databases to investigation of the drivers of vegetation change over time and also provides extensive data sets for testing hypotheses with wide ecological relevance. © 2010 Elsevier Ltd. All rights reserved.
Integration of Multidisciplinary Sensory Data:
Miller, Perry L.; Nadkarni, Prakash; Singer, Michael; Marenco, Luis; Hines, Michael; Shepherd, Gordon
2001-01-01
The paper provides an overview of neuroinformatics research at Yale University being performed as part of the national Human Brain Project. This research is exploring the integration of multidisciplinary sensory data, using the olfactory system as a model domain. The neuroinformatics activities fall into three main areas: 1) building databases and related tools that support experimental olfactory research at Yale and can also serve as resources for the field as a whole, 2) using computer models (molecular models and neuronal models) to help understand data being collected experimentally and to help guide further laboratory experiments, 3) performing basic neuroinformatics research to develop new informatics technologies, including a flexible data model (EAV/CR, entity-attribute-value with classes and relationships) designed to facilitate the integration of diverse heterogeneous data within a single unifying framework. PMID:11141511
Li, Shijun; Ehrhardt, David W.; Rhee, Seung Y.
2006-01-01
Cells are organized into a complex network of subcellular compartments that are specialized for various biological functions. Subcellular location is an important attribute of protein function. To facilitate systematic elucidation of protein subcellular location, we analyzed experimentally verified protein localization data of 1,300 Arabidopsis (Arabidopsis thaliana) proteins. The 1,300 experimentally verified proteins are distributed among 40 different compartments, with most of the proteins localized to four compartments: mitochondria (36%), nucleus (28%), plastid (17%), and cytosol (13.3%). About 19% of the proteins are found in multiple compartments, in which a high proportion (36.4%) is localized to both cytosol and nucleus. Characterization of the overrepresented Gene Ontology molecular functions and biological processes suggests that the Golgi apparatus and peroxisome may play more diverse functions but are involved in more specialized processes than other compartments. To support systematic empirical determination of protein subcellular localization using a technology called fluorescent tagging of full-length proteins, we developed a database and Web application to provide preselected green fluorescent protein insertion position and primer sequences for all Arabidopsis proteins to study their subcellular localization and to store experimentally verified protein localization images, videos, and their annotations of proteins generated using the fluorescent tagging of full-length proteins technology. The database can be searched, browsed, and downloaded using a Web browser at http://aztec.stanford.edu/gfp/. The software can also be downloaded from the same Web site for local installation. PMID:16617091
Concept analysis of nurses' happiness.
Ozkara San, Eda
2015-01-01
The purpose of this analysis is to examine and clarify the concept of nurses' happiness (NH), understand the different uses of the concept, explore the conditions that foster it, and consider the consequences of NH, including the phenomena that emerge as a result of NH occurrence. The author utilizes Walker and Avant's eight-stage concept analysis. Computer and manual searches were conducted of articles in the English language addressing NH from 1990 to present. EBSCO and PubMed are the electronic databases used to access literature for this paper. For both databases, the researcher has examined this new term by splitting the term nurses' happiness into its two root words, namely nurses and happiness. An inductive analysis of articles produced descriptive themes. Definitions of happiness and NH are analyzed. Antecedents, attributes, and consequences of NH are described. Model, borderline, contrary, and related cases for NH are also identified. This concept analysis helps in the understanding of the definition of NH, the attributes that contribute to the occurrence of NH in clinical practice, as well as the consequences of NH, and how it should be measured from a nursing perspective. Ozkara San. © 2014 Wiley Periodicals, Inc.
Climate Signals: An On-Line Digital Platform for Mapping Climate Change Impacts in Real Time
NASA Astrophysics Data System (ADS)
Cutting, H.
2016-12-01
Climate Signals is an on-line digital platform for cataloging and mapping the impacts of climate change. The CS platform specifies and details the chains of connections between greenhouse gas emissions and individual climate events. Currently in open-beta release, the platform is designed to to engage and serve the general public, news media, and policy-makers, particularly in real-time during extreme climate events. Climate Signals consists of a curated relational database of events and their links to climate change, a mapping engine, and a gallery of climate change monitors offering real-time data. For each event in the database, an infographic engine provides a custom attribution "tree" that illustrates the connections to climate change. In addition, links to key contextual resources are aggregated and curated for each event. All event records are fully annotated with detailed source citations and corresponding hyper links. The system of attribution used to link events to climate change in real-time is detailed here. This open-beta release is offered for public user testing and engagement. Launched in May 2016, the operation of this platform offers lessons for public engagement in climate change impacts.
Reported load carriage injuries of the Australian army soldier.
Orr, Robin M; Johnston, Venerina; Coyle, Julia; Pope, Rodney
2015-06-01
Many injuries experienced by soldiers can be attributed to the occupational loads they are required to carry. The aim of this study was to determine whether contemporary military load carriage is a source of injuries to Australian Regular Army soldiers and to profile these injuries. The Australian Defence Force 'Occupational Health, Safety and Compensation Analysis and Reporting' database was searched to identify all reported injuries sustained during load carriage events. Key search terms were employed and narrative description fields were interrogated to increase data accuracy. A total of 1,954 injury records were extracted from the database. Of these, 404 injuries were attributed to load carriage. The majority of these load carriage injuries involved either the lower limb or back, with bones and joints accounting for the most frequently reported body structures to be injured. Field activities were the leading activities being performed at the time that load carriage injuries occurred, and muscular stress was identified as the mechanism of injury for over half of reported load carriage injuries. This study suggests that load carriage is a substantial source of injury risk to Australian Army soldiers. Physical training may fail to adequately prepare soldiers for load carriage tasks during field training exercises.
Pathological speech signal analysis and classification using empirical mode decomposition.
Kaleem, Muhammad; Ghoraani, Behnaz; Guergachi, Aziz; Krishnan, Sridhar
2013-07-01
Automated classification of normal and pathological speech signals can provide an objective and accurate mechanism for pathological speech diagnosis, and is an active area of research. A large part of this research is based on analysis of acoustic measures extracted from sustained vowels. However, sustained vowels do not reflect real-world attributes of voice as effectively as continuous speech, which can take into account important attributes of speech such as rapid voice onset and termination, changes in voice frequency and amplitude, and sudden discontinuities in speech. This paper presents a methodology based on empirical mode decomposition (EMD) for classification of continuous normal and pathological speech signals obtained from a well-known database. EMD is used to decompose randomly chosen portions of speech signals into intrinsic mode functions, which are then analyzed to extract meaningful temporal and spectral features, including true instantaneous features which can capture discriminative information in signals hidden at local time-scales. A total of six features are extracted, and a linear classifier is used with the feature vector to classify continuous speech portions obtained from a database consisting of 51 normal and 161 pathological speakers. A classification accuracy of 95.7 % is obtained, thus demonstrating the effectiveness of the methodology.
[Short- and long-term consequences of prenatal exposure to cannabis].
Karila, L; Cazas, O; Danel, T; Reynaud, M
2006-02-01
Cannabis is one of the most commonly used drugs by pregnant women. The objective of this review of literature was to examine the association between cannabis use during pregnancy and effects upon growth, cognitive development (memory, attention, executive functions...) and behavior of newborns, children and teenagers. We searched for articles indexed in the medline database from 1970 to 2005. The following terms were used in the literature search: cannabis/marijuana, pregnancy, fetal development, newborn, prenatal exposure, neurobehavioral deficits, cognitive deficits, executive functions, cannabinoids, reproduction. Most of the articles were published in English. Cannabis use during pregnancy is related to diverse neurobehavioral and cognitive outcomes, including symptoms of inattention, impulsivity, deficits in learning and memory, and a deficiency in aspects of executive functions. It seems difficult to identify complications, such as lower birth weight, only attributable to cannabis as opposed to the multiple perinatal complications associated with tobacco smoking. In addition to alcohol and cigarettes, information should be given to women about the potentially harmful effects on fetal development, newborns, children and teenagers of smoking cannabis. Therefore, it seems necessary to develop prevention programs on this subject.
Abiiro, Gilbert Abotisem; Leppert, Gerald; Mbera, Grace Bongololo; Robyn, Paul J; De Allegri, Manuela
2014-05-22
Discrete choice experiments (DCEs) are attribute-driven experimental techniques used to elicit stakeholders' preferences to support the design and implementation of policy interventions. The validity of a DCE, therefore, depends on the appropriate specification of the attributes and their levels. There have been recent calls for greater rigor in implementing and reporting on the processes of developing attributes and attribute-levels for discrete choice experiments (DCEs). This paper responds to such calls by carefully reporting a systematic process of developing micro health insurance attributes and attribute-levels for the design of a DCE in rural Malawi. Conceptual attributes and attribute-levels were initially derived from a literature review which informed the design of qualitative data collection tools to identify context specific attributes and attribute-levels. Qualitative data was collected in August-September 2012 from 12 focus group discussions with community residents and 8 in-depth interviews with health workers. All participants were selected according to stratified purposive sampling. The material was tape-recorded, fully transcribed, and coded by three researchers to identify context-specific attributes and attribute-levels. Expert opinion was used to scale down the attributes and levels. A pilot study confirmed the appropriateness of the selected attributes and levels for a DCE. First, a consensus, emerging from an individual level analysis of the qualitative transcripts, identified 10 candidate attributes. Levels were assigned to all attributes based on data from transcripts and knowledge of the Malawian context, derived from literature. Second, through further discussions with experts, four attributes were discarded based on multiple criteria. The 6 remaining attributes were: premium level, unit of enrollment, management structure, health service benefit package, transportation coverage and copayment levels. A final step of revision and piloting confirmed that the retained attributes satisfied the credibility criteria of DCE attributes. This detailed description makes our attribute development process transparent, and provides the reader with a basis to assess the rigor of this stage of constructing the DCE. This paper contributes empirical evidence to the limited methodological literature on attributes and levels development for DCE, thereby providing further empirical guidance on the matter, specifically within rural communities of low- and middle-income countries.
ERIC Educational Resources Information Center
Kollegger, James G.; And Others
1988-01-01
In the first of three articles, the producer of Energyline, Energynet, and Tele/Scope recalls the development of the databases and database business strategies. The second describes the development of biomedical online databases, and the third discusses future developments, including full text databases, database producers as online host, and…
The "common good" phenomenon: Why similarities are positive and differences are negative.
Alves, Hans; Koch, Alex; Unkelbach, Christian
2017-04-01
Positive attributes are more prevalent than negative attributes in the social environment. From this basic assumption, 2 implications that have been overlooked thus far: Positive compared with negative attributes are more likely to be shared by individuals, and people's shared attributes (similarities) are more positive than their unshared attributes (differences). Consequently, similarity-based comparisons should lead to more positive evaluations than difference-based comparisons. We formalized our probabilistic reasoning in a model and tested its predictions in a simulation and 8 experiments (N = 1,181). When participants generated traits about 2 target persons, positive compared with negative traits were more likely to be shared by the targets (Experiment 1a) and by other participants' targets (Experiment 1b). Conversely, searching for targets' shared traits resulted in more positive traits than searching for unshared traits (Experiments 2, 4a, and 4b). In addition, positive traits were more accessible than negative traits among shared traits but not among unshared traits (Experiment 3). Finally, shared traits were only more positive when positive traits were indeed prevalent (Experiments 5 and 6). The current framework has a number of implications for comparison processes and provides a new interpretation of well-known evaluative asymmetries such as intergroup bias and self-superiority effects. (PsycINFO Database Record (c) 2017 APA, all rights reserved).
Le, T Hoang Ngan; Luu, Khoa; Savvides, Marios
2013-08-01
Robust facial hair detection and segmentation is a highly valued soft biometric attribute for carrying out forensic facial analysis. In this paper, we propose a novel and fully automatic system, called SparCLeS, for beard/moustache detection and segmentation in challenging facial images. SparCLeS uses the multiscale self-quotient (MSQ) algorithm to preprocess facial images and deal with illumination variation. Histogram of oriented gradients (HOG) features are extracted from the preprocessed images and a dynamic sparse classifier is built using these features to classify a facial region as either containing skin or facial hair. A level set based approach, which makes use of the advantages of both global and local information, is then used to segment the regions of a face containing facial hair. Experimental results demonstrate the effectiveness of our proposed system in detecting and segmenting facial hair regions in images drawn from three databases, i.e., the NIST Multiple Biometric Grand Challenge (MBGC) still face database, the NIST Color Facial Recognition Technology FERET database, and the Labeled Faces in the Wild (LFW) database.
Integration of Web-based and PC-based clinical research databases.
Brandt, C A; Sun, K; Charpentier, P; Nadkarni, P M
2004-01-01
We have created a Web-based repository or data library of information about measurement instruments used in studies of multi-factorial geriatric health conditions (the Geriatrics Research Instrument Library - GRIL) based upon existing features of two separate clinical study data management systems. GRIL allows browsing, searching, and selecting measurement instruments based upon criteria such as keywords and areas of applicability. Measurement instruments selected can be printed and/or included in an automatically generated standalone microcomputer database application, which can be downloaded by investigators for use in data collection and data management. Integration of database applications requires the creation of a common semantic model, and mapping from each system to this model. Various database schema conflicts at the table and attribute level must be identified and resolved prior to integration. Using a conflict taxonomy and a mapping schema facilitates this process. Critical conflicts at the table level that required resolution included name and relationship differences. A major benefit of integration efforts is the sharing of features and cross-fertilization of applications created for similar purposes in different operating environments. Integration of applications mandates some degree of metadata model unification.
Carlson, Mary H.; Zientek, Michael L.; Causey, J. Douglas; Kayser, Helen Z.; Spanski, Gregory T.; Wilson, Anna B.; Van Gosen, Bradley S.; Trautwein, Charles M.
2007-01-01
This report compiles selected results from 13 U.S. Geological Survey (USGS) mineral resource assessment studies conducted in Idaho and Montana into consistent spatial databases that can be used in a geographic information system. The 183 spatial databases represent areas of mineral potential delineated in these studies and include attributes on mineral deposit type, level of mineral potential, certainty, and a reference. The assessments were conducted for five 1? x 2? quadrangles (Butte, Challis, Choteau, Dillon, and Wallace), several U.S. Forest Service (USFS) National Forests (including Challis, Custer, Gallatin, Helena, and Payette), and one Bureau of Land Management (BLM) Resource Area (Dillon). The data contained in the spatial databases are based on published information: no new interpretations are made. This digital compilation is part of an ongoing effort to provide mineral resource information formatted for use in spatial analysis. In particular, this is one of several reports prepared to address USFS needs for science information as forest management plans are revised in the Northern Rocky Mountains.
NASA Astrophysics Data System (ADS)
Oppikofer, Thierry; Nordahl, Bobo; Bunkholt, Halvor; Nicolaisen, Magnus; Jarna, Alexandra; Iversen, Sverre; Hermanns, Reginald L.; Böhme, Martina; Yugsi Molina, Freddy X.
2015-11-01
The unstable rock slope database is developed and maintained by the Geological Survey of Norway as part of the systematic mapping of unstable rock slopes in Norway. This mapping aims to detect catastrophic rock slope failures before they occur. More than 250 unstable slopes with post-glacial deformation are detected up to now. The main aims of the unstable rock slope database are (1) to serve as a national archive for unstable rock slopes in Norway; (2) to serve for data collection and storage during field mapping; (3) to provide decision-makers with hazard zones and other necessary information on unstable rock slopes for land-use planning and mitigation; and (4) to inform the public through an online map service. The database is organized hierarchically with a main point for each unstable rock slope to which several feature classes and tables are linked. This main point feature class includes several general attributes of the unstable rock slopes, such as site name, general and geological descriptions, executed works, recommendations, technical parameters (volume, lithology, mechanism and others), displacement rates, possible consequences, as well as hazard and risk classification. Feature classes and tables linked to the main feature class include different scenarios of an unstable rock slope, field observation points, sampling points for dating, displacement measurement stations, lineaments, unstable areas, run-out areas, areas affected by secondary effects, along with tables for hazard and risk classification and URL links to further documentation and references. The database on unstable rock slopes in Norway will be publicly consultable through an online map service. Factsheets with key information on unstable rock slopes can be automatically generated and downloaded for each site. Areas of possible rock avalanche run-out and their secondary effects displayed in the online map service, along with hazard and risk assessments, will become important tools for land-use planning. The present database will further evolve in the coming years as the systematic mapping progresses and as available techniques and tools evolve.
Volcanoes of the World: Reconfiguring a scientific database to meet new goals and expectations
NASA Astrophysics Data System (ADS)
Venzke, Edward; Andrews, Ben; Cottrell, Elizabeth
2015-04-01
The Smithsonian Global Volcanism Program's (GVP) database of Holocene volcanoes and eruptions, Volcanoes of the World (VOTW), originated in 1971, and was largely populated with content from the IAVCEI Catalog of Volcanoes of Active Volcanoes and some independent datasets. Volcanic activity reported by Smithsonian's Bulletin of the Global Volcanism Network and USGS/SI Weekly Activity Reports (and their predecessors), published research, and other varied sources has expanded the database significantly over the years. Three editions of the VOTW were published in book form, creating a catalog with new ways to display data that included regional directories, a gazetteer, and a 10,000-year chronology of eruptions. The widespread dissemination of the data in electronic media since the first GVP website in 1995 has created new challenges and opportunities for this unique collection of information. To better meet current and future goals and expectations, we have recently transitioned VOTW into a SQL Server database. This process included significant schema changes to the previous relational database, data auditing, and content review. We replaced a disparate, confusing, and changeable volcano numbering system with unique and permanent volcano numbers. We reconfigured structures for recording eruption data to allow greater flexibility in describing the complexity of observed activity, adding in the ability to distinguish episodes within eruptions (in time and space) and events (including dates) rather than characteristics that take place during an episode. We have added a reference link field in multiple tables to enable attribution of sources at finer levels of detail. We now store and connect synonyms and feature names in a more consistent manner, which will allow for morphological features to be given unique numbers and linked to specific eruptions or samples; if the designated overall volcano name is also a morphological feature, it is then also listed and described as that feature. One especially significant audit involved re-evaluating the categories of evidence used to include a volcano in the Holocene list, and reviewing in detail the entries in low-certainty categories. Concurrently, we developed a new data entry system that may in the future allow trusted users outside of Smithsonian to input data into VOTW. A redesigned website now provides new search tools and data download options. We are collaborating with organizations that manage volcano and eruption databases, physical sample databases, and geochemical databases to allow real-time connections and complex queries. VOTW serves the volcanological community by providing a clear and consistent core database of distinctly identified volcanoes and eruptions to advance goals in research, civil defense, and public outreach.
Dodge, Kenneth A.
2009-01-01
A model of the development of hostile attributional style and its role in children's aggressive behavior is proposed, based on the translation of basic science in ethology, neuroscience, social psychology, personality psychology, and developmental psychology. Theory and findings from these domains are reviewed and synthesized in the proposed model, which posits that (a) aggressive behavior and hostile attributions are universal human characteristics, (b) socialization leads to the development of benign attributions, (c) individual differences in attributional style account for differences in aggressive behavior, and (d) interventions to change attributions have the potential to alter antisocial development. Challenges for future research are described. PMID:17152401
Kline, Jeffrey A; Courtney, D Mark; Than, Martin P; Hogg, Kerstin; Miller, Chadwick D; Johnson, Charles L; Smithline, Howard A
2010-02-01
Attribute matching matches an explicit clinical profile of a patient to a reference database to estimate the numeric value for the pretest probability of an acute disease. The authors tested the accuracy of this method for forecasting a very low probability of venous thromboembolism (VTE) in symptomatic emergency department (ED) patients. The authors performed a secondary analysis of five data sets from 15 hospitals in three countries. All patients had data collected at the time of clinical evaluation for suspected pulmonary embolism (PE). The criterion standard to exclude VTE required no evidence of PE or deep venous thrombosis (DVT) within 45 days of enrollment. To estimate pretest probabilities, a computer program selected, from a large reference database of patients previously evaluated for PE, patients who matched 10 predictor variables recorded for each current test patient. The authors compared the outcome frequency of having VTE [VTE(+)] in patients with a pretest probability estimate of <2.5% by attribute matching, compared with a value of 0 from the Wells score. The five data sets included 10,734 patients, and 747 (7.0%, 95% confidence interval [CI] = 6.5% to 7.5%) were VTE(+) within 45 days. The pretest probability estimate for PE was <2.5% in 2,975 of 10,734 (27.7%) patients, and within this subset, the observed frequency of VTE(+) was 48 of 2,975 (1.6%, 95% CI = 1.2% to 2.1%). The lowest possible Wells score (0) was observed in 3,412 (31.7%) patients, and within this subset, the observed frequency of VTE(+) was 79 of 3,412 (2.3%, 95% CI = 1.8% to 2.9%) patients. Attribute matching categorizes over one-quarter of patients tested for PE as having a pretest probability of <2.5%, and the observed rate of VTE within 45 days in this subset was <2.5%. (c) 2010 by the Society for Academic Emergency Medicine.
The transition into veterinary practice: Opinions of recent graduates and final year students
2011-01-01
Background The transition from veterinary student to member of the veterinary profession is known to be challenging. This study aimed to determine and compare the opinions of final year veterinary students and recent graduates on graduate attributes that ease this transition. Methods The study was carried out across 3 veterinary schools in the United Kingdom. Paper based or electronic surveys were used. Final year students in the 3 schools were surveyed either electronically (school A) or on paper (schools B and C). Student cohort sizes were 112, 227 and 102 respectively. Recent graduates were contacted either at a reunion event (school A) or electronically from database records (school B and school C). Cohort sizes of contacted graduates were 80, 175 and 91 respectively. Respondents were asked to rate 42 individual attributes on a 5 point Likert scale. Focus groups with final year students and recent graduates and telephone interviews with recent graduates were carried out. Data were analysed by two researchers through a combination of manual coding and thematic analysis. Data were grouped into broad themes then sorted into narrower themes. Data were then searched for counter examples. Results Response rates for final year students were 34% (school A), 36% (school B) and 40% (school C). Response rates for recent graduates were 56% (school A), 20% (school B) and 11% (school C). There was a high level of agreement between the cohorts with respect to communication skills, problem solving and decision making skills, recognition of own limitations and the ability to cope with pressure all rated unanimously important or very important. Business acumen, knowledge of veterinary practice management and research skills were the 3 attributes ranked at the bottom of the list. Nine attributes were identified with a significantly different (p < 0.05) ranking between the cohorts. Final year students ranked veterinary clinical knowledge, knowledge of veterinary public health and zoonotic issues, veterinary legislation and veterinary practice management, commitment to continuing professional development and ability to evaluate information higher than recent graduates. Recent graduates ranked the attributes of integrity, friendliness and compassion higher than final year students. Conclusions Recent graduates and final year students rate highly the attributes which help foster the client/veterinarian relationship. Recent graduates reflect that a focus on knowledge based attributes is less important once in practice when compared to final year. The study confirms the importance to recent graduates and final year students of attributes considered as non-technical in the transition to working in the veterinary profession. PMID:21939551
The transition into veterinary practice: opinions of recent graduates and final year students.
Rhind, Susan M; Baillie, Sarah; Kinnison, Tierney; Shaw, Darren J; Bell, Catriona E; Mellanby, Richard J; Hammond, Jenny; Hudson, Neil P H; Whittington, Rachel E; Donnelly, Ruth
2011-09-22
The transition from veterinary student to member of the veterinary profession is known to be challenging. This study aimed to determine and compare the opinions of final year veterinary students and recent graduates on graduate attributes that ease this transition. The study was carried out across 3 veterinary schools in the United Kingdom. Paper based or electronic surveys were used. Final year students in the 3 schools were surveyed either electronically (school A) or on paper (schools B and C). Student cohort sizes were 112, 227 and 102 respectively. Recent graduates were contacted either at a reunion event (school A) or electronically from database records (school B and school C). Cohort sizes of contacted graduates were 80, 175 and 91 respectively. Respondents were asked to rate 42 individual attributes on a 5 point Likert scale. Focus groups with final year students and recent graduates and telephone interviews with recent graduates were carried out. Data were analysed by two researchers through a combination of manual coding and thematic analysis. Data were grouped into broad themes then sorted into narrower themes. Data were then searched for counter examples. Response rates for final year students were 34% (school A), 36% (school B) and 40% (school C). Response rates for recent graduates were 56% (school A), 20% (school B) and 11% (school C). There was a high level of agreement between the cohorts with respect to communication skills, problem solving and decision making skills, recognition of own limitations and the ability to cope with pressure all rated unanimously important or very important. Business acumen, knowledge of veterinary practice management and research skills were the 3 attributes ranked at the bottom of the list. Nine attributes were identified with a significantly different (p < 0.05) ranking between the cohorts. Final year students ranked veterinary clinical knowledge, knowledge of veterinary public health and zoonotic issues, veterinary legislation and veterinary practice management, commitment to continuing professional development and ability to evaluate information higher than recent graduates. Recent graduates ranked the attributes of integrity, friendliness and compassion higher than final year students. Recent graduates and final year students rate highly the attributes which help foster the client/veterinarian relationship. Recent graduates reflect that a focus on knowledge based attributes is less important once in practice when compared to final year. The study confirms the importance to recent graduates and final year students of attributes considered as non-technical in the transition to working in the veterinary profession.
Learning-based stochastic object models for characterizing anatomical variations
NASA Astrophysics Data System (ADS)
Dolly, Steven R.; Lou, Yang; Anastasio, Mark A.; Li, Hua
2018-03-01
It is widely known that the optimization of imaging systems based on objective, task-based measures of image quality via computer-simulation requires the use of a stochastic object model (SOM). However, the development of computationally tractable SOMs that can accurately model the statistical variations in human anatomy within a specified ensemble of patients remains a challenging task. Previously reported numerical anatomic models lack the ability to accurately model inter-patient and inter-organ variations in human anatomy among a broad patient population, mainly because they are established on image data corresponding to a few of patients and individual anatomic organs. This may introduce phantom-specific bias into computer-simulation studies, where the study result is heavily dependent on which phantom is used. In certain applications, however, databases of high-quality volumetric images and organ contours are available that can facilitate this SOM development. In this work, a novel and tractable methodology for learning a SOM and generating numerical phantoms from a set of volumetric training images is developed. The proposed methodology learns geometric attribute distributions (GAD) of human anatomic organs from a broad patient population, which characterize both centroid relationships between neighboring organs and anatomic shape similarity of individual organs among patients. By randomly sampling the learned centroid and shape GADs with the constraints of the respective principal attribute variations learned from the training data, an ensemble of stochastic objects can be created. The randomness in organ shape and position reflects the learned variability of human anatomy. To demonstrate the methodology, a SOM of an adult male pelvis is computed and examples of corresponding numerical phantoms are created.
Epidemiology of international terrorism involving fatal outcomes in developed countries (1994-2003).
Wilson, Nick; Thomson, George
2005-01-01
We aimed to describe the public health burden and epidemiology of international terrorism (i.e. involving foreign nationals) with fatal outcomes in developed countries. Data was abstracted from a United States Department of State database for 21 'established market economy' countries and 18 'former socialist economies of Europe' for 1994-2003. To put the findings in a wider context, comparisons were made with WHO data on all homicides for each country. A total of 32 international terrorist attacks causing fatalities were identified over the 10-year period. These resulted in 3299 deaths, giving a crude annual mortality rate of 0.3 per million population. The mortality burden attributable to international terrorism in these countries was 208 times less than that attributable to other homicide. Even for the country with the highest mortality burden from international terrorism (the United States), this ratio was 60. There was no statistically significant trend in the number of attacks over time, but the attack severity (in terms of deaths per attack) was higher in the latter part of the 10-year period. A number of limitations with this data set were identified. If a more rigorous definition of 'international terrorism' was used, then this would substantially reduce the total number of such attacks defined in this way. In conclusion, there is a need for better quality data and improved classification systems for describing international terrorism. Nevertheless, the available data indicates that the mortality burden from international terrorism in developed countries is small compared to that from other homicide.
Israel Marine Bio-geographic Database (ISRAMAR-BIO)
NASA Astrophysics Data System (ADS)
Greengrass, Eyal; Krivenko, Yevgeniya; Ozer, Tal; Ben Yosef, Dafna; Tom, Moshe; Gertman, Isaac
2015-04-01
The knowledge of the space/time variations of species is the basis for any ecological investigations. While historical observations containing integral concentrations of biological parameters (chlorophyll, abundance, biomass…) are organized partly in ISRAMAR Cast Database, the taxon-specific data collected in Israel has not been sufficiently organized. This has been hindered by the lack of standards, variability of methods and complexity of biological data formalization. The ISRAMAR-BIO DB was developed to store various types of historical and future available information related to marine species observations and related metadata. Currently the DB allows to store biological data acquired by the following sampling devices such as: van veer grab, box corer, sampling bottles, nets (plankton, trawls and fish), quadrates, and cameras. The DB's logical unit is information regarding a specimen (taxa name, barcode, image), related attributes (abundance, size, age, contaminants…), habitat description, sampling device and method, time and space of sampling, responsible organization and scientist, source of information (cruise, project and publication). The following standardization of specimen and attributes naming were implemented: Taxonomy according to World Register of Marine Species (WoRMS: http://www.marinespecies.org). Habitat description according to Coastal and Marine Ecological Classification Standards (CMECS: http://www.cmecscatalog.org) Parameter name; Unit; Device name; Developmental stage; Institution name; Country name; Marine region according to SeaDataNet Vocabularies (http://www.seadatanet.org/Standards-Software/Common-Vocabularies). This system supports two types of data submission procedures, which support the above stated data structure. The first is a downloadable excel file with drop-down fields based on the ISRAMAR-BIO vocabularies. The file is filled and uploaded online by the data contributor. Alternatively, the same dataset can be assembled by filling online forms and then submitted to the DB. Online access to the ISRAMAR-BIO is available through taxon search page, where one can get both biological and geographical data regarding a certain taxon. Further development of the online data access is ongoing. It will include interactive geographical map interface where data may be queried, analyzed and downloaded.
The Iranian National Geodata Revision Strategy and Realization Based on Geodatabase
NASA Astrophysics Data System (ADS)
Haeri, M.; Fasihi, A.; Ayazi, S. M.
2012-07-01
In recent years, using of spatial database for storing and managing spatial data has become a hot topic in the field of GIS. Accordingly National Cartographic Center of Iran (NCC) produces - from time to time - some spatial data which is usually included in some databases. One of the NCC major projects was designing National Topographic Database (NTDB). NCC decided to create National Topographic Database of the entire country-based on 1:25000 coverage maps. The standard of NTDB was published in 1994 and its database was created at the same time. In NTDB geometric data was stored in MicroStation design format (DGN) which each feature has a link to its attribute data (stored in Microsoft Access file). Also NTDB file was produced in a sheet-wise mode and then stored in a file-based style. Besides map compilation, revision of existing maps has already been started. Key problems of NCC are revision strategy, NTDB file-based style storage and operator challenges (NCC operators are almost preferred to edit and revise geometry data in CAD environments). A GeoDatabase solution for national Geodata, based on NTDB map files and operators' revision preferences, is introduced and released herein. The proposed solution extends the traditional methods to have a seamless spatial database which it can be revised in CAD and GIS environment, simultaneously. The proposed system is the common data framework to create a central data repository for spatial data storage and management.
NASA Astrophysics Data System (ADS)
Bartolini, S.; Becerril, L.; Martí, J.
2014-11-01
One of the most important issues in modern volcanology is the assessment of volcanic risk, which will depend - among other factors - on both the quantity and quality of the available data and an optimum storage mechanism. This will require the design of purpose-built databases that take into account data format and availability and afford easy data storage and sharing, and will provide for a more complete risk assessment that combines different analyses but avoids any duplication of information. Data contained in any such database should facilitate spatial and temporal analysis that will (1) produce probabilistic hazard models for future vent opening, (2) simulate volcanic hazards and (3) assess their socio-economic impact. We describe the design of a new spatial database structure, VERDI (Volcanic managEment Risk Database desIgn), which allows different types of data, including geological, volcanological, meteorological, monitoring and socio-economic information, to be manipulated, organized and managed. The root of the question is to ensure that VERDI will serve as a tool for connecting different kinds of data sources, GIS platforms and modeling applications. We present an overview of the database design, its components and the attributes that play an important role in the database model. The potential of the VERDI structure and the possibilities it offers in regard to data organization are here shown through its application on El Hierro (Canary Islands). The VERDI database will provide scientists and decision makers with a useful tool that will assist to conduct volcanic risk assessment and management.
Powell, Robert E.
2001-01-01
This data set maps and describes the geology of the Porcupine Wash 7.5 minute quadrangle, Riverside County, southern California. The quadrangle, situated in Joshua Tree National Park in the eastern Transverse Ranges physiographic and structural province, encompasses parts of the Hexie Mountains, Cottonwood Mountains, northern Eagle Mountains, and south flank of Pinto Basin. It is underlain by a basement terrane comprising Proterozoic metamorphic rocks, Mesozoic plutonic rocks, and Mesozoic and Mesozoic or Cenozoic hypabyssal dikes. The basement terrane is capped by a widespread Tertiary erosion surface preserved in remnants in the Eagle and Cottonwood Mountains and buried beneath Cenozoic deposits in Pinto Basin. Locally, Miocene basalt overlies the erosion surface. A sequence of at least three Quaternary pediments is planed into the north piedmont of the Eagle and Hexie Mountains, each in turn overlain by successively younger residual and alluvial deposits. The Tertiary erosion surface is deformed and broken by north-northwest-trending, high-angle, dip-slip faults and an east-west trending system of high-angle dip- and left-slip faults. East-west trending faults are younger than and perhaps in part coeval with faults of the northwest-trending set. The Porcupine Wash database was created using ARCVIEW and ARC/INFO, which are geographical information system (GIS) software products of Envronmental Systems Research Institute (ESRI). The database consists of the following items: (1) a map coverage showing faults and geologic contacts and units, (2) a separate coverage showing dikes, (3) a coverage showing structural data, (4) a scanned topographic base at a scale of 1:24,000, and (5) attribute tables for geologic units (polygons and regions), contacts (arcs), and site-specific data (points). The database, accompanied by a pamphlet file and this metadata file, also includes the following graphic and text products: (1) A portable document file (.pdf) containing a navigable graphic of the geologic map on a 1:24,000 topographic base. The map is accompanied by a marginal explanation consisting of a Description of Map and Database Units (DMU), a Correlation of Map and Database Units (CMU), and a key to point-and line-symbols. (2) Separate .pdf files of the DMU and CMU, individually. (3) A PostScript graphic-file containing the geologic map on a 1:24,000 topographic base accompanied by the marginal explanation. (4) A pamphlet that describes the database and how to access it. Within the database, geologic contacts , faults, and dikes are represented as lines (arcs), geologic units as polygons and regions, and site-specific data as points. Polygon, arc, and point attribute tables (.pat, .aat, and .pat, respectively) uniquely identify each geologic datum and link it to other tables (.rel) that provide more detailed geologic information.
2014-01-01
Background Discrete choice experiments (DCEs) are attribute-driven experimental techniques used to elicit stakeholders’ preferences to support the design and implementation of policy interventions. The validity of a DCE, therefore, depends on the appropriate specification of the attributes and their levels. There have been recent calls for greater rigor in implementing and reporting on the processes of developing attributes and attribute-levels for discrete choice experiments (DCEs). This paper responds to such calls by carefully reporting a systematic process of developing micro health insurance attributes and attribute-levels for the design of a DCE in rural Malawi. Methods Conceptual attributes and attribute-levels were initially derived from a literature review which informed the design of qualitative data collection tools to identify context specific attributes and attribute-levels. Qualitative data was collected in August-September 2012 from 12 focus group discussions with community residents and 8 in-depth interviews with health workers. All participants were selected according to stratified purposive sampling. The material was tape-recorded, fully transcribed, and coded by three researchers to identify context-specific attributes and attribute-levels. Expert opinion was used to scale down the attributes and levels. A pilot study confirmed the appropriateness of the selected attributes and levels for a DCE. Results First, a consensus, emerging from an individual level analysis of the qualitative transcripts, identified 10 candidate attributes. Levels were assigned to all attributes based on data from transcripts and knowledge of the Malawian context, derived from literature. Second, through further discussions with experts, four attributes were discarded based on multiple criteria. The 6 remaining attributes were: premium level, unit of enrollment, management structure, health service benefit package, transportation coverage and copayment levels. A final step of revision and piloting confirmed that the retained attributes satisfied the credibility criteria of DCE attributes. Conclusion This detailed description makes our attribute development process transparent, and provides the reader with a basis to assess the rigor of this stage of constructing the DCE. This paper contributes empirical evidence to the limited methodological literature on attributes and levels development for DCE, thereby providing further empirical guidance on the matter, specifically within rural communities of low- and middle-income countries. PMID:24884920
Taylor, Benjamin; Rehm, Jürgen; Patra, Jayadeep; Popova, Svetlana; Baliunas, Dolly
2007-01-01
Alcohol is one of the most important risk factors for burden of disease, particularly in high-income countries such as Canada. The purpose of this article was to estimate the number of hospitalizations, hospital days, and the resulting costs attributable to alcohol for Canada in 2002. Exposure distribution was taken from the Canadian Addiction Survey and corrected for per capita consumption from production and sales. For chronic disease, risk relations were taken from the published literature and combined with exposure to calculate age- and gender-specific alcohol-attributable fractions. For injury, alcohol-attributable fractions were taken directly from available statistics. Data on the most responsible diagnosis, length of stay for hospitalizations, and costs were obtained from the national Canadian databases. For Canada in 2002, there were 195,970 alcohol-related diagnoses among acute care hospitalizations, 2,058 alcohol-attributable psychiatric hospitalizations, and 183,589 alcohol-attributable admissions to specialized treatment centers. These accounted for 1,246,945 hospital days in acute care facilities, 54,114 hospital days in psychiatric hospitals, and 3,018,688 hospital days in specialized treatment centers (inpatient and outpatient). The main causes of alcohol-attributable morbidity were neuropsychiatric conditions, cardiovascular disease, and unintentional injuries. In total, Can. $2.29 billion were spent on alcohol-related health care. Alcohol poses a heavy burden of disease as well as a financial strain on Canadian society. However, there are evidence-based effective and cost-effective policy and legislative interventions as well as measures to better enforce these laws.
EPIRUS-NET: A Wireless Health Telematics Network in Greece
2001-10-25
has also to be made. The system is represented of three basic layers: the database layer, the middleware and the Hospital Daily Progress Anamnesis ... Anamnesis . The Encounter entry is uniquely identified by the incremental ID, IID, attribute. Each encounter entry is associated with a clinical...the main entities of the system (Hospital, Patient, Anamnesis , Encounter, Clinical Examination, Daily Progress, Examination, Release Ticket), along
Peter U. Kennedy; Victor B. Shelburne
2002-01-01
Geographic Information Systems (GIS) data and historical plats ranging from 1716 to 1894 in the Coastal Flatwoods Region of South Carolina were used to quantify changes on a temporal scale. Combining the historic plats and associated witness trees (trees marking the boundaries of historic plats) with an existing database of the soils and other attributes was the basis...
Converting analog interpretive data to digital formats for use in database and GIS applications
Flocks, James G.
2004-01-01
There is a growing need by researchers and managers for comprehensive and unified nationwide datasets of scientific data. These datasets must be in a digital format that is easily accessible using database and GIS applications, providing the user with access to a wide variety of current and historical information. Although most data currently being collected by scientists are already in a digital format, there is still a large repository of information in the literature and paper archive. Converting this information into a format accessible by computer applications is typically very difficult and can result in loss of data. However, since scientific data are commonly collected in a repetitious, concise matter (i.e., forms, tables, graphs, etc.), these data can be recovered digitally by using a conversion process that relates the position of an attribute in two-dimensional space to the information that the attribute signifies. For example, if a table contains a certain piece of information in a specific row and column, then the space that the row and column occupies becomes an index of that information. An index key is used to identify the relation between the physical location of the attribute and the information the attribute contains. The conversion process can be achieved rapidly, easily and inexpensively using widely available digitizing and spreadsheet software, and simple programming code. In the geological sciences, sedimentary character is commonly interpreted from geophysical profiles and descriptions of sediment cores. In the field and laboratory, these interpretations were typically transcribed to paper. The information from these paper archives is still relevant and increasingly important to scientists, engineers and managers to understand geologic processes affecting our environment. Direct scanning of this information produces a raster facsimile of the data, which allows it to be linked to the electronic world. But true integration of the content with database and GIS software as point, vector or text information is commonly lost. Sediment core descriptions and interpretation of geophysical profiles are usually portrayed as lines, curves, symbols and text information. They have vertical and horizontal dimensions associated with depth, category, time, or geographic position. These dimensions are displayed in consistent positions, which can be digitized and converted to a digital format, such as a spreadsheet. Once this data is in a digital, tabulated form it can easily be made available to a wide variety of imaging and data manipulation software for compilation and world-wide dissemination.
Sharma, Vishal K; Fraulin, Frankie Og; Harrop, A Robertson; McPhalen, Donald F
2011-01-01
Databases are useful tools in clinical settings. The authors review the benefits and challenges associated with the development and implementation of an efficient electronic database for the multidisciplinary Vascular Birthmark Clinic at the Alberta Children's Hospital, Calgary, Alberta. The content and structure of the database were designed using the technical expertise of a data analyst from the Calgary Health Region. Relevant clinical and demographic data fields were included with the goal of documenting ongoing care of individual patients, and facilitating future epidemiological studies of this patient population. After completion of this database, 10 challenges encountered during development were retrospectively identified. Practical solutions for these challenges are presented. THE CHALLENGES IDENTIFIED DURING THE DATABASE DEVELOPMENT PROCESS INCLUDED: identification of relevant data fields; balancing simplicity and user-friendliness with complexity and comprehensive data storage; database expertise versus clinical expertise; software platform selection; linkage of data from the previous spreadsheet to a new data management system; ethics approval for the development of the database and its utilization for research studies; ensuring privacy and limited access to the database; integration of digital photographs into the database; adoption of the database by support staff in the clinic; and maintaining up-to-date entries in the database. There are several challenges involved in the development of a useful and efficient clinical database. Awareness of these potential obstacles, in advance, may simplify the development of clinical databases by others in various surgical settings.
Online Petroleum Industry Bibliographic Databases: A Review.
ERIC Educational Resources Information Center
Anderson, Margaret B.
This paper discusses the present status of the bibliographic database industry, reviews the development of online databases of interest to the petroleum industry, and considers future developments in online searching and their effect on libraries and information centers. Three groups of databases are described: (1) databases developed by the…
Effect of Temporal Relationships in Associative Rule Mining for Web Log Data
Mohd Khairudin, Nazli; Mustapha, Aida
2014-01-01
The advent of web-based applications and services has created such diverse and voluminous web log data stored in web servers, proxy servers, client machines, or organizational databases. This paper attempts to investigate the effect of temporal attribute in relational rule mining for web log data. We incorporated the characteristics of time in the rule mining process and analysed the effect of various temporal parameters. The rules generated from temporal relational rule mining are then compared against the rules generated from the classical rule mining approach such as the Apriori and FP-Growth algorithms. The results showed that by incorporating the temporal attribute via time, the number of rules generated is subsequently smaller but is comparable in terms of quality. PMID:24587757
Resilient ageing: a concept analysis.
Hicks, Maxine M; Conner, Norma E
2014-04-01
This paper is a report of an analysis of the concept resilient ageing. Unique in comparison with other healthy ageing concepts, resilient ageing can be applied to all older people, regardless of age or affliction. The state of global population expansion in older people over the next 50 years calls for increased health promotion research efforts to ensure the maintenance of health and optimal quality of life for all older people. Literature for this concept analysis was retrieved from several databases, CINAHL, PubMed PsycINFO, for the years 1990-2012. Rodgers's evolutionary method of concept analysis was used because of its applicability to concepts that are still evolving. An integrative research review methodology was applied to peer-reviewed journal articles (n = 46) for an inductive analysis of the concept of resilient ageing. The antecedents, defining attributes, and consequence of resilient ageing were identified. Antecedents to resilient ageing were found to be adversity and protective factors, while the core attributes include coping, hardiness and self-concept. The consequence of the process of resilient ageing was optimal quality of life. Sense of coherence was found to be the surrogate term. The results obtained were further substantiated using Antonovsky's (1979) theory of salutogenesis. A theoretical definition and a model of resilient ageing were developed. In addition, a discussion was provided on the practice, policy and research implications for promoting the development of protective factors and resilient ageing. © 2013 John Wiley & Sons Ltd.
Peissig, Peggy L; Rasmussen, Luke V; Berg, Richard L; Linneman, James G; McCarty, Catherine A; Waudby, Carol; Chen, Lin; Denny, Joshua C; Wilke, Russell A; Pathak, Jyotishman; Carrell, David; Kho, Abel N; Starren, Justin B
2012-01-01
There is increasing interest in using electronic health records (EHRs) to identify subjects for genomic association studies, due in part to the availability of large amounts of clinical data and the expected cost efficiencies of subject identification. We describe the construction and validation of an EHR-based algorithm to identify subjects with age-related cataracts. We used a multi-modal strategy consisting of structured database querying, natural language processing on free-text documents, and optical character recognition on scanned clinical images to identify cataract subjects and related cataract attributes. Extensive validation on 3657 subjects compared the multi-modal results to manual chart review. The algorithm was also implemented at participating electronic MEdical Records and GEnomics (eMERGE) institutions. An EHR-based cataract phenotyping algorithm was successfully developed and validated, resulting in positive predictive values (PPVs) >95%. The multi-modal approach increased the identification of cataract subject attributes by a factor of three compared to single-mode approaches while maintaining high PPV. Components of the cataract algorithm were successfully deployed at three other institutions with similar accuracy. A multi-modal strategy incorporating optical character recognition and natural language processing may increase the number of cases identified while maintaining similar PPVs. Such algorithms, however, require that the needed information be embedded within clinical documents. We have demonstrated that algorithms to identify and characterize cataracts can be developed utilizing data collected via the EHR. These algorithms provide a high level of accuracy even when implemented across multiple EHRs and institutional boundaries.
Effects of compassion meditation on a psychological model of charitable donation.
Ashar, Yoni K; Andrews-Hanna, Jessica R; Yarkoni, Tal; Sills, Jenifer; Halifax, Joan; Dimidjian, Sona; Wager, Tor D
2016-08-01
Compassion is critical for societal wellbeing. Yet, it remains unclear how specific thoughts and feelings motivate compassionate behavior, and we lack a scientific understanding of how to effectively cultivate compassion. Here, we conducted 2 studies designed to a) develop a psychological model predicting compassionate behavior, and b) test this model as a mediator of a Compassion Meditation (CM) intervention and identify the "active ingredients" of CM. In Study 1, we developed a model predicting compassionate behavior, operationalized as real-money charitable donation, from a linear combination of self-reported tenderness, personal distress, perceived blamelessness, and perceived instrumental value of helping with high cross-validated accuracy, r = .67, p < .0001. Perceived similarity to suffering others did not predict charitable donation when controlling for other feelings and attributions. In Study 2, a randomized controlled trial, we tested the Study 1 model as a mediator of CM and investigated active ingredients. We compared a smartphone-based CM program to 2 conditions-placebo oxytocin and a Familiarity intervention-to control for expectancy effects, demand characteristics, and familiarity effects. Relative to control conditions, CM increased charitable donations, and changes in the Study 1 model of feelings and attributions mediated this effect (pab = .002). The Familiarity intervention led to decreases in primary outcomes, while placebo oxytocin had no significant effects on primary outcomes. Overall, this work contributes a quantitative model of compassionate behavior, and informs our understanding of the change processes and intervention components of CM. (PsycINFO Database Record (c) 2016 APA, all rights reserved).
Elam, Kit K; Harold, Gordon T; Neiderhiser, Jenae M; Reiss, David; Shaw, Daniel S; Natsuaki, Misaki N; Gaysina, Darya; Barrett, Doug; Leve, Leslie D
2014-05-01
Socially disruptive behavior during peer interactions in early childhood is detrimental to children's social, emotional, and academic development. Few studies have investigated the developmental underpinnings of children's socially disruptive behavior using genetically sensitive research designs that allow examination of parent-on-child and child-on-parent (evocative genotype-environment correlation [rGE]) effects when examining family process and child outcome associations. Using an adoption-at-birth design, the present study controlled for passive genotype-environment correlation and directly examined evocative rGE while examining the associations between family processes and children's peer behavior. Specifically, the present study examined the evocative effect of genetic influences underlying toddler low social motivation on mother-child and father-child hostility and the subsequent influence of parent hostility on disruptive peer behavior during the preschool period. Participants were 316 linked triads of birth mothers, adoptive parents, and adopted children. Path analysis showed that birth mother low behavioral motivation predicted toddler low social motivation, which predicted both adoptive mother-child and father-child hostility, suggesting the presence of an evocative genotype-environment association. In addition, both mother-child and father-child hostility predicted children's later disruptive peer behavior. Results highlight the importance of considering genetically influenced child attributes on parental hostility that in turn links to later child social behavior. Implications for intervention programs focusing on early family processes and the precursors of disrupted child social development are discussed. (PsycINFO Database Record (c) 2014 APA, all rights reserved).
NASA Astrophysics Data System (ADS)
Hsu, L.; Bristol, S.; Lehnert, K. A.; Arko, R. A.; Peters, S. E.; Uhen, M. D.; Song, L.
2014-12-01
The U.S. Geological Survey (USGS) is an exemplar of the need for improved cyberinfrastructure for its vast holdings of invaluable physical geoscience data. Millions of discrete paleobiological and geological specimens lie in USGS warehouses and at the Smithsonian Institution. These specimens serve as the basis for many geologic maps and geochemical databases, and are a potential treasure trove of new scientific knowledge. The extent of this treasure is virtually unknown and inaccessible outside a small group of paleogeoscientists and geochemists. A team from the USGS, the Integrated Earth Data Applications (IEDA) facility, and the Paleobiology Database (PBDB) are working to expose information on paleontological and geochemical specimens for discovery by scientists and citizens. This project uses existing infrastructure of the System for Earth Sample Registration (SESAR) and PBDB, which already contains much of the fundamental data schemas that are necessary to accommodate USGS records. The project is also developing a new Linked Data interface for the USGS National Geochemical Database (NGDB). The International Geo Sample Number (IGSN) is the identifier that links samples between all systems. For paleontological specimens, SESAR and PBDB will be the primary repositories for USGS records, with a data syncing process to archive records within the USGS ScienceBase system. The process began with mapping the metadata fields necessary for USGS collections to the existing SESAR and PBDB data structures, while aligning them with the Observations & Measurements and Darwin Core standards. New functionality needed in SESAR included links to a USGS locality registry, fossil classifications, a spatial qualifier attribution for samples with sensitive locations, and acknowledgement of data and metadata licensing. The team is developing a harvesting mechanism to periodically transfer USGS records from within PBDB and SESAR to ScienceBase. For the NGDB, the samples are being registered with IGSNs in SESAR and the geochemical data are being published as Linked Data. This system allows the USGS collections to benefit from disciplinary and institutional strengths of the participating resources, while simultaneously increasing the discovery, accessibility, and citation of USGS physical collection holdings.
Stauffer, Andrew J.; Webinger, Seth; Roche, Brittany
2016-01-01
The US Geological Survey’s (USGS) National Geospatial Technical Operations Center is prototyping and evaluating the ability to filter data through a range of scales using 1:24,000-scale The National Map (TNM) datasets as the source. A “VisibilityFilter” attribute is under evaluation that can be added to all TNM vector data themes and will permit filtering of data to eight target scales between 1:24,000 and 1:5,000,000, thus defining each feature’s smallest applicable scale-of-use. For a prototype implementation, map specifications for 1:100,000- and 1:250,000-scale USGS Topographic Map Series are being utilized to define feature content appropriate at fixed mapping scales to guide generalization decisions that are documented in a ScaleMaster diagram. This paper defines the VisibilityFilter attribute, the generalization decisions made for each TNM data theme, and how these decisions are embedded into the data to support efficient data filtering.
Co-morbidities associated with influenza-attributed mortality, 1994-2000, Canada.
Schanzer, Dena L; Langley, Joanne M; Tam, Theresa W S
2008-08-26
The elderly and persons with specific chronic conditions are known to face elevated morbidity and mortality risks resulting from an influenza infection, and hence are routinely recommended for annual influenza vaccination. However, risk-specific mortality rates have not been established. We estimated age-specific influenza-attributable mortality rates stratified by the presence of chronic conditions and type of residence based on deaths of persons who were admitted to hospital with a respiratory complication captured in our national database. The majority of patients had chronic heart or respiratory conditions (80%) and were admitted from the community (80%). Influenza-attributable mortality rates clearly increase with age for all risk groups. Our influenza-specific estimates identified higher risk ratios for chronic lung or heart disease than have been suggested by other methods. These estimates identify groups most in need of improved vaccines and for whom the use of additional strategies, such as immunization of household contacts or caregivers should be considered.
Ahmadi, Farshid Farnood; Ebadi, Hamid
2009-01-01
3D spatial data acquired from aerial and remote sensing images by photogrammetric techniques is one of the most accurate and economic data sources for GIS, map production, and spatial data updating. However, there are still many problems concerning storage, structuring and appropriate management of spatial data obtained using these techniques. According to the capabilities of spatial database management systems (SDBMSs); direct integration of photogrammetric and spatial database management systems can save time and cost of producing and updating digital maps. This integration is accomplished by replacing digital maps with a single spatial database. Applying spatial databases overcomes the problem of managing spatial and attributes data in a coupled approach. This management approach is one of the main problems in GISs for using map products of photogrammetric workstations. Also by the means of these integrated systems, providing structured spatial data, based on OGC (Open GIS Consortium) standards and topological relations between different feature classes, is possible at the time of feature digitizing process. In this paper, the integration of photogrammetric systems and SDBMSs is evaluated. Then, different levels of integration are described. Finally design, implementation and test of a software package called Integrated Photogrammetric and Oracle Spatial Systems (IPOSS) is presented.
Individual differences and their measurement: A review of 100 years of research.
Sackett, Paul R; Lievens, Filip; Van Iddekinge, Chad H; Kuncel, Nathan R
2017-03-01
This article reviews 100 years of research on individual differences and their measurement, with a focus on research published in the Journal of Applied Psychology. We focus on 3 major individual differences domains: (a) knowledge, skill, and ability, including both the cognitive and physical domains; (b) personality, including integrity, emotional intelligence, stable motivational attributes (e.g., achievement motivation, core self-evaluations), and creativity; and (c) vocational interests. For each domain, we describe the evolution of the domain across the years and highlight major theoretical, empirical, and methodological developments, including relationships between individual differences and variables such as job performance, job satisfaction, and career development. We conclude by discussing future directions for individual differences research. Trends in the literature include a growing focus on substantive issues rather than on the measurement of individual differences, a differentiation between constructs and measurement methods, and the use of innovative ways of assessing individual differences, such as simulations, other-reports, and implicit measures. (PsycINFO Database Record (c) 2017 APA, all rights reserved).
Development of a 3D GIS and its application to karst areas
NASA Astrophysics Data System (ADS)
Wu, Qiang; Xu, Hua; Zhou, Wanfang
2008-05-01
There is a growing interest in modeling and analyzing karst phenomena in three dimensions. This paper integrates geology, groundwater hydrology, geographic information system (GIS), database management system (DBMS), visualization and data mining to study karst features in Huaibei, China. The 3D geo-objects retrieved from the karst area are analyzed and mapped into different abstract levels. The spatial relationships among the objects are constructed by a dual-linker. The shapes of the 3D objects and the topological models with attributes are stored and maintained in the DBMS. Spatial analysis was then used to integrate the data in the DBMS and the 3D model to form a virtual reality (VR) to provide analytical functions such as distribution analysis, correlation query, and probability assessment. The research successfully implements 3D modeling and analyses in the karst area, and meanwhile provides an efficient tool for government policy-makers to set out restrictions on water resource development in the area.
Proud Americans and lucky Japanese: cultural differences in appraisal and corresponding emotion.
Imada, Toshie; Ellsworth, Phoebe C
2011-04-01
Appraisal theories of emotion propose that the emotions people experience correspond to their appraisals of their situation. In other words, individual differences in emotional experiences reflect differing interpretations of the situation. We hypothesized that in similar situations, people in individualist and collectivist cultures experience different emotions because of culturally divergent causal attributions for success and failure (i.e., agency appraisals). In a test of this hypothesis, American and Japanese participants recalled a personal experience (Study 1) or imagined themselves to be in a situation (Study 2) in which they succeeded or failed, and then reported their agency appraisals and emotions. Supporting our hypothesis, cultural differences in emotions corresponded to differences in attributions. For example, in success situations, Americans reported stronger self-agency emotions (e.g., proud) than did Japanese, whereas Japanese reported a stronger situation-agency emotion (lucky). Also, cultural differences in attribution and emotion were largely explained by differences in self-enhancing motivation. When Japanese and Americans were induced to make the same attribution (Study 2), cultural differences in emotions became either nonsignificant or were markedly reduced. PsycINFO Database Record (c) 2011 APA, all rights reserved.
Diverse expected gradient active learning for relative attributes.
You, Xinge; Wang, Ruxin; Tao, Dacheng
2014-07-01
The use of relative attributes for semantic understanding of images and videos is a promising way to improve communication between humans and machines. However, it is extremely labor- and time-consuming to define multiple attributes for each instance in large amount of data. One option is to incorporate active learning, so that the informative samples can be actively discovered and then labeled. However, most existing active-learning methods select samples one at a time (serial mode), and may therefore lose efficiency when learning multiple attributes. In this paper, we propose a batch-mode active-learning method, called diverse expected gradient active learning. This method integrates an informativeness analysis and a diversity analysis to form a diverse batch of queries. Specifically, the informativeness analysis employs the expected pairwise gradient length as a measure of informativeness, while the diversity analysis forces a constraint on the proposed diverse gradient angle. Since simultaneous optimization of these two parts is intractable, we utilize a two-step procedure to obtain the diverse batch of queries. A heuristic method is also introduced to suppress imbalanced multiclass distributions. Empirical evaluations of three different databases demonstrate the effectiveness and efficiency of the proposed approach.
Diverse Expected Gradient Active Learning for Relative Attributes.
You, Xinge; Wang, Ruxin; Tao, Dacheng
2014-06-02
The use of relative attributes for semantic understanding of images and videos is a promising way to improve communication between humans and machines. However, it is extremely labor- and time-consuming to define multiple attributes for each instance in large amount of data. One option is to incorporate active learning, so that the informative samples can be actively discovered and then labeled. However, most existing active-learning methods select samples one at a time (serial mode), and may therefore lose efficiency when learning multiple attributes. In this paper, we propose a batch-mode active-learning method, called Diverse Expected Gradient Active Learning (DEGAL). This method integrates an informativeness analysis and a diversity analysis to form a diverse batch of queries. Specifically, the informativeness analysis employs the expected pairwise gradient length as a measure of informativeness, while the diversity analysis forces a constraint on the proposed diverse gradient angle. Since simultaneous optimization of these two parts is intractable, we utilize a two-step procedure to obtain the diverse batch of queries. A heuristic method is also introduced to suppress imbalanced multi-class distributions. Empirical evaluations of three different databases demonstrate the effectiveness and efficiency of the proposed approach.
Aktürk, Hacer; Sütçü, Murat; Somer, Ayper; Karaman, Serap; Acar, Manolya; Ünüvar, Ayşegül; Anak, Sema; Karakaş, Zeynep; Özdemir, Aslı; Sarsar, Kutay; Aydın, Derya; Salman, Nuran
2016-09-05
To investigate the clinical impact of vancomycin-resistant enterococci (VRE) colonization in patients with hematologic malignancies and associated risk factors. Patients colonized and infected with VRE were identified from an institutional surveillance database between January 2010 and December 2013. A retrospective case-control study was performed to identify the risk factors associated with development of VRE infection in VRE-colonized patients. Fecal VRE colonization was documented in 72 of 229 children (31.4%). Seven VRE-colonized patients developed subsequent systemic VRE infection (9.7%). Types of VRE infections included bacteremia (n=5), urinary tract infection (n=1), and meningitis (n=1). Enterococcus faecium was isolated in all VRE infections. Multivariate analysis revealed severe neutropenia and previous bacteremia with another pathogen as independent risk factors for VRE infection development in colonized patients [odds ratio (OR): 35.4, confidence interval (CI): 1.7-72.3, p=0.02 and OR: 20.6, CI: 1.3-48.6, p=0.03, respectively]. No deaths attributable to VRE occurred. VRE colonization has important consequences in pediatric cancer patients.
Cyberbullying in Adolescence: A Concept Analysis.
Hutson, Elizabeth
2016-01-01
The aim of this article was to present a detailed analysis of the concept of cyberbullying. Research on the topic of cyberbullying is growing exponentially, but not all studies use the same definition to examine this concept. A concept analysis in the style of Walker and Avant was used to analyze cyberbullying. Literature was retrieved from the databases of CINAHL and PubMed between the years 2009 and October 2014 using the key word "cyberbullying." Twenty-five English-language articles were located that delineated a definition of cyberbullying. First, articles were analyzed to discover the defining attributes of cyberbullying. Second, antecedents, consequences, and related terms of cyberbullying were examined. Third, an operational definition of cyberbullying is proposed on the basis of a consensus of the review. This review proposes that the defining attributes of cyberbullying are: (1) electronic form of contact (2) an aggressive act (3) intent (4) repetition (publicity), and (5) harm of the victim. The antecedents most often mentioned were lower self-esteem, higher levels of depression, and social isolation and the consequences were academic problems and affective disorders. A single concise definition of cyberbullying was created that can be utilized by health care providers to educate their patients and families about cyberbullying. This definition also can be used to guide research to develop effective interventions.
[Predatory journals: how their publishers operate and how to avoid them].
Kratochvíl, Jiří; Plch, Lukáš
Authors who publish in scientific or scholarly journals today face the risk of publishing in so-called predatory journals. These journals exploit the noble idea of the Open Access movement, whose goal is to make the latest scientific findings available for free. Predatory journals, unlike the reputable ones working on an Open Access basis, neglect the review process and publish low-quality submissions. The basic attributes of predatory journals are a very quick review process or even none at all, failure to be transparent about author fees for publishing an article, misleading potential authors by imitating the names of well-established journals, and false information on indexing in renowned databases or assigned impact factor. Some preventive measures against publishing in predatory journals or drawing information from them are: a thorough credibility check of the journals webpages, verification of the journals indexing on Bealls List and in the following databases: Web of Science Core Collection, Scopus, ERIH PLUS and DOAJ. Asking other scientists or scholars about their experience with a given journal can also be helpful. Without these necessary steps authors face an increased risk of publishing in a journal of poor quality, which will prevent them from obtaining Research and Development Council points (awarded based on the Information Register of Research & Development results); even more importantly, it may damage their reputation as well as the good name of their home institution in the professional community.Key words: academic writing - medical journals - Open Access - predatory journals - predatory publishers - scientific publications.
Gene Fusion Markup Language: a prototype for exchanging gene fusion data.
Kalyana-Sundaram, Shanker; Shanmugam, Achiraman; Chinnaiyan, Arul M
2012-10-16
An avalanche of next generation sequencing (NGS) studies has generated an unprecedented amount of genomic structural variation data. These studies have also identified many novel gene fusion candidates with more detailed resolution than previously achieved. However, in the excitement and necessity of publishing the observations from this recently developed cutting-edge technology, no community standardization approach has arisen to organize and represent the data with the essential attributes in an interchangeable manner. As transcriptome studies have been widely used for gene fusion discoveries, the current non-standard mode of data representation could potentially impede data accessibility, critical analyses, and further discoveries in the near future. Here we propose a prototype, Gene Fusion Markup Language (GFML) as an initiative to provide a standard format for organizing and representing the significant features of gene fusion data. GFML will offer the advantage of representing the data in a machine-readable format to enable data exchange, automated analysis interpretation, and independent verification. As this database-independent exchange initiative evolves it will further facilitate the formation of related databases, repositories, and analysis tools. The GFML prototype is made available at http://code.google.com/p/gfml-prototype/. The Gene Fusion Markup Language (GFML) presented here could facilitate the development of a standard format for organizing, integrating and representing the significant features of gene fusion data in an inter-operable and query-able fashion that will enable biologically intuitive access to gene fusion findings and expedite functional characterization. A similar model is envisaged for other NGS data analyses.
Yilmaz, T; Cordero-Coma, M; Gallagher, M J
2012-01-01
To assess the effectiveness of ketorolac vs control for prevention of acute pseudophakic cystoid macular edema (CME). The following databases were searched: Medline (1950–June 11, 2011), The Cochrane Library (Issue 2, 2011), and the TRIP Database (up to 11 June 2011), using no language or other limits. Randomized controlled clinical trials (RCTs) were included that consisted of patients with acute pseudophakic cystoid macular edema, those comparing ketorolac with control, and those having at least a minimum follow-up of 28 days. In the four RCTs evaluating ketorolac vs control, treatment with ketorolac significantly reduced the risk of CME development at the end of treatment (∼4 weeks) compared to control (P=0.008; 95% confidence interval (0.03–0.58)). When analyzed individually, each individual study was statistically nonsignificant in its findings with the exception of one study. When the pooled relative risk was calculated, the large sample size of this systematic review led to overall statistical significance, which is attributable to the review's large sample size and not to the individual studies themselves. In this systematic review of four RCTs, two of which compared ketorolac with no treatment and two of which evaluated ketorolac vs placebo drops, treatment with ketorolac significantly reduced the risk of developing CME at the end of ∼4 weeks of treatment compared with controls. These results, however, should be interpreted with caution considering the paucity of large randomized clinical trials in the literature. PMID:22094296
The clinical learning environment in nursing education: a concept analysis.
Flott, Elizabeth A; Linden, Lois
2016-03-01
The aim of this study was to report an analysis of the clinical learning environment concept. Nursing students are evaluated in clinical learning environments where skills and knowledge are applied to patient care. These environments affect achievement of learning outcomes, and have an impact on preparation for practice and student satisfaction with the nursing profession. Providing clarity of this concept for nursing education will assist in identifying antecedents, attributes and consequences affecting student transition to practice. The clinical learning environment was investigated using Walker and Avant's concept analysis method. A literature search was conducted using WorldCat, MEDLINE and CINAHL databases using the keywords clinical learning environment, clinical environment and clinical education. Articles reviewed were written in English and published in peer-reviewed journals between 1995-2014. All data were analysed for recurring themes and terms to determine possible antecedents, attributes and consequences of this concept. The clinical learning environment contains four attribute characteristics affecting student learning experiences. These include: (1) the physical space; (2) psychosocial and interaction factors; (3) the organizational culture and (4) teaching and learning components. These attributes often determine achievement of learning outcomes and student self-confidence. With better understanding of attributes comprising the clinical learning environment, nursing education programmes and healthcare agencies can collaborate to create meaningful clinical experiences and enhance student preparation for the professional nurse role. © 2015 John Wiley & Sons Ltd.
Vázquez, José Juan; Panadero, Sonia; Zúñiga, Claudia
2018-01-01
Causal attributions of homelessness may affect both the design and acceptance of public policies aimed at improving the situation of homeless people and the strategies that homeless people themselves decide to adopt in order to cope with their situation. This article analyzes the differences in causal attributions of homelessness based on gender, age, nationality, educational background, perceived social class, evolution of personal economic situation, and future expectations between the members of 2 groups: (a) "homeless group", consisting of a representative sample of homeless people in Madrid, Spain (n = 188); and (b) "domiciled group", consisting of a sample of people in Madrid at no risk of homelessness (n = 180), matched for sex, age and nationality. Results show that among domiciled population, women, older people, those without university education, those considering themselves to belong to lower income social classes, those who considered their economic situation to have worsened, and those who expressed negative expectations for the future attributed homelessness to individualistic courses to a greater extent. Meanwhile, among homeless group, younger people, those without university education, those considering themselves to belong to higher social classes, those who perceived their economic situation as having improved in recent years, and those who expressed positive expectations for the future generally attributed homelessness to individualistic courses to a greater extent. (PsycINFO Database Record (c) 2018 APA, all rights reserved).
Genetic essentialism: on the deceptive determinism of DNA.
Dar-Nimrod, Ilan; Heine, Steven J
2011-09-01
This article introduces the notion of genetic essentialist biases: cognitive biases associated with essentialist thinking that are elicited when people encounter arguments that genes are relevant for a behavior, condition, or social group. Learning about genetic attributions for various human conditions leads to a particular set of thoughts regarding those conditions: they are more likely to be perceived as (a) immutable and determined, (b) having a specific etiology, (c) homogeneous and discrete, and (d) natural, which can lead to the naturalistic fallacy. There are rare cases of "strong genetic explanation" when such responses to genetic attributions may be appropriate; however, people tend to overweigh genetic attributions compared with competing attributions even in cases of "weak genetic explanation," which are far more common. The authors reviewed research on people's understanding of race, gender, sexual orientation, criminality, mental illness, and obesity through a genetic essentialism lens, highlighting attitudinal, cognitive, and behavioral changes that stem from consideration of genetic attributions as bases of these categories. Scientific and media portrayals of genetic discoveries are discussed with respect to genetic essentialism, as is the role that genetic essentialism has played (and continues to play) in various public policies, legislation, scientific endeavors, and ideological movements in recent history. Last, moderating factors and interventions to reduce the magnitude of genetic essentialism, which identify promising directions to explore in order to reduce these biases, are discussed. (PsycINFO Database Record (c) 2011 APA, all rights reserved).
Compression technique for large statistical data bases
DOE Office of Scientific and Technical Information (OSTI.GOV)
Eggers, S.J.; Olken, F.; Shoshani, A.
1981-03-01
The compression of large statistical databases is explored and are proposed for organizing the compressed data, such that the time required to access the data is logarithmic. The techniques exploit special characteristics of statistical databases, namely, variation in the space required for the natural encoding of integer attributes, a prevalence of a few repeating values or constants, and the clustering of both data of the same length and constants in long, separate series. The techniques are variations of run-length encoding, in which modified run-lengths for the series are extracted from the data stream and stored in a header, which ismore » used to form the base level of a B-tree index into the database. The run-lengths are cumulative, and therefore the access time of the data is logarithmic in the size of the header. The details of the compression scheme and its implementation are discussed, several special cases are presented, and an analysis is given of the relative performance of the various versions.« less
Seabird databases and the new paradigm for scientific publication and attribution
Hatch, Scott A.
2010-01-01
For more than 300 years, the peer-reviewed journal article has been the principal medium for packaging and delivering scientific data. With new tools for managing digital data, a new paradigm is emerging—one that demands open and direct access to data and that enables and rewards a broad-based approach to scientific questions. Ground-breaking papers in the future will increasingly be those that creatively mine and synthesize vast stores of data available on the Internet. This is especially true for conservation science, in which essential data can be readily captured in standard record formats. For seabird professionals, a number of globally shared databases are in the offing, or should be. These databases will capture the salient results of inventories and monitoring, pelagic surveys, diet studies, and telemetry. A number of real or perceived barriers to data sharing exist, but none is insurmountable. Our discipline should take an important stride now by adopting a specially designed markup language for annotating and sharing seabird data.
usSEABED: Pacific coast (California, Oregon, Washington) offshore surficial-sediment data release
Reid, Jane A.; Reid, Jamey M.; Jenkins, Chris J.; Zimmermann, Mark; Williams, S. Jeffress; Field, Michael E.
2006-01-01
Over the past 50 years there has been an explosion in scientific interest, research effort, and information gathered on the geologic sedimentary character of the continental margin of the United States. Data and information from thousands of publications have greatly increased our scientific understanding of the geologic origins of the margin surface but rarely have those data been combined and integrated. This publication is the first release of the Pacific coast data from the usSEABED database. The report contains a compilation of published and unpublished sediment texture and other geologic data about the sea floor from diverse sources. usSEABED is an innovative database system developed to unify assorted data, with the data processed by the dbSEABED system. Examples of maps displaying attributes such as grain size and sediment color are included. This database contains information that is a scientific foundation for the U.S. Geological Survey (USGS) Sea floor Mapping and Benthic Habitats project and the Marine Aggregate Resources and Processes assessment project, and will be useful to the marine science community for other studies of the Pacific coast continental margin. The publication is divided into 10 sections: Home, Introduction, Content, usSEABED (data), dbSEABED (processing), Data Catalog, References, Contacts, Acknowledgments, and Frequently Asked Questions. Use the navigation bar on the left to navigate to specific sections of this report. Underlined topics throughout the publication are links to more information. Links to specific and detailed information on processing and to those to pages outside this report will open in a new browser window.
NASA Astrophysics Data System (ADS)
Liu, S.; Wei, Y.; Post, W. M.; Cook, R. B.; Schaefer, K.; Thornton, M. M.
2013-05-01
The Unified North American Soil Map (UNASM) was developed to provide more accurate regional soil information for terrestrial biosphere modeling. The UNASM combines information from state-of-the-art US STATSGO2 and Soil Landscape of Canada (SLCs) databases. The area not covered by these datasets is filled by using the Harmonized World Soil Database version 1.21 (HWSD1.21). The UNASM contains maximum soil depth derived from the data source as well as seven soil attributes (including sand, silt, and clay content, gravel content, organic carbon content, pH, and bulk density) for the topsoil layer (0-30 cm) and the subsoil layer (30-100 cm), respectively, of the spatial resolution of 0.25 degrees in latitude and longitude. There are pronounced differences in the spatial distributions of soil properties and soil organic carbon between UNASM and HWSD, but the UNASM overall provides more detailed and higher-quality information particularly in Alaska and central Canada. To provide more accurate and up-to-date estimate of soil organic carbon stock in North America, we incorporated Northern Circumpolar Soil Carbon Database (NCSCD) into the UNASM. The estimate of total soil organic carbon mass in the upper 100 cm soil profile based on the improved UNASM is 365.96 Pg, of which 23.1% is under trees, 14.1% is in shrubland, and 4.6% is in grassland and cropland. This UNASM data will provide a resource for use in terrestrial ecosystem modeling both for input of soil characteristics and for benchmarking model output.
NASA Astrophysics Data System (ADS)
Liu, S.; Wei, Y.; Post, W. M.; Cook, R. B.; Schaefer, K.; Thornton, M. M.
2012-10-01
The Unified North American Soil Map (UNASM) was developed to provide more accurate regional soil information for terrestrial biosphere modeling. The UNASM combines information from state-of-the-art US STATSGO2 and Soil Landscape of Canada (SLCs) databases. The area not covered by these datasets is filled with the Harmonized World Soil Database version 1.1 (HWSD1.1). The UNASM contains maximum soil depth derived from the data source as well as seven soil attributes (including sand, silt, and clay content, gravel content, organic carbon content, pH, and bulk density) for the top soil layer (0-30 cm) and the sub soil layer (30-100 cm) respectively, of the spatial resolution of 0.25° in latitude and longitude. There are pronounced differences in the spatial distributions of soil properties and soil organic carbon between UNASM and HWSD, but the UNASM overall provides more detailed and higher-quality information particularly in Alaska and Central Canada. To provide more accurate and up-to-date estimate of soil organic carbon stock in North America, we incorporated Northern Circumpolar Soil Carbon Database (NCSCD) into the UNASM. The estimate of total soil organic carbon mass in the upper 100 cm soil profile based on the improved UNASM is 347.70 Pg, of which 24.7% is under trees, 14.2% is under shrubs, and 1.3% is under grasses and 3.8% under crops. This UNASM data will provide a resource for use in land surface and terrestrial biogeochemistry modeling both for input of soil characteristics and for benchmarking model output.
A database for assessment of effect of lossy compression on digital mammograms
NASA Astrophysics Data System (ADS)
Wang, Jiheng; Sahiner, Berkman; Petrick, Nicholas; Pezeshk, Aria
2018-03-01
With widespread use of screening digital mammography, efficient storage of the vast amounts of data has become a challenge. While lossless image compression causes no risk to the interpretation of the data, it does not allow for high compression rates. Lossy compression and the associated higher compression ratios are therefore more desirable. The U.S. Food and Drug Administration (FDA) currently interprets the Mammography Quality Standards Act as prohibiting lossy compression of digital mammograms for primary image interpretation, image retention, or transfer to the patient or her designated recipient. Previous work has used reader studies to determine proper usage criteria for evaluating lossy image compression in mammography, and utilized different measures and metrics to characterize medical image quality. The drawback of such studies is that they rely on a threshold on compression ratio as the fundamental criterion for preserving the quality of images. However, compression ratio is not a useful indicator of image quality. On the other hand, many objective image quality metrics (IQMs) have shown excellent performance for natural image content for consumer electronic applications. In this paper, we create a new synthetic mammogram database with several unique features. We compare and characterize the impact of image compression on several clinically relevant image attributes such as perceived contrast and mass appearance for different kinds of masses. We plan to use this database to develop a new objective IQM for measuring the quality of compressed mammographic images to help determine the allowed maximum compression for different kinds of breasts and masses in terms of visual and diagnostic quality.
DFAS Needs More Effective Controls Over Managing DoD Contractor Debt
2011-07-15
Acrobat fil e onl y) to audc lev@dod ig.m il. Copies of management comments must have the actual signature of the authori zing official for yo ur...Attribute Testing Results 18 E. Sampling Methodology 20 Glossary of Technical Terms 21 Management...Bill of Collection number to each contractor debt as a debt management tool. CDS allows users to prepare a demand letter from the database, post
Molecular Biogeochemistry of Modern and Ancient Marine Microbes
2010-02-01
number distributions in the late Archean bitumens fall within the range of compositions of Phanerozoic petroleum ( gray line in Fig. 7), suggesting that...bitumen extracts. The gray line indicates the range of compositions observed in Phanerozoic petroleum systems, from the GeoMark Reservoir Fluid Database...than that of mRNA are attributable to noisy, non-cycling protein timecourses ( gray points above 1:1 line). For clarity, only genes whose protein
An infrastructure to mine molecular descriptors for ligand selection on virtual screening.
Seus, Vinicius Rosa; Perazzo, Giovanni Xavier; Winck, Ana T; Werhli, Adriano V; Machado, Karina S
2014-01-01
The receptor-ligand interaction evaluation is one important step in rational drug design. The databases that provide the structures of the ligands are growing on a daily basis. This makes it impossible to test all the ligands for a target receptor. Hence, a ligand selection before testing the ligands is needed. One possible approach is to evaluate a set of molecular descriptors. With the aim of describing the characteristics of promising compounds for a specific receptor we introduce a data warehouse-based infrastructure to mine molecular descriptors for virtual screening (VS). We performed experiments that consider as target the receptor HIV-1 protease and different compounds for this protein. A set of 9 molecular descriptors are taken as the predictive attributes and the free energy of binding is taken as a target attribute. By applying the J48 algorithm over the data we obtain decision tree models that achieved up to 84% of accuracy. The models indicate which molecular descriptors and their respective values are relevant to influence good FEB results. Using their rules we performed ligand selection on ZINC database. Our results show important reduction in ligands selection to be applied in VS experiments; for instance, the best selection model picked only 0.21% of the total amount of drug-like ligands.
NASA Astrophysics Data System (ADS)
Thies, Christian; Ostwald, Tamara; Fischer, Benedikt; Lehmann, Thomas M.
2005-04-01
The classification and measuring of objects in medical images is important in radiological diagnostics and education, especially when using large databases as knowledge resources, for instance a picture archiving and communication system (PACS). The main challenge is the modeling of medical knowledge and the diagnostic context to label the sought objects. This task is referred to as closing the semantic gap between low-level pixel information and high level application knowledge. This work describes an approach which allows labeling of a-priori unknown objects in an intuitive way. Our approach consists of four main components. At first an image is completely decomposed into all visually relevant partitions on different scales. This provides a hierarchical organized set of regions. Afterwards, for each of the obtained regions a set of descriptive features is computed. In this data structure objects are represented by regions with characteristic attributes. The actual object identification is the formulation of a query. It consists of attributes on which intervals are defined describing those regions that correspond to the sought objects. Since the objects are a-priori unknown, they are described by a medical expert by means of an intuitive graphical user interface (GUI). This GUI is the fourth component. It enables complex object definitions by browsing the data structure and examinating the attributes to formulate the query. The query is executed and if the sought objects have not been identified its parameterization is refined. By using this heuristic approach, object models for hand radiographs have been developed to extract bones from a single hand in different anatomical contexts. This demonstrates the applicability of the labeling concept. By using a rule for metacarpal bones on a series of 105 images, this type of bone could be retrieved with a precision of 0.53 % and a recall of 0.6%.
Development and Validation of the Poverty Attributions Survey
ERIC Educational Resources Information Center
Bennett, Robert M.; Raiz, Lisa; Davis, Tamara S.
2016-01-01
This article describes the process of developing and testing the Poverty Attribution Survey (PAS), a measure of poverty attributions. The PAS is theory based and includes original items as well as items from previously tested poverty attribution instruments. The PAS was electronically administered to a sample of state-licensed professional social…
University Real Estate Development Database: A Database-Driven Internet Research Tool
ERIC Educational Resources Information Center
Wiewel, Wim; Kunst, Kara
2008-01-01
The University Real Estate Development Database is an Internet resource developed by the University of Baltimore for the Lincoln Institute of Land Policy, containing over six hundred cases of university expansion outside of traditional campus boundaries. The University Real Estate Development database is a searchable collection of real estate…
Dwyer, Johanna T.; Picciano, Mary Frances; Betz, Joseph M.; Fisher, Kenneth D.; Saldanha, Leila G.; Yetley, Elizabeth A.; Coates, Paul M.; Radimer, Kathy; Bindewald, Bernadette; Sharpless, Katherine E.; Holden, Joanne; Andrews, Karen; Zhao, Cuiwei; Harnly, James; Wolf, Wayne R.; Perry, Charles R.
2013-01-01
Several activities of the Office of Dietary Supplements (ODS) at the National Institutes of Health involve enhancement of dietary supplement databases. These include an initiative with US Department of Agriculture to develop an analytically substantiated dietary supplement ingredient database (DSID) and collaboration with the National Center for Health Statistics to enhance the dietary supplement label database in the National Health and Nutrition Examination Survey (NHANES). The many challenges that must be dealt with in developing an analytically supported DSID include categorizing product types in the database, identifying nutrients, and other components of public health interest in these products and prioritizing which will be entered in the database first. Additional tasks include developing methods and reference materials for quantifying the constituents, finding qualified laboratories to measure the constituents, developing appropriate sample handling procedures, and finally developing representative sampling plans. Developing the NHANES dietary supplement label database has other challenges such as collecting information on dietary supplement use from NHANES respondents, constant updating and refining of information obtained, developing default values that can be used if the respondent cannot supply the exact supplement or strength that was consumed, and developing a publicly available label database. Federal partners and the research community are assisting in making an analytically supported dietary supplement database a reality. PMID:25309034
NASA Astrophysics Data System (ADS)
Essa, Salem M.; Loughland, R.; Khogali, Mohamed E.
2005-10-01
AL Sammalyah Island is considered an important protected area in Abu Dhabi Emirate. The island has witnessed high rates of change in land use in the past few years starting from the early 1990s. Change detection analysis is conducted to monitor rate and spatial distribution of change occurring on the island. A three-phase research project has been implemented, an integrated Geographic Information System (GIS) database for the Island is the focus; the current phase main objective was to assess rate and spatial distribution of the change on the island using multi-date large scale aerial photos. Results of the current study demonstrated that total vegetation cover extent has increased from 3.742 km2 in 1994 to 5.101 km2 in 2005, an increase of 36.3% between 1994 and 2005. The study also showed that this increase in vegetation extent is mostly attributed to the increase in mangrove planted areas with an increase from 2.256 km2 in 1994 to 3.568 km2 in 2005, an increase of 58.2% in ten years. Remote sensing and GIS have been successfully used to quantify change extent, distribution and trajectories of change. The next step will be to complete the GIS database for AL Sammalyah Island.
Delalibera, Mayra; Presa, Joana; Barbosa, António; Leal, Isabel
2015-09-01
Caring for a family member with an advanced and/or terminal illness can be a great emotional, physical and financial burden that has an impact on the quality of life of the caregivers. The scope of this study was to conduct a systematic review of the literature on the burden of caregiving, related factors and the consequences for family caregivers of advanced stage cancer patients or patients in end-of-life or palliative care. A search for scientific papers published in the EBSCO, Web of Knowledge and BIREME databases was conducted since records on this topic began in the databases through March 2014. Of the 582 articles found, only 27 were selected. The majority of the articles found that family caregivers were overburdened. Some studies found that the care-giving burden was associated with characteristics of the patients and their illnesses while, in other studies, it was associated with poor health of the caregiver, greater psychopathological symptoms (anxiety, depression, emotional distress) and with the development of complications in the grieving process. However, hope, social support, the ability of the caregiver to attribute meaning to the experience of caring and feeling comfortable with the tasks of caring were associated with lower levels of burden.
Blok, Amanda C
2017-04-01
To report an analysis of the concept of self-management behaviors. Self-management behaviors are typically associated with disease management, with frequent use by nurse researchers related to chronic illness management and by international health organizations for development of disease management interventions. A concept analysis was conducted within the context of Orem's self-care framework. Walker and Avant's eight-step concept analysis approach guided the analysis. Academic databases were searched for relevant literature including CIHAHL, Cochrane Databases of Systematic Reviews and Register of Controlled Trials, MEDLINE, PsycARTICLES and PsycINFO, and SocINDEX. Literature using the term "self-management behavior" and published between April 2001 and March 2015 was analyzed for attributes, antecedents, and consequences. A total of 189 journal articles were reviewed. Self-management behaviors are defined as proactive actions related to lifestyle, a problem, planning, collaborating, and mental support, as well as reactive actions related to a circumstantial change, to achieve a goal influenced by the antecedents of physical, psychological, socioeconomic, and cultural characteristics, as well as collaborative and received support. The theoretical definition and middle-range explanatory theory of self-management behaviors will guide future collaborative research and clinical practice for disease management. © 2016 Wiley Periodicals, Inc.
Design and Implementation of a Metadata-rich File System
DOE Office of Scientific and Technical Information (OSTI.GOV)
Ames, S; Gokhale, M B; Maltzahn, C
2010-01-19
Despite continual improvements in the performance and reliability of large scale file systems, the management of user-defined file system metadata has changed little in the past decade. The mismatch between the size and complexity of large scale data stores and their ability to organize and query their metadata has led to a de facto standard in which raw data is stored in traditional file systems, while related, application-specific metadata is stored in relational databases. This separation of data and semantic metadata requires considerable effort to maintain consistency and can result in complex, slow, and inflexible system operation. To address thesemore » problems, we have developed the Quasar File System (QFS), a metadata-rich file system in which files, user-defined attributes, and file relationships are all first class objects. In contrast to hierarchical file systems and relational databases, QFS defines a graph data model composed of files and their relationships. QFS incorporates Quasar, an XPATH-extended query language for searching the file system. Results from our QFS prototype show the effectiveness of this approach. Compared to the de facto standard, the QFS prototype shows superior ingest performance and comparable query performance on user metadata-intensive operations and superior performance on normal file metadata operations.« less
NASA Astrophysics Data System (ADS)
Aleman, A.; Olsen, L. M.; Ritz, S.; Stevens, T.; Morahan, M.; Grebas, S. K.
2011-12-01
NASA's Global Change Master Directory provides the scientific community with the ability to discover, access, and use Earth science data, data-related services, and climate diagnostics worldwide.The GCMD offers descriptions of Earth science data sets using the Directory Interchange Format (DIF) metadata standard; Earth science related data services are described using the Service Entry Resource Format (SERF); and climate visualizations are described using the Climate Diagnostic (CD) standard. The DIF, SERF and CD standards each capture data attributes used to determine whether a data set, service, or climate visualization is relevant to a user's needs.Metadata fields include: title, summary, science keywords, service keywords, data center, data set citation, personnel, instrument, platform, quality, related URL, temporal and spatial coverage, data resolution and distribution information.In addition, nine valuable sets of controlled vocabularies have been developed to assist users in normalizing the search for data descriptions. An update to the GCMD's search functionality is planned to further capitalize on the controlled vocabularies during database queries.By implementing a dynamic keyword "tree", users will have the ability to search for data sets by combining keywords in new ways.This will allow users to conduct more relevant and efficient database searches to support the free exchange and re-use of Earth science data.
Keshtiari, Niloofar; Kuhlmann, Michael; Eslami, Moharram; Klann-Delius, Gisela
2015-03-01
Research on emotional speech often requires valid stimuli for assessing perceived emotion through prosody and lexical content. To date, no comprehensive emotional speech database for Persian is officially available. The present article reports the process of designing, compiling, and evaluating a comprehensive emotional speech database for colloquial Persian. The database contains a set of 90 validated novel Persian sentences classified in five basic emotional categories (anger, disgust, fear, happiness, and sadness), as well as a neutral category. These sentences were validated in two experiments by a group of 1,126 native Persian speakers. The sentences were articulated by two native Persian speakers (one male, one female) in three conditions: (1) congruent (emotional lexical content articulated in a congruent emotional voice), (2) incongruent (neutral sentences articulated in an emotional voice), and (3) baseline (all emotional and neutral sentences articulated in neutral voice). The speech materials comprise about 470 sentences. The validity of the database was evaluated by a group of 34 native speakers in a perception test. Utterances recognized better than five times chance performance (71.4 %) were regarded as valid portrayals of the target emotions. Acoustic analysis of the valid emotional utterances revealed differences in pitch, intensity, and duration, attributes that may help listeners to correctly classify the intended emotion. The database is designed to be used as a reliable material source (for both text and speech) in future cross-cultural or cross-linguistic studies of emotional speech, and it is available for academic research purposes free of charge. To access the database, please contact the first author.
HAEdb: a novel interactive, locus-specific mutation database for the C1 inhibitor gene.
Kalmár, Lajos; Hegedüs, Tamás; Farkas, Henriette; Nagy, Melinda; Tordai, Attila
2005-01-01
Hereditary angioneurotic edema (HAE) is an autosomal dominant disorder characterized by episodic local subcutaneous and submucosal edema and is caused by the deficiency of the activated C1 esterase inhibitor protein (C1-INH or C1INH; approved gene symbol SERPING1). Published C1-INH mutations are represented in large universal databases (e.g., OMIM, HGMD), but these databases update their data rather infrequently, they are not interactive, and they do not allow searches according to different criteria. The HAEdb, a C1-INH gene mutation database (http://hae.biomembrane.hu) was created to contribute to the following expectations: 1) help the comprehensive collection of information on genetic alterations of the C1-INH gene; 2) create a database in which data can be searched and compared according to several flexible criteria; and 3) provide additional help in new mutation identification. The website uses MySQL, an open-source, multithreaded, relational database management system. The user-friendly graphical interface was written in the PHP web programming language. The website consists of two main parts, the freely browsable search function, and the password-protected data deposition function. Mutations of the C1-INH gene are divided in two parts: gross mutations involving DNA fragments >1 kb, and micro mutations encompassing all non-gross mutations. Several attributes (e.g., affected exon, molecular consequence, family history) are collected for each mutation in a standardized form. This database may facilitate future comprehensive analyses of C1-INH mutations and also provide regular help for molecular diagnostic testing of HAE patients in different centers.
Privacy-preserving matching of similar patients.
Vatsalan, Dinusha; Christen, Peter
2016-02-01
The identification of similar entities represented by records in different databases has drawn considerable attention in many application areas, including in the health domain. One important type of entity matching application that is vital for quality healthcare analytics is the identification of similar patients, known as similar patient matching. A key component of identifying similar records is the calculation of similarity of the values in attributes (fields) between these records. Due to increasing privacy and confidentiality concerns, using the actual attribute values of patient records to identify similar records across different organizations is becoming non-trivial because the attributes in such records often contain highly sensitive information such as personal and medical details of patients. Therefore, the matching needs to be based on masked (encoded) values while being effective and efficient to allow matching of large databases. Bloom filter encoding has widely been used as an efficient masking technique for privacy-preserving matching of string and categorical values. However, no work on Bloom filter-based masking of numerical data, such as integer (e.g. age), floating point (e.g. body mass index), and modulus (numbers wrap around upon reaching a certain value, e.g. date and time), which are commonly required in the health domain, has been presented in the literature. We propose a framework with novel methods for masking numerical data using Bloom filters, thereby facilitating the calculation of similarities between records. We conduct an empirical study on publicly available real-world datasets which shows that our framework provides efficient masking and achieves similar matching accuracy compared to the matching of actual unencoded patient records. Copyright © 2015 Elsevier Inc. All rights reserved.
Kleij, Kim-Sarah; Tangermann, Ulla; Amelung, Volker E; Krauth, Christian
2017-07-11
Primary care is a key element of health care systems and addresses the main health problems of the population. Due to the demographic change, primary care even gains in importance. The knowledge of the patients' preferences can help policy makers as well as physicians to set priorities in their effort to make health care delivery more responsive to patients' needs. Our objective was to describe which aspects of primary care were included in preference studies and which of them were the most preferred aspects. In order to elicit the preferences for primary care, a systematic literature search was conducted. Two researchers searched three electronic databases (PubMed, Scopus, and PsycINFO) and conducted a narrative synthesis. Inclusion criteria were: focus on primary health care delivery, discrete choice experiment as elicitation method, and studies published between 2006 and 2015 in English language. We identified 18 studies that elicited either the patients' or the population's preferences for primary care based on a discrete choice experiment. Altogether the studies used 16 structure attributes, ten process attributes and four outcome attributes. The most commonly applied structure attribute was "Waiting time till appointment", the most frequently used process attribute was "Shared decision making / professional's attention paid to your views". "Receiving the 'best' treatment" was the most commonly applied outcome attribute. Process attributes were most often the ones of highest importance for patients or the population. The attributes and attribute levels used in the discrete choice experiments were identified by literature research, qualitative research, expert interviews, or the analysis of policy documents. The results of the DCE studies show different preferences for primary health care. The diversity of the results may have several reasons, such as the method of analysis, the selection procedure of the attributes and their levels or the specific research question of the study. As the results of discrete choice experiments depend on many different factors, it is important for a better comprehensibility of the studies to transparently report the steps undertaken in a study as well as the interim results regarding the identification of attributes and levels.
Vaughn, Lisa M; McLinden, Daniel J; Shellmer, Diana; Baker, Raymond C
2011-01-01
The causes attributed to childhood health and illness across cultures (cultural health attributions) are key factors that are now more frequently identified as affecting the health outcomes of children. Research suggests that the causes attributed to an event such as illness are thought to affect subsequent motivation, emotional response, decision making, and behavior. To date, there is no measure of health attributions appropriate for use with parents of pediatric patients. Using the Many-Facets approach to Rasch analysis, this study assesses the psychometrics of a newly developed instrument, the Pediatric Health Attributions Questionnaire (Pedi-CHAQ), a measure designed to assess the cultural health attributions of parents in diverse communities. Results suggest acceptable Rasch model statistics of fit and reliability for the Pedi-CHAQ. A shortened version of the questionnaire was developed as a result of this study and next steps are discussed.
Gene: a gene-centered information resource at NCBI.
Brown, Garth R; Hem, Vichet; Katz, Kenneth S; Ovetsky, Michael; Wallin, Craig; Ermolaeva, Olga; Tolstoy, Igor; Tatusova, Tatiana; Pruitt, Kim D; Maglott, Donna R; Murphy, Terence D
2015-01-01
The National Center for Biotechnology Information's (NCBI) Gene database (www.ncbi.nlm.nih.gov/gene) integrates gene-specific information from multiple data sources. NCBI Reference Sequence (RefSeq) genomes for viruses, prokaryotes and eukaryotes are the primary foundation for Gene records in that they form the critical association between sequence and a tracked gene upon which additional functional and descriptive content is anchored. Additional content is integrated based on the genomic location and RefSeq transcript and protein sequence data. The content of a Gene record represents the integration of curation and automated processing from RefSeq, collaborating model organism databases, consortia such as Gene Ontology, and other databases within NCBI. Records in Gene are assigned unique, tracked integers as identifiers. The content (citations, nomenclature, genomic location, gene products and their attributes, phenotypes, sequences, interactions, variation details, maps, expression, homologs, protein domains and external databases) is available via interactive browsing through NCBI's Entrez system, via NCBI's Entrez programming utilities (E-Utilities and Entrez Direct) and for bulk transfer by FTP. Published by Oxford University Press on behalf of Nucleic Acids Research 2014. This work is written by (a) US Government employee(s) and is in the public domain in the US.
Optimizing a Query by Transformation and Expansion.
Glocker, Katrin; Knurr, Alexander; Dieter, Julia; Dominick, Friederike; Forche, Melanie; Koch, Christian; Pascoe Pérez, Analie; Roth, Benjamin; Ückert, Frank
2017-01-01
In the biomedical sector not only the amount of information produced and uploaded into the web is enormous, but also the number of sources where these data can be found. Clinicians and researchers spend huge amounts of time on trying to access this information and to filter the most important answers to a given question. As the formulation of these queries is crucial, automated query expansion is an effective tool to optimize a query and receive the best possible results. In this paper we introduce the concept of a workflow for an optimization of queries in the medical and biological sector by using a series of tools for expansion and transformation of the query. After the definition of attributes by the user, the query string is compared to previous queries in order to add semantic co-occurring terms to the query. Additionally, the query is enlarged by an inclusion of synonyms. The translation into database specific ontologies ensures the optimal query formulation for the chosen database(s). As this process can be performed in various databases at once, the results are ranked and normalized in order to achieve a comparable list of answers for a question.
Massive Scale Cyber Traffic Analysis: A Driver for Graph Database Research
DOE Office of Scientific and Technical Information (OSTI.GOV)
Joslyn, Cliff A.; Choudhury, S.; Haglin, David J.
2013-06-19
We describe the significance and prominence of network traffic analysis (TA) as a graph- and network-theoretical domain for advancing research in graph database systems. TA involves observing and analyzing the connections between clients, servers, hosts, and actors within IP networks, both at particular times and as extended over times. Towards that end, NetFlow (or more generically, IPFLOW) data are available from routers and servers which summarize coherent groups of IP packets flowing through the network. IPFLOW databases are routinely interrogated statistically and visualized for suspicious patterns. But the ability to cast IPFLOW data as a massive graph and query itmore » interactively, in order to e.g.\\ identify connectivity patterns, is less well advanced, due to a number of factors including scaling, and their hybrid nature combining graph connectivity and quantitative attributes. In this paper, we outline requirements and opportunities for graph-structured IPFLOW analytics based on our experience with real IPFLOW databases. Specifically, we describe real use cases from the security domain, cast them as graph patterns, show how to express them in two graph-oriented query languages SPARQL and Datalog, and use these examples to motivate a new class of "hybrid" graph-relational systems.« less
CellLineNavigator: a workbench for cancer cell line analysis
Krupp, Markus; Itzel, Timo; Maass, Thorsten; Hildebrandt, Andreas; Galle, Peter R.; Teufel, Andreas
2013-01-01
The CellLineNavigator database, freely available at http://www.medicalgenomics.org/celllinenavigator, is a web-based workbench for large scale comparisons of a large collection of diverse cell lines. It aims to support experimental design in the fields of genomics, systems biology and translational biomedical research. Currently, this compendium holds genome wide expression profiles of 317 different cancer cell lines, categorized into 57 different pathological states and 28 individual tissues. To enlarge the scope of CellLineNavigator, the database was furthermore closely linked to commonly used bioinformatics databases and knowledge repositories. To ensure easy data access and search ability, a simple data and an intuitive querying interface were implemented. It allows the user to explore and filter gene expression, focusing on pathological or physiological conditions. For a more complex search, the advanced query interface may be used to query for (i) differentially expressed genes; (ii) pathological or physiological conditions; or (iii) gene names or functional attributes, such as Kyoto Encyclopaedia of Genes and Genomes pathway maps. These queries may also be combined. Finally, CellLineNavigator allows additional advanced analysis of differentially regulated genes by a direct link to the Database for Annotation, Visualization and Integrated Discovery (DAVID) Bioinformatics Resources. PMID:23118487
Using Assessment to Develop Social Responsibility as a Graduate Attribute in Teacher Education
ERIC Educational Resources Information Center
Howells, Kerry; Fitzallen, Noleine; Adams, Christine
2016-01-01
Australian higher education institutions have struggled to develop clear strategies for developing and assessing graduate attributes within their specific disciplinary contexts. Using the example of the graduate attribute of social responsibility, this paper explores the outcomes of using assessment tasks to raise the awareness of development of…
Amand, Caroline; Tong, Sabine; Kieffer, Alexia; Kyaw, Moe H
2018-04-20
Despite several studies that have estimated the economic impact of Respiratory Syncytial Virus (RSV) in infants, limited data are available on healthcare resource use and costs attributable to RSV across age groups. The aim of this study was to quantify age-specific RSV-related healthcare resource use and costs on the US healthcare system. This retrospective case-control study identified patients aged ≥1 year with an RSV event in the Truven Health Marketscan® Commercial Claims and Encounters and Medicare Supplemental and Coordination of Benefits databases between August 31, 2012 and August 1, 2013. RSV patients were matched 1:1 with non-RSV controls for age, gender, region, healthcare plan and index date (n = 11,432 in each group). Stratified analyses for healthcare resource use and costs were conducted by age groups. RSV-attributable resource use and costs were assessed based on the incremental differences between RSV cases and controls using multivariate analysis. RSV patients had a higher healthcare resource use (hospital stays, emergency room/urgent care visits, ambulatory visits and outpatient visits) than non-RSV matched controls for all age groups (all p < 0.0001), particularly in the elderly age groups with RSV (1.9 to 3 days length of stay, 0.4 to 0.5 more ER/UC visits, 0.7 to 2.7 more ambulatory visits, 12.1 to 18.6 more outpatient visits and 9.5 to 14.6 more prescriptions than elderly in the control groups). The incremental difference in adjusted mean annual costs between RSV and non-RSV controls was higher in elderly (≥65; $12,030 to $23,194) than in those aged < 65 years ($2251 to $5391). Among children, adjusted costs attributable to RSV were higher in children aged 5-17 years ($3192), than those 1-4 years ($2251 to $2521). Our findings showed a substantial annual RSV-attributable healthcare resource use and costs in the US across age groups, with the highest burden in those aged ≥65 years. These data can be used in cost-effectiveness analyses, and may be useful for policymakers to guide future RSV vaccination and other prevention programs.
Zhang, Dongmu; Prabhu, Vimalanand S; Marcella, Stephen W
2018-04-17
The economic burden of Clostridium difficile infection (CDI), the leading cause of nosocomial infectious diarrhea, is not well understood. The objective of this study was to estimate the healthcare resource utilization (HCRU) and costs attributable to primary CDI and recurrent CDI (rCDI). This is a database (MarketScan) study. Patients without CDI were matched 1:1 by propensity score to those with primary CDI but no recurrences to obtain HCRU and costs attributable to primary CDI. Patients with primary CDI but no recurrences were matched 1:1 by propensity score to those with primary CDI plus 1 recurrence in order to obtain HCRU and costs attributable to rCDI. Adjusted estimates for incremental cumulative hospitalized days and healthcare costs over a 6-month follow-up period were obtained by generalized linear models with a Poisson or gamma distribution and a log link. Bootstrapping was used to obtain 95% confidence intervals (CIs). A total of 55504 eligible CDI patients were identified. Approximately 25% of these CDI patients had rCDI. The cumulative hospitalized days attributable to primary CDI and rCDI over the 6-month follow-up period were 5.20 days (95% CI, 5.01-5.39) and 1.95 days (95% CI, 1.48-2.43), respectively. The healthcare costs attributable to primary CDI and rCDI over the 6-month follow-up period were $24205 (95% CI, $23436-$25013) and $10580 (95% CI, $8849-$12446), respectively. The HCRU and costs attributable to primary CDI and rCDI are quite substantial. It is necessary to reduce the burden of CDI, especially rCDI.
Zhang, Dongmu; Prabhu, Vimalanand S; Marcella, Stephen W
2018-01-01
Abstract Background The economic burden of Clostridium difficile infection (CDI), the leading cause of nosocomial infectious diarrhea, is not well understood. The objective of this study was to estimate the healthcare resource utilization (HCRU) and costs attributable to primary CDI and recurrent CDI (rCDI). Methods This is a database (MarketScan) study. Patients without CDI were matched 1:1 by propensity score to those with primary CDI but no recurrences to obtain HCRU and costs attributable to primary CDI. Patients with primary CDI but no recurrences were matched 1:1 by propensity score to those with primary CDI plus 1 recurrence in order to obtain HCRU and costs attributable to rCDI. Adjusted estimates for incremental cumulative hospitalized days and healthcare costs over a 6-month follow-up period were obtained by generalized linear models with a Poisson or gamma distribution and a log link. Bootstrapping was used to obtain 95% confidence intervals (CIs). Results A total of 55504 eligible CDI patients were identified. Approximately 25% of these CDI patients had rCDI. The cumulative hospitalized days attributable to primary CDI and rCDI over the 6-month follow-up period were 5.20 days (95% CI, 5.01–5.39) and 1.95 days (95% CI, 1.48–2.43), respectively. The healthcare costs attributable to primary CDI and rCDI over the 6-month follow-up period were $24205 (95% CI, $23436–$25013) and $10580 (95% CI, $8849–$12446), respectively. Conclusions The HCRU and costs attributable to primary CDI and rCDI are quite substantial. It is necessary to reduce the burden of CDI, especially rCDI. PMID:29360950
Development of expert systems for analyzing electronic documents
NASA Astrophysics Data System (ADS)
Abeer Yassin, Al-Azzawi; Shidlovskiy, S.; Jamal, A. A.
2018-05-01
The paper analyses a Database Management System (DBMS). Expert systems, Databases, and database technology have become an essential component of everyday life in the modern society. As databases are widely used in every organization with a computer system, data resource control and data management are very important [1]. DBMS is the most significant tool developed to serve multiple users in a database environment consisting of programs that enable users to create and maintain a database. This paper focuses on development of a database management system for General Directorate for education of Diyala in Iraq (GDED) using Clips, java Net-beans and Alfresco and system components, which were previously developed in Tomsk State University at the Faculty of Innovative Technology.
IMPROVED SEARCH OF PRINCIPAL COMPONENT ANALYSIS DATABASES FOR SPECTRO-POLARIMETRIC INVERSION
DOE Office of Scientific and Technical Information (OSTI.GOV)
Casini, R.; Lites, B. W.; Ramos, A. Asensio
2013-08-20
We describe a simple technique for the acceleration of spectro-polarimetric inversions based on principal component analysis (PCA) of Stokes profiles. This technique involves the indexing of the database models based on the sign of the projections (PCA coefficients) of the first few relevant orders of principal components of the four Stokes parameters. In this way, each model in the database can be attributed a distinctive binary number of 2{sup 4n} bits, where n is the number of PCA orders used for the indexing. Each of these binary numbers (indices) identifies a group of ''compatible'' models for the inversion of amore » given set of observed Stokes profiles sharing the same index. The complete set of the binary numbers so constructed evidently determines a partition of the database. The search of the database for the PCA inversion of spectro-polarimetric data can profit greatly from this indexing. In practical cases it becomes possible to approach the ideal acceleration factor of 2{sup 4n} as compared to the systematic search of a non-indexed database for a traditional PCA inversion. This indexing method relies on the existence of a physical meaning in the sign of the PCA coefficients of a model. For this reason, the presence of model ambiguities and of spectro-polarimetric noise in the observations limits in practice the number n of relevant PCA orders that can be used for the indexing.« less
GIS-project: geodynamic globe for global monitoring of geological processes
NASA Astrophysics Data System (ADS)
Ryakhovsky, V.; Rundquist, D.; Gatinsky, Yu.; Chesalova, E.
2003-04-01
A multilayer geodynamic globe at the scale 1:10,000,000 was created at the end of the nineties in the GIS Center of the Vernadsky Museum. A special soft-and-hardware complex was elaborated for its visualization with a set of multitarget object directed databases. The globe includes separate thematic covers represented by digital sets of spatial geological, geochemical, and geophysical information (maps, schemes, profiles, stratigraphic columns, arranged databases etc.). At present the largest databases included in the globe program are connected with petrochemical and isotopic data on magmatic rocks of the World Ocean and with the large and supperlarge mineral deposits. Software by the Environmental Scientific Research Institute (ESRI), USA as well as ArcScan vectrorizator were used for covers digitizing and database adaptation (ARC/INFO 7.0, 8.0). All layers of the geoinformational project were obtained by scanning of separate objects and their transfer to the real geographic co-ordinates of an equiintermediate conic projection. Then the covers were projected on plane degree-system geographic co-ordinates. Some attributive databases were formed for each thematic layer, and in the last stage all covers were combined into the single information system. Separate digital covers represent mathematical descriptions of geological objects and relations between them, such as Earth's altimetry, active fault systems, seismicity etc. Some grounds of the cartographic generalization were taken into consideration in time of covers compilation with projection and co-ordinate systems precisely answered a given scale. The globe allows us to carry out in the interactive regime the formation of coordinated with each other object-oriented databases and thematic covers directly connected with them. They can be spread for all the Earth and the near-Earth space, and for the most well known parts of divergent and convergent boundaries of the lithosphere plates. Such covers and time series reflect in diagram form a total combination and dynamics of data on the geological structure, geophysical fields, seismicity, geomagnetism, composition of rock complexes, and metalloge-ny of different areas on the Earth's surface. They give us possibility to scale, detail, and develop 3D spatial visualization. Information filling the covers could be replenished as in the existing so in newly formed databases with new data. The integrated analyses of the data allows us more precisely to define our ideas on regularities in development of lithosphere and mantle unhomogeneities using some original technologies. It also enables us to work out 3D digital models for geodynamic development of tectonic zones in convergent and divergent plate boundaries with the purpose of integrated monitoring of mineral resources and establishing correlation between seismicity, magmatic activity, and metallogeny in time-spatial co-ordinates. The created multifold geoinformation system gives a chance to execute an integral analyses of geoinformation flows in the interactive regime and, in particular, to establish some regularities in the time-spatial distribution and dynamics of main structural units in the lithosphere, as well as illuminate the connection between stages of their development and epochs of large and supperlarge mineral deposit formation. Now we try to use the system for prediction of large oil and gas concentration in the main sedimentary basins. The work was supported by RFBR, (grants 93-07-14680, 96-07-89499, 99-07-90030, 00-15-98535, 02-07-90140) and MTC.
Drozda, Joseph P; Roach, James; Forsyth, Thomas; Helmering, Paul; Dummitt, Benjamin; Tcheng, James E
2018-02-01
The US Food and Drug Administration (FDA) has recognized the need to improve the tracking of medical device safety and performance, with implementation of Unique Device Identifiers (UDIs) in electronic health information as a key strategy. The FDA funded a demonstration by Mercy Health wherein prototype UDIs were incorporated into its electronic information systems. This report describes the demonstration's informatics architecture. Prototype UDIs for coronary stents were created and implemented across a series of information systems, resulting in UDI-associated data flow from manufacture through point of use to long-term follow-up, with barcode scanning linking clinical data with UDI-associated device attributes. A reference database containing device attributes and the UDI Research and Surveillance Database (UDIR) containing the linked clinical and device information were created, enabling longitudinal assessment of device performance. The demonstration included many stakeholders: multiple Mercy departments, manufacturers, health system partners, the FDA, professional societies, the National Cardiovascular Data Registry, and information system vendors. The resulting system of systems is described in detail, including entities, functions, linkage between the UDIR and proprietary systems using UDIs as the index key, data flow, roles and responsibilities of actors, and the UDIR data model. The demonstration provided proof of concept that UDIs can be incorporated into provider and enterprise electronic information systems and used as the index key to combine device and clinical data in a database useful for device evaluation. Keys to success and challenges to achieving this goal were identified. Fundamental informatics principles were central to accomplishing the system of systems model. © The Author 2017. Published by Oxford University Press on behalf of the American Medical Informatics Association. All rights reserved. For Permissions, please email: journals.permissions@oup.com
[A study of relation between hopelessness and causal attribution in school-aged children].
Sakurai, S
1989-12-01
This study was conducted to investigate the relation between hopelessness and causal attribution in Japanese school-aged children. In Study 1, the Japanese edition of hopelessness scale for children developed by Kazdin, French, Unis, Esveldt-Dawsan, and Sherick (1983) was constructed. Seventeen original items were translated into Japanese and they were administrated to 405 fifth- and sixth-graders. All of the items could be included to the Japanese edition of hopelessness scale. The reliability and validity was examined. In Study 2, the relation between hopelessness and causal attribution in children were investigated. The causal attribution questionnaire developed by Higuchi, Kambare, and Otsuka (1983) and the hopelessness scale developed by Study 1 were administered to 188 sixth-graders. Children with high scores in hopelessness scale significantly attributed negative events to much more effort factor than children with low scores. It supports neither the reformulated learned helplessness model nor the causal attribution theory of achievement motivation. It was explained mainly from points of self-serving attribution, cultural difference, and social desirability. Some questions were discussed for developing studies on depression and causal attribution in Japan.
Issues central to a useful image understanding environment
NASA Astrophysics Data System (ADS)
Beveridge, J. Ross; Draper, Bruce A.; Hanson, Allen R.; Riseman, Edward M.
1992-04-01
A recent DARPA initiative has sparked interested in software environments for computer vision. The goal is a single environment to support both basic research and technology transfer. This paper lays out six fundamental attributes such a system must possess: (1) support for both C and Lisp, (2) extensibility, (3) data sharing, (4) data query facilities tailored to vision, (5) graphics, and (6) code sharing. The first three attributes fundamentally constrain the system design. Support for both C and Lisp demands some form of database or data-store for passing data between languages. Extensibility demands that system support facilities, such as spatial retrieval of data, be readily extended to new user-defined datatypes. Finally, data sharing demands that data saved by one user, including data of a user-defined type, must be readable by another user.
Lubinski, David
2010-04-01
Invited commentary on Armstrong and Vogel's (2009) article on interpreting the interest-efficacy association stimulated an appraisal from a broader perspective. Like empirical research, scale development, and theorizing emanating from social cognitive career theory (SCCT), their conclusion about the importance of assessing both interests and self-efficacy in applied settings and speculations about the developmental sequencing of these attributes need to be evaluated in the context of what decades of longitudinal research reveal are critical determinants of educational and vocational choice, performance after choice, and persistence. For our interventions to be effective and our theory development to be meaningful, we must ensure that innovative measures possess incremental validity relative to cognitive abilities and educational-vocational interests, which are already well established as salient predictors of long-term educational-vocational outcomes. Broader historical, philosophical, and scientific perspectives are provided to enhance practice, research, and theory development. These broader perspectives reveal how well-positioned vocational counseling is for further advances if it builds on (rather than neglects) its longstanding tradition of developing a cumulative psychological science. PsycINFO Database Record (c) 2010 APA, all rights reserved.
RNA-Seq and molecular docking reveal multi-level pesticide resistance in the bed bug
2012-01-01
Background Bed bugs (Cimex lectularius) are hematophagous nocturnal parasites of humans that have attained high impact status due to their worldwide resurgence. The sudden and rampant resurgence of C. lectularius has been attributed to numerous factors including frequent international travel, narrower pest management practices, and insecticide resistance. Results We performed a next-generation RNA sequencing (RNA-Seq) experiment to find differentially expressed genes between pesticide-resistant (PR) and pesticide-susceptible (PS) strains of C. lectularius. A reference transcriptome database of 51,492 expressed sequence tags (ESTs) was created by combining the databases derived from de novo assembled mRNA-Seq tags (30,404 ESTs) and our previous 454 pyrosequenced database (21,088 ESTs). The two-way GLMseq analysis revealed ~15,000 highly significant differentially expressed ESTs between the PR and PS strains. Among the top 5,000 differentially expressed ESTs, 109 putative defense genes (cuticular proteins, cytochrome P450s, antioxidant genes, ABC transporters, glutathione S-transferases, carboxylesterases and acetyl cholinesterase) involved in penetration resistance and metabolic resistance were identified. Tissue and development-specific expression of P450 CYP3 clan members showed high mRNA levels in the cuticle, Malpighian tubules, and midgut; and in early instar nymphs, respectively. Lastly, molecular modeling and docking of a candidate cytochrome P450 (CYP397A1V2) revealed the flexibility of the deduced protein to metabolize a broad range of insecticide substrates including DDT, deltamethrin, permethrin, and imidacloprid. Conclusions We developed significant molecular resources for C. lectularius putatively involved in metabolic resistance as well as those participating in other modes of insecticide resistance. RNA-Seq profiles of PR strains combined with tissue-specific profiles and molecular docking revealed multi-level insecticide resistance in C. lectularius. Future research that is targeted towards RNA interference (RNAi) on the identified metabolic targets such as cytochrome P450s and cuticular proteins could lay the foundation for a better understanding of the genetic basis of insecticide resistance in C. lectularius. PMID:22226239