NASA Astrophysics Data System (ADS)
Hu, Haibin
2017-05-01
Among numerous WEB security issues, SQL injection is the most notable and dangerous. In this study, characteristics and procedures of SQL injection are analyzed, and the method for detecting the SQL injection attack is illustrated. The defense resistance and remedy model of SQL injection attack is established from the perspective of non-intrusive SQL injection attack and defense. Moreover, the ability of resisting the SQL injection attack of the server has been comprehensively improved through the security strategies on operation system, IIS and database, etc.. Corresponding codes are realized. The method is well applied in the actual projects.
Ultrabroadband photonic internet: safety aspects
NASA Astrophysics Data System (ADS)
Kalicki, Arkadiusz; Romaniuk, Ryszard
2008-11-01
Web applications became most popular medium in the Internet. Popularity, easiness of web application frameworks together with careless development results in high number of vulnerabilities and attacks. There are several types of attacks possible because of improper input validation. SQL injection is ability to execute arbitrary SQL queries in a database through an existing application. Cross-site scripting is the vulnerability which allows malicious web users to inject code into the web pages viewed by other users. Cross-Site Request Forgery (CSRF) is an attack that tricks the victim into loading a page that contains malicious request. Web spam in blogs. There are several techniques to mitigate attacks. Most important are web application strong design, correct input validation, defined data types for each field and parameterized statements in SQL queries. Server hardening with firewall, modern security policies systems and safe web framework interpreter configuration are essential. It is advised to keep proper security level on client side, keep updated software and install personal web firewalls or IDS/IPS systems. Good habits are logging out from services just after finishing work and using even separate web browser for most important sites, like e-banking.
Automatically Preparing Safe SQL Queries
NASA Astrophysics Data System (ADS)
Bisht, Prithvi; Sistla, A. Prasad; Venkatakrishnan, V. N.
We present the first sound program source transformation approach for automatically transforming the code of a legacy web application to employ PREPARE statements in place of unsafe SQL queries. Our approach therefore opens the way for eradicating the SQL injection threat vector from legacy web applications.
Protecting Database Centric Web Services against SQL/XPath Injection Attacks
NASA Astrophysics Data System (ADS)
Laranjeiro, Nuno; Vieira, Marco; Madeira, Henrique
Web services represent a powerful interface for back-end database systems and are increasingly being used in business critical applications. However, field studies show that a large number of web services are deployed with security flaws (e.g., having SQL Injection vulnerabilities). Although several techniques for the identification of security vulnerabilities have been proposed, developing non-vulnerable web services is still a difficult task. In fact, security-related concerns are hard to apply as they involve adding complexity to already complex code. This paper proposes an approach to secure web services against SQL and XPath Injection attacks, by transparently detecting and aborting service invocations that try to take advantage of potential vulnerabilities. Our mechanism was applied to secure several web services specified by the TPC-App benchmark, showing to be 100% effective in stopping attacks, non-intrusive and very easy to use.
Methods to Secure Databases Against Vulnerabilities
2015-12-01
for several languages such as C, C++, PHP, Java and Python [16]. MySQL will work well with very large databases. The documentation references...using Eclipse and connected to each database management system using Python and Java drivers provided by MySQL , MongoDB, and Datastax (for Cassandra...tiers in Python and Java . Problem MySQL MongoDB Cassandra 1. Injection a. Tautologies Vulnerable Vulnerable Not Vulnerable b. Illegal query
Fall 2014 Data-Intensive Systems
2014-10-29
Oct 2014 © 2014 Carnegie Mellon University Big Data Systems NoSQL and horizontal scaling are changing architecture principles by creating...University Status LEAP4BD • Ready to pilot QuABase • Prototype is complete – covers 8 NoSQL /NewSQL implementations • Completing validation testing Big...machine learning to automate population of knowledge base • Initial focus on NoSQL /NewSQL technology domain • Extend to create knowledge bases in other
Detecting Distributed SQL Injection Attacks in a Eucalyptus Cloud Environment
NASA Technical Reports Server (NTRS)
Kebert, Alan; Barnejee, Bikramjit; Solano, Juan; Solano, Wanda
2013-01-01
The cloud computing environment offers malicious users the ability to spawn multiple instances of cloud nodes that are similar to virtual machines, except that they can have separate external IP addresses. In this paper we demonstrate how this ability can be exploited by an attacker to distribute his/her attack, in particular SQL injection attacks, in such a way that an intrusion detection system (IDS) could fail to identify this attack. To demonstrate this, we set up a small private cloud, established a vulnerable website in one instance, and placed an IDS within the cloud to monitor the network traffic. We found that an attacker could quite easily defeat the IDS by periodically altering its IP address. To detect such an attacker, we propose to use multi-agent plan recognition, where the multiple source IPs are considered as different agents who are mounting a collaborative attack. We show that such a formulation of this problem yields a more sophisticated approach to detecting SQL injection attacks within a cloud computing environment.
SiC: An Agent Based Architecture for Preventing and Detecting Attacks to Ubiquitous Databases
NASA Astrophysics Data System (ADS)
Pinzón, Cristian; de Paz, Yanira; Bajo, Javier; Abraham, Ajith; Corchado, Juan M.
One of the main attacks to ubiquitous databases is the structure query language (SQL) injection attack, which causes severe damages both in the commercial aspect and in the user’s confidence. This chapter proposes the SiC architecture as a solution to the SQL injection attack problem. This is a hierarchical distributed multiagent architecture, which involves an entirely new approach with respect to existing architectures for the prevention and detection of SQL injections. SiC incorporates a kind of intelligent agent, which integrates a case-based reasoning system. This agent, which is the core of the architecture, allows the application of detection techniques based on anomalies as well as those based on patterns, providing a great degree of autonomy, flexibility, robustness and dynamic scalability. The characteristics of the multiagent system allow an architecture to detect attacks from different types of devices, regardless of the physical location. The architecture has been tested on a medical database, guaranteeing safe access from various devices such as PDAs and notebook computers.
Preventing SQL Code Injection by Combining Static and Runtime Analysis
2008-05-01
attacker changes the developer’s intended structure of an SQ L com- mand by inserting new SQ L keywords or operators. (Su and Wasser - mann provide a...FROM b o o k s WHERE a u t h o r = ’ ’ GROUP BY r a t i n g We use symbol as a placeholder for the indeterminate part of the command (in this...dialects of SQL.) In our model, we mark transitions that correspond to externally defined strings with the symbol . To illustrate, Figure 2 shows the SQL
Flexible network reconstruction from relational databases with Cytoscape and CytoSQL
2010-01-01
Background Molecular interaction networks can be efficiently studied using network visualization software such as Cytoscape. The relevant nodes, edges and their attributes can be imported in Cytoscape in various file formats, or directly from external databases through specialized third party plugins. However, molecular data are often stored in relational databases with their own specific structure, for which dedicated plugins do not exist. Therefore, a more generic solution is presented. Results A new Cytoscape plugin 'CytoSQL' is developed to connect Cytoscape to any relational database. It allows to launch SQL ('Structured Query Language') queries from within Cytoscape, with the option to inject node or edge features of an existing network as SQL arguments, and to convert the retrieved data to Cytoscape network components. Supported by a set of case studies we demonstrate the flexibility and the power of the CytoSQL plugin in converting specific data subsets into meaningful network representations. Conclusions CytoSQL offers a unified approach to let Cytoscape interact with relational databases. Thanks to the power of the SQL syntax, this tool can rapidly generate and enrich networks according to very complex criteria. The plugin is available at http://www.ptools.ua.ac.be/CytoSQL. PMID:20594316
Flexible network reconstruction from relational databases with Cytoscape and CytoSQL.
Laukens, Kris; Hollunder, Jens; Dang, Thanh Hai; De Jaeger, Geert; Kuiper, Martin; Witters, Erwin; Verschoren, Alain; Van Leemput, Koenraad
2010-07-01
Molecular interaction networks can be efficiently studied using network visualization software such as Cytoscape. The relevant nodes, edges and their attributes can be imported in Cytoscape in various file formats, or directly from external databases through specialized third party plugins. However, molecular data are often stored in relational databases with their own specific structure, for which dedicated plugins do not exist. Therefore, a more generic solution is presented. A new Cytoscape plugin 'CytoSQL' is developed to connect Cytoscape to any relational database. It allows to launch SQL ('Structured Query Language') queries from within Cytoscape, with the option to inject node or edge features of an existing network as SQL arguments, and to convert the retrieved data to Cytoscape network components. Supported by a set of case studies we demonstrate the flexibility and the power of the CytoSQL plugin in converting specific data subsets into meaningful network representations. CytoSQL offers a unified approach to let Cytoscape interact with relational databases. Thanks to the power of the SQL syntax, this tool can rapidly generate and enrich networks according to very complex criteria. The plugin is available at http://www.ptools.ua.ac.be/CytoSQL.
Vulnerabilities in Bytecode Removed by Analysis, Nuanced Confinement and Diversification (VIBRANCE)
2015-06-01
VIBRANCE tool starts with a vulnerable Java application and automatically hardens it against SQL injection, OS command injection, file path traversal...7 2.2 Java Front End...7 2.2.2 Java Byte Code Parser
Epstein, Richard H; Dexter, Franklin
2017-07-01
Comorbidity adjustment is often performed during outcomes and health care resource utilization research. Our goal was to develop an efficient algorithm in structured query language (SQL) to determine the Elixhauser comorbidity index. We wrote an SQL algorithm to calculate the Elixhauser comorbidities from Diagnosis Related Group and International Classification of Diseases (ICD) codes. Validation was by comparison to expected comorbidities from combinations of these codes and to the 2013 Nationwide Readmissions Database (NRD). The SQL algorithm matched perfectly with expected comorbidities for all combinations of ICD-9 or ICD-10, and Diagnosis Related Groups. Of 13 585 859 evaluable NRD records, the algorithm matched 100% of the listed comorbidities. Processing time was ∼0.05 ms/record. The SQL Elixhauser code was efficient and computationally identical to the SAS algorithm used for the NRD. This algorithm may be useful where preprocessing of large datasets in a relational database environment and comorbidity determination is desired before statistical analysis. A validated SQL procedure to calculate Elixhauser comorbidities and the van Walraven index from ICD-9 or ICD-10 discharge diagnosis codes has been published. © The Author 2017. Published by Oxford University Press on behalf of the American Medical Informatics Association. All rights reserved. For Permissions, please email: journals.permissions@oup.com
Information Flow Integrity for Systems of Independently-Developed Components
2015-06-22
We also examined three programs (Apache, MySQL , and PHP) in detail to evaluate the efficacy of using the provided package test suites to generate...method are just as effective as hooks that were manually placed over the course of years while greatly reducing the burden on programmers. ”Leveraging...to validate optimizations of real-world, mature applications: the Apache software suite, the Mozilla Suite, and the MySQL database. ”Validating Library
NASA Astrophysics Data System (ADS)
Ivankovic, D.; Dadic, V.
2009-04-01
Some of oceanographic parameters have to be manually inserted into database; some (for example data from CTD probe) are inserted from various files. All this parameters requires visualization, validation and manipulation from research vessel or scientific institution, and also public presentation. For these purposes is developed web based system, containing dynamic sql procedures and java applets. Technology background is Oracle 10g relational database, and Oracle application server. Web interfaces are developed using PL/SQL stored database procedures (mod PL/SQL). Additional parts for data visualization include use of Java applets and JavaScript. Mapping tool is Google maps API (javascript) and as alternative java applet. Graph is realized as dynamically generated web page containing java applet. Mapping tool and graph are georeferenced. That means that click on some part of graph, automatically initiate zoom or marker onto location where parameter was measured. This feature is very useful for data validation. Code for data manipulation and visualization are partially realized with dynamic SQL and that allow as to separate data definition and code for data manipulation. Adding new parameter in system requires only data definition and description without programming interface for this kind of data.
Operational Exercise Integration Recommendations for DoD Cyber Ranges
2015-08-05
be the precision and recall of a security information and event management (SIEM) system ’s notifications of unauthorized access to that directory...network traffic, port scanning Deplete Resources TCP flooding, memory leak exploitation Injection Cross-site scripting attacks, SQL injection Deceptive...requirements for personnel development; tactics, techniques, and procedures (TTPs) devel- opment; and mission rehearsals . While unique in their own
76 FR 26729 - Ceridian Corporation; Analysis of Proposed Consent Order to Aid Public Comment
Federal Register 2010, 2011, 2012, 2013, 2014
2011-05-09
... result of these failures, hackers executed an SQL injection attack on the Powerpay Web site and Web application. Through this attack, the hackers found personal information stored in Powerpay on Ceridian's...
Vulnerability Assessment of IPv6 Websites to SQL Injection and Other Application Level Attacks
Cho, Ying-Chiang; Pan, Jen-Yi
2013-01-01
Given the proliferation of internet connected devices, IPv6 has been proposed to replace IPv4. Aside from providing a larger address space which can be assigned to internet enabled devices, it has been suggested that the IPv6 protocol offers increased security due to the fact that with the large number of addresses available, standard IP scanning attacks will no longer become feasible. However, given the interest in attacking organizations rather than individual devices, most initial points of entry onto an organization's network and their attendant devices are visible and reachable through web crawling techniques, and, therefore, attacks on the visible application layer may offer ways to compromise the overall network. In this evaluation, we provide a straightforward implementation of a web crawler in conjunction with a benign black box penetration testing system and analyze the ease at which SQL injection attacks can be carried out. PMID:24574863
Vulnerability assessment of IPv6 websites to SQL injection and other application level attacks.
Cho, Ying-Chiang; Pan, Jen-Yi
2013-01-01
Given the proliferation of internet connected devices, IPv6 has been proposed to replace IPv4. Aside from providing a larger address space which can be assigned to internet enabled devices, it has been suggested that the IPv6 protocol offers increased security due to the fact that with the large number of addresses available, standard IP scanning attacks will no longer become feasible. However, given the interest in attacking organizations rather than individual devices, most initial points of entry onto an organization's network and their attendant devices are visible and reachable through web crawling techniques, and, therefore, attacks on the visible application layer may offer ways to compromise the overall network. In this evaluation, we provide a straightforward implementation of a web crawler in conjunction with a benign black box penetration testing system and analyze the ease at which SQL injection attacks can be carried out.
Multiple-Feature Extracting Modules Based Leak Mining System Design
Cho, Ying-Chiang; Pan, Jen-Yi
2013-01-01
Over the years, human dependence on the Internet has increased dramatically. A large amount of information is placed on the Internet and retrieved from it daily, which makes web security in terms of online information a major concern. In recent years, the most problematic issues in web security have been e-mail address leakage and SQL injection attacks. There are many possible causes of information leakage, such as inadequate precautions during the programming process, which lead to the leakage of e-mail addresses entered online or insufficient protection of database information, a loophole that enables malicious users to steal online content. In this paper, we implement a crawler mining system that is equipped with SQL injection vulnerability detection, by means of an algorithm developed for the web crawler. In addition, we analyze portal sites of the governments of various countries or regions in order to investigate the information leaking status of each site. Subsequently, we analyze the database structure and content of each site, using the data collected. Thus, we make use of practical verification in order to focus on information security and privacy through black-box testing. PMID:24453892
Multiple-feature extracting modules based leak mining system design.
Cho, Ying-Chiang; Pan, Jen-Yi
2013-01-01
Over the years, human dependence on the Internet has increased dramatically. A large amount of information is placed on the Internet and retrieved from it daily, which makes web security in terms of online information a major concern. In recent years, the most problematic issues in web security have been e-mail address leakage and SQL injection attacks. There are many possible causes of information leakage, such as inadequate precautions during the programming process, which lead to the leakage of e-mail addresses entered online or insufficient protection of database information, a loophole that enables malicious users to steal online content. In this paper, we implement a crawler mining system that is equipped with SQL injection vulnerability detection, by means of an algorithm developed for the web crawler. In addition, we analyze portal sites of the governments of various countries or regions in order to investigate the information leaking status of each site. Subsequently, we analyze the database structure and content of each site, using the data collected. Thus, we make use of practical verification in order to focus on information security and privacy through black-box testing.
PKI Layer Cake: New Collision Attacks against the Global X.509 Infrastructure
NASA Astrophysics Data System (ADS)
Kaminsky, Dan; Patterson, Meredith L.; Sassaman, Len
Research unveiled in December of 2008 [15] showed how MD5's long-known flaws could be actively exploited to attack the real-worldCertification Authority infrastructure. In this paper, we demonstrate two new classes of collision, which will be somewhat trickier to address than previous attacks against X.509: the applicability of MD2 preimage attacks against the primary root certificate for Verisign, and the difficulty of validating X.509 Names contained within PKCS#10 Certificate Requests.We also draw particular attention to two possibly unrecognized vectors for implementation flaws that have been problematic in the past: the ASN.1 BER decoder required to parsePKCS#10, and the potential for SQL injection fromtext contained within its requests. Finally, we explore why the implications of these attacks are broader than some have realized - first, because Client Authentication is sometimes tied to X.509, and second, because Extended Validation certificates were only intended to stop phishing attacks from names similar to trusted brands. As per the work of Adam Barth and Collin Jackson [4], EV does not prevent an attacker who can synthesize or acquire a "low assurance" certificate for a given name from acquiring the "green bar" EV experience.
Network traffic anomaly prediction using Artificial Neural Network
NASA Astrophysics Data System (ADS)
Ciptaningtyas, Hening Titi; Fatichah, Chastine; Sabila, Altea
2017-03-01
As the excessive increase of internet usage, the malicious software (malware) has also increase significantly. Malware is software developed by hacker for illegal purpose(s), such as stealing data and identity, causing computer damage, or denying service to other user[1]. Malware which attack computer or server often triggers network traffic anomaly phenomena. Based on Sophos's report[2], Indonesia is the riskiest country of malware attack and it also has high network traffic anomaly. This research uses Artificial Neural Network (ANN) to predict network traffic anomaly based on malware attack in Indonesia which is recorded by Id-SIRTII/CC (Indonesia Security Incident Response Team on Internet Infrastructure/Coordination Center). The case study is the highest malware attack (SQL injection) which has happened in three consecutive years: 2012, 2013, and 2014[4]. The data series is preprocessed first, then the network traffic anomaly is predicted using Artificial Neural Network and using two weight update algorithms: Gradient Descent and Momentum. Error of prediction is calculated using Mean Squared Error (MSE) [7]. The experimental result shows that MSE for SQL Injection is 0.03856. So, this approach can be used to predict network traffic anomaly.
Supporting temporal queries on clinical relational databases: the S-WATCH-QL language.
Combi, C.; Missora, L.; Pinciroli, F.
1996-01-01
Due to the ubiquitous and special nature of time, specially in clinical datábases there's the need of particular temporal data and operators. In this paper we describe S-WATCH-QL (Structured Watch Query Language), a temporal extension of SQL, the widespread query language based on the relational model. S-WATCH-QL extends the well-known SQL by the addition of: a) temporal data types that allow the storage of information with different levels of granularity; b) historical relations that can store together both instantaneous valid times and intervals; c) some temporal clauses, functions and predicates allowing to define complex temporal queries. PMID:8947722
Assembling proteomics data as a prerequisite for the analysis of large scale experiments
Schmidt, Frank; Schmid, Monika; Thiede, Bernd; Pleißner, Klaus-Peter; Böhme, Martina; Jungblut, Peter R
2009-01-01
Background Despite the complete determination of the genome sequence of a huge number of bacteria, their proteomes remain relatively poorly defined. Beside new methods to increase the number of identified proteins new database applications are necessary to store and present results of large- scale proteomics experiments. Results In the present study, a database concept has been developed to address these issues and to offer complete information via a web interface. In our concept, the Oracle based data repository system SQL-LIMS plays the central role in the proteomics workflow and was applied to the proteomes of Mycobacterium tuberculosis, Helicobacter pylori, Salmonella typhimurium and protein complexes such as 20S proteasome. Technical operations of our proteomics labs were used as the standard for SQL-LIMS template creation. By means of a Java based data parser, post-processed data of different approaches, such as LC/ESI-MS, MALDI-MS and 2-D gel electrophoresis (2-DE), were stored in SQL-LIMS. A minimum set of the proteomics data were transferred in our public 2D-PAGE database using a Java based interface (Data Transfer Tool) with the requirements of the PEDRo standardization. Furthermore, the stored proteomics data were extractable out of SQL-LIMS via XML. Conclusion The Oracle based data repository system SQL-LIMS played the central role in the proteomics workflow concept. Technical operations of our proteomics labs were used as standards for SQL-LIMS templates. Using a Java based parser, post-processed data of different approaches such as LC/ESI-MS, MALDI-MS and 1-DE and 2-DE were stored in SQL-LIMS. Thus, unique data formats of different instruments were unified and stored in SQL-LIMS tables. Moreover, a unique submission identifier allowed fast access to all experimental data. This was the main advantage compared to multi software solutions, especially if personnel fluctuations are high. Moreover, large scale and high-throughput experiments must be managed in a comprehensive repository system such as SQL-LIMS, to query results in a systematic manner. On the other hand, these database systems are expensive and require at least one full time administrator and specialized lab manager. Moreover, the high technical dynamics in proteomics may cause problems to adjust new data formats. To summarize, SQL-LIMS met the requirements of proteomics data handling especially in skilled processes such as gel-electrophoresis or mass spectrometry and fulfilled the PSI standardization criteria. The data transfer into a public domain via DTT facilitated validation of proteomics data. Additionally, evaluation of mass spectra by post-processing using MS-Screener improved the reliability of mass analysis and prevented storage of data junk. PMID:19166578
Kortüm, Karsten U; Müller, Michael; Kern, Christoph; Babenko, Alexander; Mayer, Wolfgang J; Kampik, Anselm; Kreutzer, Thomas C; Priglinger, Siegfried; Hirneiss, Christoph
2017-06-01
To develop a near-real-time data warehouse (DW) in an academic ophthalmologic center to gain scientific use of increasing digital data from electronic medical records (EMR) and diagnostic devices. Database development. Specific macular clinic user interfaces within the institutional hospital information system were created. Orders for imaging modalities were sent by an EMR-linked picture-archiving and communications system to the respective devices. All data of 325 767 patients since 2002 were gathered in a DW running on an SQL database. A data discovery tool was developed. An exemplary search for patients with age-related macular degeneration, performed cataract surgery, and at least 10 intravitreal (excluding bevacizumab) injections was conducted. Data related to those patients (3 142 204 diagnoses [including diagnoses from other fields of medicine], 720 721 procedures [eg, surgery], and 45 416 intravitreal injections) were stored, including 81 274 optical coherence tomography measurements. A web-based browsing tool was successfully developed for data visualization and filtering data by several linked criteria, for example, minimum number of intravitreal injections of a specific drug and visual acuity interval. The exemplary search identified 450 patients with 516 eyes meeting all criteria. A DW was successfully implemented in an ophthalmologic academic environment to support and facilitate research by using increasing EMR and measurement data. The identification of eligible patients for studies was simplified. In future, software for decision support can be developed based on the DW and its structured data. The improved classification of diseases and semiautomatic validation of data via machine learning are warranted. Copyright © 2017 Elsevier Inc. All rights reserved.
Improved Information Retrieval Performance on SQL Database Using Data Adapter
NASA Astrophysics Data System (ADS)
Husni, M.; Djanali, S.; Ciptaningtyas, H. T.; Wicaksana, I. G. N. A.
2018-02-01
The NoSQL databases, short for Not Only SQL, are increasingly being used as the number of big data applications increases. Most systems still use relational databases (RDBs), but as the number of data increases each year, the system handles big data with NoSQL databases to analyze and access data more quickly. NoSQL emerged as a result of the exponential growth of the internet and the development of web applications. The query syntax in the NoSQL database differs from the SQL database, therefore requiring code changes in the application. Data adapter allow applications to not change their SQL query syntax. Data adapters provide methods that can synchronize SQL databases with NotSQL databases. In addition, the data adapter provides an interface which is application can access to run SQL queries. Hence, this research applied data adapter system to synchronize data between MySQL database and Apache HBase using direct access query approach, where system allows application to accept query while synchronization process in progress. From the test performed using data adapter, the results obtained that the data adapter can synchronize between SQL databases, MySQL, and NoSQL database, Apache HBase. This system spends the percentage of memory resources in the range of 40% to 60%, and the percentage of processor moving from 10% to 90%. In addition, from this system also obtained the performance of database NoSQL better than SQL database.
The HARPS-N archive through a Cassandra, NoSQL database suite?
NASA Astrophysics Data System (ADS)
Molinari, Emilio; Guerra, Jose; Harutyunyan, Avet; Lodi, Marcello; Martin, Adrian
2016-07-01
The TNG-INAF is developing the science archive for the WEAVE instrument. The underlying architecture of the archive is based on a non relational database, more precisely, on Apache Cassandra cluster, which uses a NoSQL technology. In order to test and validate the use of this architecture, we created a local archive which we populated with all the HARPSN spectra collected at the TNG since the instrument's start of operations in mid-2012, as well as developed tools for the analysis of this data set. The HARPS-N data set is two orders of magnitude smaller than WEAVE, but we want to demonstrate the ability to walk through a complete data set and produce scientific output, as valuable as that produced by an ordinary pipeline, though without accessing directly the FITS files. The analytics is done by Apache Solr and Spark and on a relational PostgreSQL database. As an example, we produce observables like metallicity indexes for the targets in the archive and compare the results with the ones coming from the HARPS-N regular data reduction software. The aim of this experiment is to explore the viability of a high availability cluster and distributed NoSQL database as a platform for complex scientific analytics on a large data set, which will then be ported to the WEAVE Archive System (WAS) which we are developing for the WEAVE multi object, fiber spectrograph.
Offense-Defense Balance in Cyberspace: A Proposed Model
2012-12-01
enable faster multitasking , the more computations per second can be accomplished. Increased processing speed will allow an attacker to send attacks...depend on what spam or DDoS needs to be accomplished with prices ranging from $30-$70 a day for DDoS to $10 for 1 million spam emails . 21 For...In addition to DDoS attacks on webservers, there were web defacements (using various tools such as SQL injections) and email flooding. The attacks
2015-03-01
domains. Major model functions include: • Ground combat: Light and heavy forces. • Air mobile forces. • Future forces. • Fixed-wing and rotary-wing...Constraints: • Study must be completed no later than 31 December 2014. • Entity behavior limited to select COMBATXXI Mobility , Unmanned Aerial System...and SQL backend , as well as any open application programming interface API. • Allows data transparency and data driven navigation through the model
Research on high availability architecture of SQL and NoSQL
NASA Astrophysics Data System (ADS)
Wang, Zhiguo; Wei, Zhiqiang; Liu, Hao
2017-03-01
With the advent of the era of big data, amount and importance of data have increased dramatically. SQL database develops in performance and scalability, but more and more companies tend to use NoSQL database as their databases, because NoSQL database has simpler data model and stronger extension capacity than SQL database. Almost all database designers including SQL database and NoSQL database aim to improve performance and ensure availability by reasonable architecture which can reduce the effects of software failures and hardware failures, so that they can provide better experiences for their customers. In this paper, I mainly discuss the architectures of MySQL, MongoDB, and Redis, which are high available and have been deployed in practical application environment, and design a hybrid architecture.
Maintaining Multimedia Data in a Geospatial Database
2012-09-01
at PostgreSQL and MySQL as spatial databases was offered. Given their results, as each database produced result sets from zero to 100,000, it was...excelled given multiple conditions. A different look at PostgreSQL and MySQL as spatial databases was offered. Given their results, as each database... MySQL ................................................................................................14 B. BENCHMARKING DATA RETRIEVED FROM TABLE
Sánchez-de-Madariaga, Ricardo; Muñoz, Adolfo; Castro, Antonio L; Moreno, Oscar; Pascual, Mario
2018-01-01
This research shows a protocol to assess the computational complexity of querying relational and non-relational (NoSQL (not only Structured Query Language)) standardized electronic health record (EHR) medical information database systems (DBMS). It uses a set of three doubling-sized databases, i.e. databases storing 5000, 10,000 and 20,000 realistic standardized EHR extracts, in three different database management systems (DBMS): relational MySQL object-relational mapping (ORM), document-based NoSQL MongoDB, and native extensible markup language (XML) NoSQL eXist. The average response times to six complexity-increasing queries were computed, and the results showed a linear behavior in the NoSQL cases. In the NoSQL field, MongoDB presents a much flatter linear slope than eXist. NoSQL systems may also be more appropriate to maintain standardized medical information systems due to the special nature of the updating policies of medical information, which should not affect the consistency and efficiency of the data stored in NoSQL databases. One limitation of this protocol is the lack of direct results of improved relational systems such as archetype relational mapping (ARM) with the same data. However, the interpolation of doubling-size database results to those presented in the literature and other published results suggests that NoSQL systems might be more appropriate in many specific scenarios and problems to be solved. For example, NoSQL may be appropriate for document-based tasks such as EHR extracts used in clinical practice, or edition and visualization, or situations where the aim is not only to query medical information, but also to restore the EHR in exactly its original form. PMID:29608174
Sánchez-de-Madariaga, Ricardo; Muñoz, Adolfo; Castro, Antonio L; Moreno, Oscar; Pascual, Mario
2018-03-19
This research shows a protocol to assess the computational complexity of querying relational and non-relational (NoSQL (not only Structured Query Language)) standardized electronic health record (EHR) medical information database systems (DBMS). It uses a set of three doubling-sized databases, i.e. databases storing 5000, 10,000 and 20,000 realistic standardized EHR extracts, in three different database management systems (DBMS): relational MySQL object-relational mapping (ORM), document-based NoSQL MongoDB, and native extensible markup language (XML) NoSQL eXist. The average response times to six complexity-increasing queries were computed, and the results showed a linear behavior in the NoSQL cases. In the NoSQL field, MongoDB presents a much flatter linear slope than eXist. NoSQL systems may also be more appropriate to maintain standardized medical information systems due to the special nature of the updating policies of medical information, which should not affect the consistency and efficiency of the data stored in NoSQL databases. One limitation of this protocol is the lack of direct results of improved relational systems such as archetype relational mapping (ARM) with the same data. However, the interpolation of doubling-size database results to those presented in the literature and other published results suggests that NoSQL systems might be more appropriate in many specific scenarios and problems to be solved. For example, NoSQL may be appropriate for document-based tasks such as EHR extracts used in clinical practice, or edition and visualization, or situations where the aim is not only to query medical information, but also to restore the EHR in exactly its original form.
2009-01-01
Oracle 9i, 10g MySQL MS SQL Server MS SQL Server Operating System Supported Windows 2003 Server Windows 2000 Server (32 bit...WebStar (Mac OS X) SunOne Internet Information Services (IIS) Database Server Supported MS SQL Server MS SQL Server Oracle 9i, 10g...challenges of Web-based surveys are: 1) identifying the best Commercial Off the Shelf (COTS) Web-based survey packages to serve the particular
An Optimization of the Basic School Military Occupational Skill Assignment Process
2003-06-01
Corps Intranet (NMCI)23 supports it. We evaluated the use of Microsoft’s SQL Server, but dismissed this after learning that TBS did not possess a SQL ...Server license or a qualified SQL Server administrator.24 SQL Server would have provided for additional security measures not available in MS...administrator. Although not has powerful as SQL Server, MS Access can handle the multi-user environment necessary for this system.25 The training
Photo-z-SQL: Photometric redshift estimation framework
NASA Astrophysics Data System (ADS)
Beck, Róbert; Dobos, László; Budavári, Tamás; Szalay, Alexander S.; Csabai, István
2017-04-01
Photo-z-SQL is a flexible template-based photometric redshift estimation framework that can be seamlessly integrated into a SQL database (or DB) server and executed on demand in SQL. The DB integration eliminates the need to move large photometric datasets outside a database for redshift estimation, and uses the computational capabilities of DB hardware. Photo-z-SQL performs both maximum likelihood and Bayesian estimation and handles inputs of variable photometric filter sets and corresponding broad-band magnitudes.
Quantifying Uncertainty in Expert Judgment: Initial Results
2013-03-01
lines of source code were added in . ---------- C++ = 32%; JavaScript = 29%; XML = 15%; C = 7%; CSS = 7%; Java = 5%; Oth- er = 5% LOC = 927,266...much total effort in person years has been spent on this project? CMU/SEI-2013-TR-001 | 33 5 MySQL , the most popular Open Source SQL...as MySQL , Oracle, PostgreSQL, MS SQL Server, ODBC, or Interbase. Features include email reminders, iCal/vCal import/export, re- mote subscriptions
Sánchez-de-Madariaga, Ricardo; Muñoz, Adolfo; Lozano-Rubí, Raimundo; Serrano-Balazote, Pablo; Castro, Antonio L; Moreno, Oscar; Pascual, Mario
2017-08-18
The objective of this research is to compare the relational and non-relational (NoSQL) database systems approaches in order to store, recover, query and persist standardized medical information in the form of ISO/EN 13606 normalized Electronic Health Record XML extracts, both in isolation and concurrently. NoSQL database systems have recently attracted much attention, but few studies in the literature address their direct comparison with relational databases when applied to build the persistence layer of a standardized medical information system. One relational and two NoSQL databases (one document-based and one native XML database) of three different sizes have been created in order to evaluate and compare the response times (algorithmic complexity) of six different complexity growing queries, which have been performed on them. Similar appropriate results available in the literature have also been considered. Relational and non-relational NoSQL database systems show almost linear algorithmic complexity query execution. However, they show very different linear slopes, the former being much steeper than the two latter. Document-based NoSQL databases perform better in concurrency than in isolation, and also better than relational databases in concurrency. Non-relational NoSQL databases seem to be more appropriate than standard relational SQL databases when database size is extremely high (secondary use, research applications). Document-based NoSQL databases perform in general better than native XML NoSQL databases. EHR extracts visualization and edition are also document-based tasks more appropriate to NoSQL database systems. However, the appropriate database solution much depends on each particular situation and specific problem.
Comprehensive Routing Security Development and Deployment for the Internet
2015-02-01
feature enhancement and bug fixes. • MySQL : MySQL is a widely used and popular open source database package. It was chosen for database support in the...RPSTIR depends on several other open source packages. • MySQL : MySQL is used for the the local RPKI database cache. • OpenSSL: OpenSSL is used for...cryptographic libraries for X.509 certificates. • ODBC mySql Connector: ODBC (Open Database Connectivity) is a standard programming interface (API) for
TabSQL: a MySQL tool to facilitate mapping user data to public databases.
Xia, Xiao-Qin; McClelland, Michael; Wang, Yipeng
2010-06-23
With advances in high-throughput genomics and proteomics, it is challenging for biologists to deal with large data files and to map their data to annotations in public databases. We developed TabSQL, a MySQL-based application tool, for viewing, filtering and querying data files with large numbers of rows. TabSQL provides functions for downloading and installing table files from public databases including the Gene Ontology database (GO), the Ensembl databases, and genome databases from the UCSC genome bioinformatics site. Any other database that provides tab-delimited flat files can also be imported. The downloaded gene annotation tables can be queried together with users' data in TabSQL using either a graphic interface or command line. TabSQL allows queries across the user's data and public databases without programming. It is a convenient tool for biologists to annotate and enrich their data.
TabSQL: a MySQL tool to facilitate mapping user data to public databases
2010-01-01
Background With advances in high-throughput genomics and proteomics, it is challenging for biologists to deal with large data files and to map their data to annotations in public databases. Results We developed TabSQL, a MySQL-based application tool, for viewing, filtering and querying data files with large numbers of rows. TabSQL provides functions for downloading and installing table files from public databases including the Gene Ontology database (GO), the Ensembl databases, and genome databases from the UCSC genome bioinformatics site. Any other database that provides tab-delimited flat files can also be imported. The downloaded gene annotation tables can be queried together with users' data in TabSQL using either a graphic interface or command line. Conclusions TabSQL allows queries across the user's data and public databases without programming. It is a convenient tool for biologists to annotate and enrich their data. PMID:20573251
Migration from relational to NoSQL database
NASA Astrophysics Data System (ADS)
Ghotiya, Sunita; Mandal, Juhi; Kandasamy, Saravanakumar
2017-11-01
Data generated by various real time applications, social networking sites and sensor devices is of very huge amount and unstructured, which makes it difficult for Relational database management systems to handle the data. Data is very precious component of any application and needs to be analysed after arranging it in some structure. Relational databases are only able to deal with structured data, so there is need of NoSQL Database management System which can deal with semi -structured data also. Relational database provides the easiest way to manage the data but as the use of NoSQL is increasing it is becoming necessary to migrate the data from Relational to NoSQL databases. Various frameworks has been proposed previously which provides mechanisms for migration of data stored at warehouses in SQL, middle layer solutions which can provide facility of data to be stored in NoSQL databases to handle data which is not structured. This paper provides a literature review of some of the recent approaches proposed by various researchers to migrate data from relational to NoSQL databases. Some researchers proposed mechanisms for the co-existence of NoSQL and Relational databases together. This paper provides a summary of mechanisms which can be used for mapping data stored in Relational databases to NoSQL databases. Various techniques for data transformation and middle layer solutions are summarised in the paper.
2014-09-01
NoSQL Data Store Technologies John Klein, Software Engineering Institute Patrick Donohoe, Software Engineering Institute Neil Ernst...REPORT TYPE N/A 3. DATES COVERED 4. TITLE AND SUBTITLE NoSQL Data Store Technologies 5a. CONTRACT NUMBER 5b. GRANT NUMBER 5c. PROGRAM ELEMENT...distribute data 4. Data Replication – determines how a NoSQL database facilitates reliable, high performance data replication to build
NASA Technical Reports Server (NTRS)
Alfaro, Victor O.; Casey, Nancy J.
2005-01-01
SQL-RAMS (where "SQL" signifies Structured Query Language and "RAMS" signifies Rocketdyne Automated Management System) is a successor to the legacy version of RAMS -- a computer program used to manage all work, nonconformance, corrective action, and configuration management on rocket engines and ground support equipment at Stennis Space Center. The legacy version resided in the File-Maker Pro software system and was constructed in modules that could act as standalone programs. There was little or no integration among modules. Because of limitations on file-management capabilities in FileMaker Pro, and because of difficulty of integration of FileMaker Pro with other software systems for exchange of data using such industry standards as SQL, the legacy version of RAMS proved to be limited, and working to circumvent its limitations too time-consuming. In contrast, SQL-RAMS is an integrated SQL-server-based program that supports all data-exchange software industry standards. Whereas in the legacy version, it was necessary to access individual modules to gain insight into a particular workstatus document, SQL-RAMS provides access through a single-screen presentation of core modules. In addition, SQL-RAMS enables rapid and efficient filtering of displayed statuses by predefined categories and test numbers. SQL-RAMS is rich in functionality and encompasses significant improvements over the legacy system. It provides users the ability to perform many tasks, which in the past required administrator intervention. Additionally, many of the design limitations have been corrected, allowing for a robust application that is user centric.
NASA Technical Reports Server (NTRS)
Alfaro, Victor O.; Casey, Nancy J.
2005-01-01
SQL-RAMS (where "SQL" signifies Structured Query Language and "RAMS" signifies Rocketdyne Automated Management System) is a successor to the legacy version of RAMS a computer program used to manage all work, nonconformance, corrective action, and configuration management on rocket engines and ground support equipment at Stennis Space Center. The legacy version resided in the FileMaker Pro software system and was constructed in modules that could act as stand-alone programs. There was little or no integration among modules. Because of limitations on file-management capabilities in FileMaker Pro, and because of difficulty of integration of FileMaker Pro with other software systems for exchange of data using such industry standards as SQL, the legacy version of RAMS proved to be limited, and working to circumvent its limitations too time-consuming. In contrast, SQL-RAMS is an integrated SQL-server-based program that supports all data-exchange software industry standards. Whereas in the legacy version, it was necessary to access individual modules to gain insight to a particular work-status documents, SQL-RAMS provides access through a single-screen presentation of core modules. In addition, SQL-RAMS enable rapid and efficient filtering of displayed statuses by predefined categories and test numbers. SQL-RAMS is rich in functionality and encompasses significant improvements over the legacy system. It provides users the ability to perform many tasks which in the past required administrator intervention. Additionally many of the design limitations have been corrected allowing for a robust application that is user centric.
2004-03-01
with MySQL . This choice was made because MySQL is open source. Any significant database engine such as Oracle or MS- SQL or even MS Access can be used...10 Figure 6. The DoD vs . Commercial Life Cycle...necessarily be interested in SCADA network security 13. MySQL (Database server) – This station represents a typical data server for a web page
Analysis and Development of a Web-Enabled Planning and Scheduling Database Application
2013-09-01
establishes an entity—relationship diagram for the desired process, constructs an operable database using MySQL , and provides a web- enabled interface for...development, develop, design, process, re- engineering, reengineering, MySQL , structured query language, SQL, myPHPadmin. 15. NUMBER OF PAGES 107 16...relationship diagram for the desired process, constructs an operable database using MySQL , and provides a web-enabled interface for the population of
Petaminer: Using ROOT for efficient data storage in MySQL database
NASA Astrophysics Data System (ADS)
Cranshaw, J.; Malon, D.; Vaniachine, A.; Fine, V.; Lauret, J.; Hamill, P.
2010-04-01
High Energy and Nuclear Physics (HENP) experiments store Petabytes of event data and Terabytes of calibration data in ROOT files. The Petaminer project is developing a custom MySQL storage engine to enable the MySQL query processor to directly access experimental data stored in ROOT files. Our project is addressing the problem of efficient navigation to PetaBytes of HENP experimental data described with event-level TAG metadata, which is required by data intensive physics communities such as the LHC and RHIC experiments. Physicists need to be able to compose a metadata query and rapidly retrieve the set of matching events, where improved efficiency will facilitate the discovery process by permitting rapid iterations of data evaluation and retrieval. Our custom MySQL storage engine enables the MySQL query processor to directly access TAG data stored in ROOT TTrees. As ROOT TTrees are column-oriented, reading them directly provides improved performance over traditional row-oriented TAG databases. Leveraging the flexible and powerful SQL query language to access data stored in ROOT TTrees, the Petaminer approach enables rich MySQL index-building capabilities for further performance optimization.
Evaluation of relational and NoSQL database architectures to manage genomic annotations.
Schulz, Wade L; Nelson, Brent G; Felker, Donn K; Durant, Thomas J S; Torres, Richard
2016-12-01
While the adoption of next generation sequencing has rapidly expanded, the informatics infrastructure used to manage the data generated by this technology has not kept pace. Historically, relational databases have provided much of the framework for data storage and retrieval. Newer technologies based on NoSQL architectures may provide significant advantages in storage and query efficiency, thereby reducing the cost of data management. But their relative advantage when applied to biomedical data sets, such as genetic data, has not been characterized. To this end, we compared the storage, indexing, and query efficiency of a common relational database (MySQL), a document-oriented NoSQL database (MongoDB), and a relational database with NoSQL support (PostgreSQL). When used to store genomic annotations from the dbSNP database, we found the NoSQL architectures to outperform traditional, relational models for speed of data storage, indexing, and query retrieval in nearly every operation. These findings strongly support the use of novel database technologies to improve the efficiency of data management within the biological sciences. Copyright © 2016 Elsevier Inc. All rights reserved.
Knowledge Data Base for Amorphous Metals
2007-07-26
not programmatic, updates. Over 100 custom SQL statements that maintain the domain specific data are attached to the workflow entries in a generic...for the form by populating the SQL and run generation tables. Application data may be prepared in different ways for two steps that invoke the same form...run generation mode). There is a single table of SQL commands. Each record has a user-definable ID, the SQL code, and a comment. The run generation
NASA Astrophysics Data System (ADS)
Dziedzic, Adam; Mulawka, Jan
2014-11-01
NoSQL is a new approach to data storage and manipulation. The aim of this paper is to gain more insight into NoSQL databases, as we are still in the early stages of understanding when to use them and how to use them in an appropriate way. In this submission descriptions of selected NoSQL databases are presented. Each of the databases is analysed with primary focus on its data model, data access, architecture and practical usage in real applications. Furthemore, the NoSQL databases are compared in fields of data references. The relational databases offer foreign keys, whereas NoSQL databases provide us with limited references. An intermediate model between graph theory and relational algebra which can address the problem should be created. Finally, the proposal of a new approach to the problem of inconsistent references in Big Data storage systems is introduced.
Validation of the United States Marine Corps Qualified Candidate Population Model
2003-03-01
time. Fields are created in the database to support this forecasting. User forms and a macro are programmed in Microsoft VBA to develop the...at 0.001. To accomplish 50,000 iterations of a minimization problem, this study wrote a macro in the VBA programming language that guides the solver...success in the commissioning process. **To improve the diagnostics of this propensity model, other factors were considered as well. Applying SQL
Risk Assessment of the Naval Postgraduate School Gigabit Network
2004-09-01
Management Server (1) • Ras Server (1) • Remedy Server (1) • Samba Server(2) • SQL Servers (3) • Web Servers (3) • WINS Server (1) • Library...Server Bob Sharp INCA Windows 2000 Advanced Server NPGS Landesk SQL 2000 Alan Pires eagle Microsoft Windows 2000 Advanced Server EWS NPGS Landesk...Advanced Server Special Projects NPGS SQL Alan Pires MC01BDB Microsoft Windows 2000 Advanced Server Special Projects NPGS SQL 2000 Alan Pires
Blind Seer: A Scalable Private DBMS
2014-05-01
searchable index terms per DB row, in time comparable to (insecure) MySQL (many practical queries can be privately executed with work 1.2-3 times slower...than MySQL , although some queries are costlier). We support a rich query set, including searching on arbitrary boolean formulas on keywords and ranges...index terms per DB row, in time comparable to (insecure) MySQL (many practical queries can be privately executed with work 1.2-3 times slower than MySQL
Performance Evaluation of NoSQL Databases: A Case Study
2015-02-01
a centralized relational database. The customer decided to consider NoSQL technologies for two specific uses, namely: the primary data store for...17 custom specific 6. FU NoSQL availab data mo arking of data g a specific wo sin benchmark f hmark for tran le workload de o publish meas their...The choice of a particular NoSQL database imposes a specific distributed software architecture and data model, and is a major determinant of the
DOMe: A deduplication optimization method for the NewSQL database backups
Wang, Longxiang; Zhu, Zhengdong; Zhang, Xingjun; Wang, Yinfeng
2017-01-01
Reducing duplicated data of database backups is an important application scenario for data deduplication technology. NewSQL is an emerging database system and is now being used more and more widely. NewSQL systems need to improve data reliability by periodically backing up in-memory data, resulting in a lot of duplicated data. The traditional deduplication method is not optimized for the NewSQL server system and cannot take full advantage of hardware resources to optimize deduplication performance. A recent research pointed out that the future NewSQL server will have thousands of CPU cores, large DRAM and huge NVRAM. Therefore, how to utilize these hardware resources to optimize the performance of data deduplication is an important issue. To solve this problem, we propose a deduplication optimization method (DOMe) for NewSQL system backup. To take advantage of the large number of CPU cores in the NewSQL server to optimize deduplication performance, DOMe parallelizes the deduplication method based on the fork-join framework. The fingerprint index, which is the key data structure in the deduplication process, is implemented as pure in-memory hash table, which makes full use of the large DRAM in NewSQL system, eliminating the performance bottleneck problem of fingerprint index existing in traditional deduplication method. The H-store is used as a typical NewSQL database system to implement DOMe method. DOMe is experimentally analyzed by two representative backup data. The experimental results show that: 1) DOMe can reduce the duplicated NewSQL backup data. 2) DOMe significantly improves deduplication performance by parallelizing CDC algorithms. In the case of the theoretical speedup ratio of the server is 20.8, the speedup ratio of DOMe can achieve up to 18; 3) DOMe improved the deduplication throughput by 1.5 times through the pure in-memory index optimization method. PMID:29049307
2012-09-01
relative performance of several conventional SQL and NoSQL databases with a set of one billion file block hashes. Digital Forensics, Sector Hashing, Full... NoSQL databases with a set of one billion file block hashes. v THIS PAGE INTENTIONALLY LEFT BLANK vi Table of Contents List of Acronyms and...Operating System NOOP No Operation assembly instruction NoSQL “Not only SQL” model for non-relational database management NSRL National Software
Global ISR: Toward a Comprehensive Defense Against Unauthorized Code Execution
2010-10-01
implementation using two of the most popular open- source servers: the Apache web server, and the MySQL database server. For Apache, we measure the effect that...utility ab. T o ta l T im e ( s e c ) 0 500 1000 1500 2000 2500 3000 Native Null ISR ISR−MP Fig. 3. The MySQL test-insert bench- mark measures...various SQL operations. The figure draws total execution time as reported by the benchmark utility. Finally, we benchmarked a MySQL database server using
SQL/NF Translator for the Triton Nested Relational Database System
1990-12-01
18as., Ohio .. 9~~ ~~ 1 4- AFIT/GCE/ENG/90D-05 SQL/Nk1 TRANSLATOR FOR THE TRITON NESTED RELATIONAL DATABASE SYSTEM THESIS Craig William Schnepf Captain...FOR THE TRITON NESTED RELATIONAL DATABASE SYSTEM THESIS Presented to the Faculty of the School of Engineering of the Air Force Institute of Technnlogy... systems . The SQL/NF query language used for the nested relationil model is an extension of the popular relational model query language SQL. The query
Flexible Decision Support in Device-Saturated Environments
2003-10-01
also output tuples to a remote MySQL or Postgres database. 3.3 GUI The GUI allows the user to pose queries using SQL and to display query...DatabaseConnection.java – handles connections to an external database (such as MySQL or Postgres ). • Debug.java – contains the code for printing out Debug messages...also provided. It is possible to output the results of queries to a MySQL or Postgres database for archival and the GUI can query those results
Evaluating a NoSQL Alternative for Chilean Virtual Observatory Services
NASA Astrophysics Data System (ADS)
Antognini, J.; Araya, M.; Solar, M.; Valenzuela, C.; Lira, F.
2015-09-01
Currently, the standards and protocols for data access in the Virtual Observatory architecture (DAL) are generally implemented with relational databases based on SQL. In particular, the Astronomical Data Query Language (ADQL), language used by IVOA to represent queries to VO services, was created to satisfy the different data access protocols, such as Simple Cone Search. ADQL is based in SQL92, and has extra functionality implemented using PgSphere. An emergent alternative to SQL are the so called NoSQL databases, which can be classified in several categories such as Column, Document, Key-Value, Graph, Object, etc.; each one recommended for different scenarios. Within their notable characteristics we can find: schema-free, easy replication support, simple API, Big Data, etc. The Chilean Virtual Observatory (ChiVO) is developing a functional prototype based on the IVOA architecture, with the following relevant factors: Performance, Scalability, Flexibility, Complexity, and Functionality. Currently, it's very difficult to compare these factors, due to a lack of alternatives. The objective of this paper is to compare NoSQL alternatives with SQL through the implementation of a Web API REST that satisfies ChiVO's needs: a SESAME-style name resolver for the data from ALMA. Therefore, we propose a test scenario by configuring a NoSQL database with data from different sources and evaluating the feasibility of creating a Simple Cone Search service and its performance. This comparison will allow to pave the way for the application of Big Data databases in the Virtual Observatory.
Urgent Virtual Machine Eviction with Enlightened Post-Copy
2015-12-01
memory is in use, almost all of which is by Memcached. MySQL : The VMs run MySQL 5.6, and the clients execute OLTPBenchmark [3] using the Twitter...workload with scale factor of 960. The VMs are each allocated 16 cores and 30 GB of memory, and MySQL is configured with a 16 GB buffer pool in memory. The...operation mix for 5 minutes as a warm-up. At the time of migration, MySQL uses approximately 17 GB of memory, and almost all of the 30 GB memory is
NoSQL: collection document and cloud by using a dynamic web query form
NASA Astrophysics Data System (ADS)
Abdalla, Hemn B.; Lin, Jinzhao; Li, Guoquan
2015-07-01
Mongo-DB (from "humongous") is an open-source document database and the leading NoSQL database. A NoSQL (Not Only SQL, next generation databases, being non-relational, deal, open-source and horizontally scalable) presenting a mechanism for storage and retrieval of documents. Previously, we stored and retrieved the data using the SQL queries. Here, we use the MonogoDB that means we are not utilizing the MySQL and SQL queries. Directly importing the documents into our Drives, retrieving the documents on that drive by not applying the SQL queries, using the IO BufferReader and Writer, BufferReader for importing our type of document files to my folder (Drive). For retrieving the document files, the usage is BufferWriter from the particular folder (or) Drive. In this sense, providing the security for those storing files for what purpose means if we store the documents in our local folder means all or views that file and modified that file. So preventing that file, we are furnishing the security. The original document files will be changed to another format like in this paper; Binary format is used. Our documents will be converting to the binary format after that direct storing in one of our folder, that time the storage space will provide the private key for accessing that file. Wherever any user tries to discover the Document files means that file data are in the binary format, the document's file owner simply views that original format using that personal key from receive the secret key from the cloud.
Recommender System for Learning SQL Using Hints
ERIC Educational Resources Information Center
Lavbic, Dejan; Matek, Tadej; Zrnec, Aljaž
2017-01-01
Today's software industry requires individuals who are proficient in as many programming languages as possible. Structured query language (SQL), as an adopted standard, is no exception, as it is the most widely used query language to retrieve and manipulate data. However, the process of learning SQL turns out to be challenging. The need for a…
Life in extra dimensions of database world or penetration of NoSQL in HEP community
NASA Astrophysics Data System (ADS)
Kuznetsov, V.; Evans, D.; Metson, S.
2012-12-01
The recent buzzword in IT world is NoSQL. Major players, such as Facebook, Yahoo, Google, etc. are widely adopted different “NoSQL” solutions for their needs. Horizontal scalability, flexible data model and management of big data volumes are only a few advantages of NoSQL. In CMS experiment we use several of them in production environment. Here, we present CMS projects based on NoSQL solutions, their strengths and weaknesses as well as our experience with those tools and their coexistence with standard RDBMS solutions in our applications.
A Database Design for the Brazilian Air Force Military Personnel Control System.
1987-06-01
GIVEN A RECNum GET MOVING HISTORICAL". 77 SEL4 PlC X(70) VALUE ". 4. GIVEN A RECNUM GET NOMINATION HISTORICAL". 77 SEL5 PIC X(70) VALUE it 5. GIVEN A...WHERE - "°RECNUM = :RECNUM". 77 SQL-SEL3-LENGTH PIC S9999 VALUE 150 COMP. 77 SQL- SEL4 PIC X(150) VALUE "SELECT ABBREV,DTNOM,DTEXO,SITN FROM...NOMINATION WHERE RECNUM 77 SQL- SEL4 -LENGTH PIC S9999 VALUE 150 COMP. 77 SQL-SEL5 PIC X(150) VALUE "SELECT ABBREVDTDES,DTWAIVER,SITD FROM DESIG WHERE RECNUM It
Sun, Yongmei; Li, Xing; Wu, Di; Pan, Qi; Ji, Yuefeng; Ren, Hong; Ding, Keyue
2016-01-01
RNA editing is one of the post- or co-transcriptional processes that can lead to amino acid substitutions in protein sequences, alternative pre-mRNA splicing, and changes in gene expression levels. Although several methods have been suggested to identify RNA editing sites, there remains challenges to be addressed in distinguishing true RNA editing sites from its counterparts on genome and technical artifacts. In addition, there lacks a software framework to identify and visualize potential RNA editing sites. Here, we presented a software - 'RED' (RNA Editing sites Detector) - for the identification of RNA editing sites by integrating multiple rule-based and statistical filters. The potential RNA editing sites can be visualized at the genome and the site levels by graphical user interface (GUI). To improve performance, we used MySQL database management system (DBMS) for high-throughput data storage and query. We demonstrated the validity and utility of RED by identifying the presence and absence of C→U RNA-editing sites experimentally validated, in comparison with REDItools, a command line tool to perform high-throughput investigation of RNA editing. In an analysis of a sample data-set with 28 experimentally validated C→U RNA editing sites, RED had sensitivity and specificity of 0.64 and 0.5. In comparison, REDItools had a better sensitivity (0.75) but similar specificity (0.5). RED is an easy-to-use, platform-independent Java-based software, and can be applied to RNA-seq data without or with DNA sequencing data. The package is freely available under the GPLv3 license at http://github.com/REDetector/RED or https://sourceforge.net/projects/redetector.
Sun, Yongmei; Li, Xing; Wu, Di; Pan, Qi; Ji, Yuefeng; Ren, Hong; Ding, Keyue
2016-01-01
RNA editing is one of the post- or co-transcriptional processes that can lead to amino acid substitutions in protein sequences, alternative pre-mRNA splicing, and changes in gene expression levels. Although several methods have been suggested to identify RNA editing sites, there remains challenges to be addressed in distinguishing true RNA editing sites from its counterparts on genome and technical artifacts. In addition, there lacks a software framework to identify and visualize potential RNA editing sites. Here, we presented a software − ‘RED’ (RNA Editing sites Detector) − for the identification of RNA editing sites by integrating multiple rule-based and statistical filters. The potential RNA editing sites can be visualized at the genome and the site levels by graphical user interface (GUI). To improve performance, we used MySQL database management system (DBMS) for high-throughput data storage and query. We demonstrated the validity and utility of RED by identifying the presence and absence of C→U RNA-editing sites experimentally validated, in comparison with REDItools, a command line tool to perform high-throughput investigation of RNA editing. In an analysis of a sample data-set with 28 experimentally validated C→U RNA editing sites, RED had sensitivity and specificity of 0.64 and 0.5. In comparison, REDItools had a better sensitivity (0.75) but similar specificity (0.5). RED is an easy-to-use, platform-independent Java-based software, and can be applied to RNA-seq data without or with DNA sequencing data. The package is freely available under the GPLv3 license at http://github.com/REDetector/RED or https://sourceforge.net/projects/redetector. PMID:26930599
Database Migration for Command and Control
2002-11-01
Sql - proprietary JDP Private Area Air defense data Defended asset list Oracle 7.3.2 - Automated process (OLTP...TADIL warnings Oracle 7.3.2 Flat File - Discrete transaction with data upds - NRT response required Pull mission data Std SQL ...level execution data Oracle 7.3 User update External interfaces Auto/manual backup Messaging Proprietary replication (internally) SQL Web server
Oracle Applications Patch Administration Tool (PAT) Beta Version
DOE Office of Scientific and Technical Information (OSTI.GOV)
2002-01-04
PAT is a Patch Administration Tool that provides analysis, tracking, and management of Oracle Application patches. This includes capabilities as outlined below: Patch Analysis & Management Tool Outline of capabilities: Administration Patch Data Maintenance -- track Oracle Application patches applied to what database instance & machine Patch Analysis capture text files (readme.txt and driver files) form comparison detail report comparison detail PL/SQL package comparison detail SQL scripts detail JSP module comparison detail Parse and load the current applptch.txt (10.7) or load patch data from Oracle Application database patch tables (11i) Display Analysis -- Compare patch to be applied with currentmore » Oracle Application installed Appl_top code versions Patch Detail Module comparison detail Analyze and display one Oracle Application module patch. Patch Management -- automatic queue and execution of patches Administration Parameter maintenance -- setting for directory structure of Oracle Application appl_top Validation data maintenance -- machine names and instances to patch Operation Patch Data Maintenance Schedule a patch (queue for later execution) Run a patch (queue for immediate execution) Review the patch logs Patch Management Reports« less
Network Configuration of Oracle and Database Programming Using SQL
NASA Technical Reports Server (NTRS)
Davis, Melton; Abdurrashid, Jibril; Diaz, Philip; Harris, W. C.
2000-01-01
A database can be defined as a collection of information organized in such a way that it can be retrieved and used. A database management system (DBMS) can further be defined as the tool that enables us to manage and interact with the database. The Oracle 8 Server is a state-of-the-art information management environment. It is a repository for very large amounts of data, and gives users rapid access to that data. The Oracle 8 Server allows for sharing of data between applications; the information is stored in one place and used by many systems. My research will focus primarily on SQL (Structured Query Language) programming. SQL is the way you define and manipulate data in Oracle's relational database. SQL is the industry standard adopted by all database vendors. When programming with SQL, you work on sets of data (i.e., information is not processed one record at a time).
Survey of Machine Learning Methods for Database Security
NASA Astrophysics Data System (ADS)
Kamra, Ashish; Ber, Elisa
Application of machine learning techniques to database security is an emerging area of research. In this chapter, we present a survey of various approaches that use machine learning/data mining techniques to enhance the traditional security mechanisms of databases. There are two key database security areas in which these techniques have found applications, namely, detection of SQL Injection attacks and anomaly detection for defending against insider threats. Apart from the research prototypes and tools, various third-party commercial products are also available that provide database activity monitoring solutions by profiling database users and applications. We present a survey of such products. We end the chapter with a primer on mechanisms for responding to database anomalies.
Teaching Case: Introduction to NoSQL in a Traditional Database Course
ERIC Educational Resources Information Center
Fowler, Brad; Godin, Joy; Geddy, Margaret
2016-01-01
Many organizations are dealing with the increasing demands of big data, so they are turning to NoSQL databases as their preferred system for handling the unique problems of capturing and storing massive amounts of data. Therefore, it is likely that employees in all sizes of organizations will encounter NoSQL databases. Thus, to be more job-ready,…
Zero tolerance for incorrect data: Best practices in SQL transaction programming
NASA Astrophysics Data System (ADS)
Laiho, M.; Skourlas, C.; Dervos, D. A.
2015-02-01
DBMS products differ in the way they support even the basic SQL transaction services. In this paper, a framework of best practices in SQL transaction programming is given and discussed. The SQL developers are advised to experiment with and verify the services supported by the DBMS product used. The framework has been developed by DBTechNet, a European network of teachers, trainers and ICT professionals. A course module on SQL transactions, offered by the LLP "DBTech VET Teachers" programme, is also presented and discussed. Aims and objectives of the programme include the introduction of the topics and content of SQL transactions and concurrency control to HE/VET curricula and addressing the need for initial and continuous training on these topics to in-company trainers, VET teachers, and Higher Education students. An overview of the course module, its learning outcomes, the education and training (E&T) content, virtual database labs with hands-on self-practicing exercises, plus instructions for the teacher/trainer on the pedagogy and the usage of the course modules' content are briefly described. The main principle adopted is to "Learn by verifying in practice" and the transactions course motto is: "Zero Tolerance for Incorrect Data".
The BiolAD-DB system : an informatics system for clinical and genetic data.
Nielsen, David A; Leidner, Marty; Haynes, Chad; Krauthammer, Michael; Kreek, Mary Jeanne
2007-01-01
The Biology of Addictive Diseases-Database (BiolAD-DB) system is a research bioinformatics system for archiving, analyzing, and processing of complex clinical and genetic data. The database schema employs design principles for handling complex clinical information, such as response items in genetic questionnaires. Data access and validation is provided by the BiolAD-DB client application, which features a data validation engine tightly coupled to a graphical user interface. Data integrity is provided by the password-protected BiolAD-DB SQL compliant server and database. BiolAD-DB tools further provide functionalities for generating customized reports and views. The BiolAD-DB system schema, client, and installation instructions are freely available at http://www.rockefeller.edu/biolad-db/.
DESPIC: Detecting Early Signatures of Persuasion in Information Cascades
2015-08-27
over NoSQL Databases, Proceedings of the 14th IEEE/ACM International Symposium on Cluster, Cloud and Grid Computing (CCGrid 2014). 26-MAY-14, . : , P...over NoSQL Databases. Proceedings of the 14th IEEE/ACM International Symposium on Cluster, Cloud and Grid Computing (CCGrid 2014). Chicago, IL, USA...distributed NoSQL databases including HBase and Riak, we finalized the requirements of the optimal computational architecture to support our framework
1990-06-30
declaring all of the valid enumeration values for a domain of type enumeration. Example-> In type color_vals is (red, blue , green), color_vals is the name...Color Vals is (red, white, blue ); package Fifth is new domainl.generateenum domain (ColorVals, Colors, enumeration, nullandnotnull); Resulting...generated domain definition-> type colorsnotnull is (red, white, blue ); package colors_ops is new sqlenumeration_pkg(colors notnull); type colors_type is
Joint Battlespace Infosphere: Information Management Within a C2 Enterprise
2005-06-01
using. In version 1.2, we support both MySQL and Oracle as underlying implementations where the XML metadata schema is mapped into relational tables in...Identity Servers, Role-Based Access Control, and Policy Representation – Databases: Oracle , MySQL , TigerLogic, Berkeley XML DB 15 Instrumentation Services...converted to SQL for execution. Invocations are then forwarded to the appropriate underlying IOR core components that have the responsibility of issuing
Analyzing Enron Data: Bitmap Indexing Outperforms MySQL Queries bySeveral Orders of Magnitude
DOE Office of Scientific and Technical Information (OSTI.GOV)
Stockinger, Kurt; Rotem, Doron; Shoshani, Arie
2006-01-28
FastBit is an efficient, compressed bitmap indexing technology that was developed in our group. In this report we evaluate the performance of MySQL and FastBit for analyzing the email traffic of the Enron dataset. The first finding shows that materializing the join results of several tables significantly improves the query performance. The second finding shows that FastBit outperforms MySQL by several orders of magnitude.
Reactive Aggregate Model Protecting Against Real-Time Threats
2014-09-01
on the underlying functionality of three core components. • MS SQL server 2008 backend database. • Microsoft IIS running on Windows server 2008...services. The capstone tested a Linux-based Apache web server with the following software implementations: • MySQL as a Linux-based backend server for...malicious compromise. 1. Assumptions • GINA could connect to a backend MS SQL database through proper configuration of DotNetNuke. • GINA had access
Information Security Considerations for Applications Using Apache Accumulo
2014-09-01
Distributed File System INSCOM United States Army Intelligence and Security Command JPA Java Persistence API JSON JavaScript Object Notation MAC Mandatory... MySQL [13]. BigTable can process 20 petabytes per day [14]. High degree of scalability on commodity hardware. NoSQL databases do not rely on highly...manipulation in relational databases. NoSQL databases each have a unique programming interface that uses a lower level procedural language (e.g., Java
Database Entity Persistence with Hibernate for the Network Connectivity Analysis Model
2014-04-01
time savings in the Java coding development process. Appendices A and B describe address setup procedures for installing the MySQL database...development environment is required: • The open source MySQL Database Management System (DBMS) from Oracle, which is a Java Database Connectivity (JDBC...compliant DBMS • MySQL JDBC Driver library that comes as a plug-in with the Netbeans distribution • The latest Java Development Kit with the latest
Learning Asset Technology Integration Support Tool Design Document
2010-05-11
language known as Hypertext Preprocessor ( PHP ) and by MySQL – a relational database management system that can also be used for content management. It...Requirements The LATIST tool will be implemented utilizing a WordPress platform with MySQL as the database. Also the LATIST system must effectively work... MySQL . When designing the LATIST system there are several considerations which must be accounted for in the working prototype. These include: • DAU
Standard Port-Visit Cost Forecasting Model for U.S. Navy Husbanding Contracts
2009-12-01
Protocol (HTTP) server.35 2. MySQL . An open-source database.36 3. PHP . A common scripting language used for Web development.37 E. IMPLEMENTATION OF...Inc. (2009). MySQL Community Server (Version 5.1) [Software]. Available from http://dev.mysql.com/downloads/ 37 The PHP Group (2009). PHP (Version...Logistics Services MySQL My Structured Query Language NAVSUP Navy Supply Systems Command NC Non-Contract Items NPS Naval Postgraduate
Cloud-Based Distributed Control of Unmanned Systems
2015-04-01
during mission execution. At best, the data is saved onto hard-drives and is accessible only by the local team. Data history in a form available and...following open source technologies: GeoServer, OpenLayers, PostgreSQL , and PostGIS are chosen to implement the back-end database and server. A brief...geospatial map data. 3. PostgreSQL : An SQL-compliant object-relational database that easily scales to accommodate large amounts of data - upwards to
SigmaCLIPSE = presentation management + NASA CLI PS + SQL
NASA Technical Reports Server (NTRS)
Weiss, Bernard P., Jr.
1990-01-01
SigmaCLIPSE provides an expert systems and 'intelligent' data base development program for diverse systems integration environments that require support for automated reasoning and expert systems technology, presentation management, and access to 'intelligent' SQL data bases. The SigmaCLIPSE technology and and its integrated ability to access 4th generation application development and decision support tools through a portable SQL interface, comprises a sophisticated software development environment for solving knowledge engineering and expert systems development problems in information intensive commercial environments -- financial services, health care, and distributed process control -- where the expert system must be extendable -- a major architectural advantage of NASA CLIPS. SigmaCLIPSE is a research effort intended to test the viability of merging SQL data bases with expert systems technology.
Using PHP/MySQL to Manage Potential Mass Impacts
NASA Technical Reports Server (NTRS)
Hager, Benjamin I.
2010-01-01
This paper presents a new application using commercially available software to manage mass properties for spaceflight vehicles. PHP/MySQL(PHP: Hypertext Preprocessor and My Structured Query Language) are a web scripting language and a database language commonly used in concert with each other. They open up new opportunities to develop cutting edge mass properties tools, and in particular, tools for the management of potential mass impacts (threats and opportunities). The paper begins by providing an overview of the functions and capabilities of PHP/MySQL. The focus of this paper is on how PHP/MySQL are being used to develop an advanced "web accessible" database system for identifying and managing mass impacts on NASA's Ares I Upper Stage program, managed by the Marshall Space Flight Center. To fully describe this application, examples of the data, search functions, and views are provided to promote, not only the function, but the security, ease of use, simplicity, and eye-appeal of this new application. This paper concludes with an overview of the other potential mass properties applications and tools that could be developed using PHP/MySQL. The premise behind this paper is that PHP/MySQL are software tools that are easy to use and readily available for the development of cutting edge mass properties applications. These tools are capable of providing "real-time" searching and status of an active database, automated report generation, and other capabilities to streamline and enhance mass properties management application. By using PHP/MySQL, proven existing methods for managing mass properties can be adapted to present-day information technology to accelerate mass properties data gathering, analysis, and reporting, allowing mass property management to keep pace with today's fast-pace design and development processes.
2002-06-01
Student memo for personnel MCLLS . . . . . . . . . . . . . . 75 i. Migrate data to SQL Server...The Web Server is on the same server as the SWORD database in the current version. 4: results set 5: dynamic HTML page 6: dynamic HTML page 3: SQL ...still be supported by Access. SQL Server would be a more viable tool for a fully developed application based on the number of potential users and
Multi-Resolution Playback of Network Trace Files
2015-06-01
a com- plete MySQL database, C++ developer tools and the libraries utilized in the development of the system (Boost and Libcrafter), and Wireshark...XE suite has a limit to the allowed size of each database. In order to be scalable, the project had to switch to the MySQL database suite. The...programs that access the database use the MySQL C++ connector, provided by Oracle, and the supplied methods and libraries. 4.4 Flow Generator Chapter 3
2013-01-01
commercial NoSQL database system. The results show that In-dexedHBase provides a data loading speed that is 6 times faster than Riak, and is...compare it with Riak, a widely adopted commercial NoSQL database system. The results show that In- dexedHBase provides a data loading speed that is 6...events. This chapter describes our research towards building an efficient and scalable storage platform for Truthy. Many existing NoSQL databases
Quality Attribute-Guided Evaluation of NoSQL Databases: A Case Study
2015-01-16
evaluations of NoSQL databases specifically, and big data systems in general, that have become apparent during our study. Keywords—NoSQL, distributed...technology, namely that of big data , software systems [1]. At the heart of big data systems are a collection of database technologies that are more...born organizations such as Google and Amazon [3][4], along with those of numerous other big data innovators, have created a variety of open source and
Novel Method of Storing and Reconstructing Events at Fermilab E-906/SeaQuest Using a MySQL Database
NASA Astrophysics Data System (ADS)
Hague, Tyler
2010-11-01
Fermilab E-906/SeaQuest is a fixed target experiment at Fermi National Accelerator Laboratory. We are investigating the antiquark asymmetry in the nucleon sea. By examining the ratio of the Drell- Yan cross sections of proton-proton and proton-deuterium collisions we can determine the asymmetry ratio. An essential feature in the development of the analysis software is to update the event reconstruction to modern software tools. We are doing this in a unique way by doing a majority of the calculations within an SQL database. Using a MySQL database allows us to take advantage of off-the-shelf software without sacrificing ROOT compatibility and avoid network bottlenecks with server-side data selection. Using our raw data we create stubs, or partial tracks, at each station which are pieced together to create full tracks. Our reconstruction process uses dynamically created SQL statements to analyze the data. These SQL statements create tables that contain the final reconstructed tracks as well as intermediate values. This poster will explain the reconstruction process and how it is being implemented.
Managing hydrological measurements for small and intermediate projects: RObsDat
NASA Astrophysics Data System (ADS)
Reusser, Dominik E.
2014-05-01
Hydrological measurements need good management for the data not to be lost. Multiple, often overlapping files from various loggers with heterogeneous formats need to be merged. Data needs to be validated and cleaned and subsequently converted to the format for the hydrological target application. Preferably, all these steps should be easily tracable. RObsDat is an R package designed to support such data management. It comes with a command line user interface to support hydrologists to enter and adjust their data in a database following the Observations Data Model (ODM) standard by QUASHI. RObsDat helps in the setup of the database within one of the free database engines MySQL, PostgreSQL or SQLite. It imports the controlled water vocabulary from the QUASHI web service and provides a smart interface between the hydrologist and the database: Already existing data entries are detected and duplicates avoided. The data import function converts different data table designes to make import simple. Cleaning and modifications of data are handled with a simple version control system. Variable and location names are treated in a user friendly way, accepting and processing multiple versions. A new development is the use of spacetime objects for subsequent processing.
Comparison of the Frontier Distributed Database Caching System to NoSQL Databases
DOE Office of Scientific and Technical Information (OSTI.GOV)
Dykstra, Dave
One of the main attractions of non-relational NoSQL databases is their ability to scale to large numbers of readers, including readers spread over a wide area. The Frontier distributed database caching system, used in production by the Large Hadron Collider CMS and ATLAS detector projects for Conditions data, is based on traditional SQL databases but also adds high scalability and the ability to be distributed over a wide-area for an important subset of applications. This paper compares the major characteristics of the two different approaches and identifies the criteria for choosing which approach to prefer over the other. It alsomore » compares in some detail the NoSQL databases used by CMS and ATLAS: MongoDB, CouchDB, HBase, and Cassandra.« less
Weissensteiner, Hansi; Schönherr, Sebastian; Specht, Günther; Kronenberg, Florian; Brandstätter, Anita
2010-03-09
Mitochondrial DNA (mtDNA) is widely being used for population genetics, forensic DNA fingerprinting and clinical disease association studies. The recent past has uncovered severe problems with mtDNA genotyping, not only due to the genotyping method itself, but mainly to the post-lab transcription, storage and report of mtDNA genotypes. eCOMPAGT, a system to store, administer and connect phenotype data to all kinds of genotype data is now enhanced by the possibility of storing mtDNA profiles and allowing their validation, linking to phenotypes and export as numerous formats. mtDNA profiles can be imported from different sequence evaluation programs, compared between evaluations and their haplogroup affiliations stored. Furthermore, eCOMPAGT has been improved in its sophisticated transparency (support of MySQL and Oracle), security aspects (by using database technology) and the option to import, manage and store genotypes derived from various genotyping methods (SNPlex, TaqMan, and STRs). It is a software solution designed for project management, laboratory work and the evaluation process all-in-one. The extended mtDNA version of eCOMPAGT was designed to enable error-free post-laboratory data handling of human mtDNA profiles. This software is suited for small to medium-sized human genetic, forensic and clinical genetic laboratories. The direct support of MySQL and the improved database security options render eCOMPAGT a powerful tool to build an automated workflow architecture for several genotyping methods. eCOMPAGT is freely available at http://dbis-informatik.uibk.ac.at/ecompagt.
2010-01-01
Background Mitochondrial DNA (mtDNA) is widely being used for population genetics, forensic DNA fingerprinting and clinical disease association studies. The recent past has uncovered severe problems with mtDNA genotyping, not only due to the genotyping method itself, but mainly to the post-lab transcription, storage and report of mtDNA genotypes. Description eCOMPAGT, a system to store, administer and connect phenotype data to all kinds of genotype data is now enhanced by the possibility of storing mtDNA profiles and allowing their validation, linking to phenotypes and export as numerous formats. mtDNA profiles can be imported from different sequence evaluation programs, compared between evaluations and their haplogroup affiliations stored. Furthermore, eCOMPAGT has been improved in its sophisticated transparency (support of MySQL and Oracle), security aspects (by using database technology) and the option to import, manage and store genotypes derived from various genotyping methods (SNPlex, TaqMan, and STRs). It is a software solution designed for project management, laboratory work and the evaluation process all-in-one. Conclusions The extended mtDNA version of eCOMPAGT was designed to enable error-free post-laboratory data handling of human mtDNA profiles. This software is suited for small to medium-sized human genetic, forensic and clinical genetic laboratories. The direct support of MySQL and the improved database security options render eCOMPAGT a powerful tool to build an automated workflow architecture for several genotyping methods. eCOMPAGT is freely available at http://dbis-informatik.uibk.ac.at/ecompagt. PMID:20214782
2001-09-01
replication) -- all from Visual Basic and VBA . In fact, we found that the SQL Server engine actually had a plethora of options, most formidable of...2002, the new SQL Server 2000 database engine, and Microsoft Visual Basic.NET. This thesis describes our use of the Spiral Development Model to...versions of Microsoft products? Specifically, the pending release of Microsoft Office 2002, the new SQL Server 2000 database engine, and Microsoft
An XML-Based Knowledge Management System of Port Information for U.S. Coast Guard Cutters
2003-03-01
using DTDs was not chosen. XML Schema performs many of the same functions as SQL type schemas, but differ by the unique structure of XML documents...to access data from content files within the developed system. XPath is not equivalent to SQL . While XPath is very powerful at reaching into an XML...document and finding nodes or node sets, it is not a complete query language. For operations like joins, unions, intersections, etc., SQL is far
Defense Information Systems Agency Technical Integration Support (DISA- TIS). MUMPS Study.
1993-01-01
usable in DoD, MUMPS must continue to improve in its support of DoD and OSE standards such as SQL , X-Windows, POSIX, PHIGS, etc. MUMPS and large AlSs...Language ( SQL ), X-Windows, and Graphical Kernel Services (GKS)) 2.2.2.3 FIPS Adoption by NIST The National Institute of Standards and Technology (NIST...many of the performance tuning mechanisms that must be performed explicitly with other systems. The VA looks forward to the SQL binding (1993 ANS) that
User Manual for Personnel Inventory Aging and Promotion Model
2009-06-01
increased by 12. Now, an SQL 9 statement deletes records where [target] = NULL, and the model calculates the number of E8s that need to be promoted to...the run, the [Likelihood] and [Expected] tables are created. The first step in this process is to dy- namically build an SQL statement, based on the...This table has individual-level, longitudinal records. Next, a dy- namically built SQL statement based on the Number of Years, cre- ates a new data
Reliability Information Analysis Center 1st Quarter 2007, Technical Area Task (TAT) Report
2007-02-05
34* Created new SQL server database for "PC Configuration" web application. Added roles for security closed 4235 and posted application to production. "e Wrote...and ran SQL Server scripts to migrate production databases to new server . "e Created backup jobs for new SQL Server databases. "* Continued...second phase of the TENA demo. Extensive tasking was established and assigned. A TENA interface to EW Server was reaffirmed after some uncertainty about
Spectrum Savings from High Performance Recording and Playback Onboard the Test Article
2013-02-20
execute within a Windows 7 environment, and data is recorded on SSDs. The underlying database is implemented using MySQL . Figure 1 illustrates the... MySQL database. This is effectively the time at which the recorded data are available for retransmission. CPU and Memory utilization were collected...17.7% MySQL avg. 3.9% EQDR Total avg. 21.6% Table 1 CPU Utilization with260 Mbits/sec Load The difference between the total System CPU (27.8
2013-01-01
website). Data mining tools are in-house code developed in Python, C++ and Java . • NGA The National Geospatial-Intelligence Agency (NGA) performs data...as PostgreSQL (with PostGIS), MySQL , Microsoft SQL Server, SQLite, etc. using the appropriate JDBC driver. 14 The documentation and ease to learn are...written in Java that is able to perform various types of regressions, classi- fications, and other data mining tasks. There is also a commercial version
A Brief Assessment of LC2IEDM, MIST and Web Services for use in Naval Tactical Data Management
2004-07-01
server software, messaging between the client and server, and a database. The MIST database is implemented in an open source DBMS named PostGreSQL ... PostGreSQL had its beginnings at the University of California, Berkley, in 1986 [11]. The development of PostGreSQL has since evolved into a...contact history from the database. DRDC Atlantic TM 2004-148 9 Request Software Request Software Server Side Response from service
Lee, Ken Ka-Yin; Tang, Wai-Choi; Choi, Kup-Sze
2013-04-01
Clinical data are dynamic in nature, often arranged hierarchically and stored as free text and numbers. Effective management of clinical data and the transformation of the data into structured format for data analysis are therefore challenging issues in electronic health records development. Despite the popularity of relational databases, the scalability of the NoSQL database model and the document-centric data structure of XML databases appear to be promising features for effective clinical data management. In this paper, three database approaches--NoSQL, XML-enabled and native XML--are investigated to evaluate their suitability for structured clinical data. The database query performance is reported, together with our experience in the databases development. The results show that NoSQL database is the best choice for query speed, whereas XML databases are advantageous in terms of scalability, flexibility and extensibility, which are essential to cope with the characteristics of clinical data. While NoSQL and XML technologies are relatively new compared to the conventional relational database, both of them demonstrate potential to become a key database technology for clinical data management as the technology further advances. Copyright © 2012 Elsevier Ireland Ltd. All rights reserved.
Assessment & Commitment Tracking System (ACTS)
DOE Office of Scientific and Technical Information (OSTI.GOV)
Bryant, Robert A.; Childs, Teresa A.; Miller, Michael A.
2004-12-20
The ACTS computer code provides a centralized tool for planning and scheduling assessments, tracking and managing actions associated with assessments or that result from an event or condition, and "mining" data for reporting and analyzing information for improving performance. The ACTS application is designed to work with the MS SQL database management system. All database interfaces are written in SQL. The following software is used to develop and support the ACTS application: Cold Fusion HTML JavaScript Quest TOAD Microsoft Visual Source Safe (VSS) HTML Mailer for sending email Microsoft SQL Microsoft Internet Information Server
NASA Technical Reports Server (NTRS)
McGlynn, T.; Santisteban, M.
2007-01-01
This chapter provides a very brief introduction to the Structured Query Language (SQL) for getting information from relational databases. We make no pretense that this is a complete or comprehensive discussion of SQL. There are many aspects of the language the will be completely ignored in the presentation. The goal here is to provide enough background so that users understand the basic concepts involved in building and using relational databases. We also go through the steps involved in building a particular astronomical database used in some of the other presentations in this volume.
Integrated Array/Metadata Analytics
NASA Astrophysics Data System (ADS)
Misev, Dimitar; Baumann, Peter
2015-04-01
Data comes in various forms and types, and integration usually presents a problem that is often simply ignored and solved with ad-hoc solutions. Multidimensional arrays are an ubiquitous data type, that we find at the core of virtually all science and engineering domains, as sensor, model, image, statistics data. Naturally, arrays are richly described by and intertwined with additional metadata (alphanumeric relational data, XML, JSON, etc). Database systems, however, a fundamental building block of what we call "Big Data", lack adequate support for modelling and expressing these array data/metadata relationships. Array analytics is hence quite primitive or non-existent at all in modern relational DBMS. Recognizing this, we extended SQL with a new SQL/MDA part seamlessly integrating multidimensional array analytics into the standard database query language. We demonstrate the benefits of SQL/MDA with real-world examples executed in ASQLDB, an open-source mediator system based on HSQLDB and rasdaman, that already implements SQL/MDA.
Processing of the WLCG monitoring data using NoSQL
NASA Astrophysics Data System (ADS)
Andreeva, J.; Beche, A.; Belov, S.; Dzhunov, I.; Kadochnikov, I.; Karavakis, E.; Saiz, P.; Schovancova, J.; Tuckett, D.
2014-06-01
The Worldwide LHC Computing Grid (WLCG) today includes more than 150 computing centres where more than 2 million jobs are being executed daily and petabytes of data are transferred between sites. Monitoring the computing activities of the LHC experiments, over such a huge heterogeneous infrastructure, is extremely demanding in terms of computation, performance and reliability. Furthermore, the generated monitoring flow is constantly increasing, which represents another challenge for the monitoring systems. While existing solutions are traditionally based on Oracle for data storage and processing, recent developments evaluate NoSQL for processing large-scale monitoring datasets. NoSQL databases are getting increasingly popular for processing datasets at the terabyte and petabyte scale using commodity hardware. In this contribution, the integration of NoSQL data processing in the Experiment Dashboard framework is described along with first experiences of using this technology for monitoring the LHC computing activities.
Integrating Scientific Array Processing into Standard SQL
NASA Astrophysics Data System (ADS)
Misev, Dimitar; Bachhuber, Johannes; Baumann, Peter
2014-05-01
We live in a time that is dominated by data. Data storage is cheap and more applications than ever accrue vast amounts of data. Storing the emerging multidimensional data sets efficiently, however, and allowing them to be queried by their inherent structure, is a challenge many databases have to face today. Despite the fact that multidimensional array data is almost always linked to additional, non-array information, array databases have mostly developed separately from relational systems, resulting in a disparity between the two database categories. The current SQL standard and SQL DBMS supports arrays - and in an extension also multidimensional arrays - but does so in a very rudimentary and inefficient way. This poster demonstrates the practicality of an SQL extension for array processing, implemented in a proof-of-concept multi-faceted system that manages a federation of array and relational database systems, providing transparent, efficient and scalable access to the heterogeneous data in them.
Interactive DataBase of Cosmic Ray Anisotropy (DB A10)
NASA Astrophysics Data System (ADS)
Asipenka, A.S.; Belov, A.V.; Eroshenko, E.F.; Klepach, E.G.; Oleneva, V.A.; Yake, V.G.
Data on the hourly means of cosmic ray density and anisotropy derived by the GSM method over the 1957-2006 are introduced in to MySQL database. This format allowed an access to data both in local and in the Internet. Using the realized combination of script-language Php and My SQL database the Internet project was created on the access for users data on the CR anisotropy in different formats (http://cr20.izmiran.ru/AnisotropyCR/main.htm/). Usage the sheaf Php and MySQL provides fast receiving data even in the Internet since a request and following process of data are accomplished on the project server. Usage of MySQL basis for the storing data on cosmic ray variations give a possibility to construct requests of different structures, extends the variety of data reflection, makes it possible the conformity data to other systems and usage them in other projects.
SQL Triggers Reacting on Time Events: An Extension Proposal
NASA Astrophysics Data System (ADS)
Behrend, Andreas; Dorau, Christian; Manthey, Rainer
Being able to activate triggers at timepoints reached or after time intervals elapsed has been acknowledged by many authors as a valuable functionality of a DBMS. Recently, the interest in time-based triggers has been renewed in the context of data stream monitoring. However, up till now SQL triggers react to data changes only, even though research proposals and prototypes have been supporting several other event types, in particular time-based ones, since long. We therefore propose a seamless extension of the SQL trigger concept by time-based triggers, focussing on semantic issues arising from such an extension.
Advancing Future Network Science through Content Understanding
2014-05-01
BitTorrent, PostgreSQL, MySQL , and GRSecurity) and emerging technologies (HadoopDFS, Tokutera, Sector/Sphere, HBase, and other BigTable-like...result. • Multi-Source Network Pulse Analyzer and Correlator provides course of action planning by enhancing the understanding of the complex dynamics
Streamlining the Process of Acquiring Secure Open Architecture Software Systems
2013-10-08
Microsoft.NET, Enterprise Java Beans, GNU Lesser General Public License (LGPL) libraries, and data communication protocols like the Hypertext Transfer...NetBeans development environments), customer relationship management (SugarCRM), database management systems (PostgreSQL, MySQL ), operating
Trends and New Directions in Software Architecture
2014-10-10
frameworks Open source Cloud strategies NoSQL Machine Learning MDD Incremental approaches Dashboards Distributed development...complexity grows NoSQL Models are not created equal 2014 Our Current Research Lightweight Evaluation and Architecture Prototyping for Big Data
Benchmarking database performance for genomic data.
Khushi, Matloob
2015-06-01
Genomic regions represent features such as gene annotations, transcription factor binding sites and epigenetic modifications. Performing various genomic operations such as identifying overlapping/non-overlapping regions or nearest gene annotations are common research needs. The data can be saved in a database system for easy management, however, there is no comprehensive database built-in algorithm at present to identify overlapping regions. Therefore I have developed a novel region-mapping (RegMap) SQL-based algorithm to perform genomic operations and have benchmarked the performance of different databases. Benchmarking identified that PostgreSQL extracts overlapping regions much faster than MySQL. Insertion and data uploads in PostgreSQL were also better, although general searching capability of both databases was almost equivalent. In addition, using the algorithm pair-wise, overlaps of >1000 datasets of transcription factor binding sites and histone marks, collected from previous publications, were reported and it was found that HNF4G significantly co-locates with cohesin subunit STAG1 (SA1).Inc. © 2015 Wiley Periodicals, Inc.
NASA Astrophysics Data System (ADS)
Krishna, B.; Gustafson, W. I., Jr.; Vogelmann, A. M.; Toto, T.; Devarakonda, R.; Palanisamy, G.
2016-12-01
This paper presents a new way of providing ARM data discovery through data analysis and visualization services. ARM stands for Atmospheric Radiation Measurement. This Program was created to study cloud formation processes and their influence on radiative transfer and also include additional measurements of aerosol and precipitation at various highly instrumented ground and mobile stations. The total volume of ARM data is roughly 900TB. The current search for ARM data is performed by using its metadata, such as the site name, instrument name, date, etc. NoSQL technologies were explored to improve the capabilities of data searching, not only by their metadata, but also by using the measurement values. Two technologies that are currently being implemented for testing are Apache Cassandra (noSQL database) and Apache Spark (noSQL based analytics framework). Both of these technologies were developed to work in a distributed environment and hence can handle large data for storing and analytics. D3.js is a JavaScript library that can generate interactive data visualizations in web browsers by making use of commonly used SVG, HTML5, and CSS standards. To test the performance of NoSQL for ARM data, we will be using ARM's popular measurements to locate the data based on its value. Recently noSQL technology has been applied to a pilot project called LASSO, which stands for LES ARM Symbiotic Simulation and Observation Workflow. LASSO will be packaging LES output and observations in "data bundles" and analyses will require the ability for users to analyze both observations and LES model output either individually or together across multiple time periods. The LASSO implementation strategy suggests that enormous data storage is required to store the above mentioned quantities. Thus noSQL was used to provide a powerful means to store portions of the data that provided users with search capabilities on each simulation's traits through a web application. Based on the user selection, plots are created dynamically along with ancillary information that enables the user to locate and download data that fulfilled their required traits.
Hewitt, Robin; Gobbi, Alberto; Lee, Man-Ling
2005-01-01
Relational databases are the current standard for storing and retrieving data in the pharmaceutical and biotech industries. However, retrieving data from a relational database requires specialized knowledge of the database schema and of the SQL query language. At Anadys, we have developed an easy-to-use system for searching and reporting data in a relational database to support our drug discovery project teams. This system is fast and flexible and allows users to access all data without having to write SQL queries. This paper presents the hierarchical, graph-based metadata representation and SQL-construction methods that, together, are the basis of this system's capabilities.
Kennedy, David J; Engel, Andrew; Kreiner, D Scott; Nampiaparampil, Devi; Duszynski, Belinda; MacVicar, John
2015-08-01
To assess the validity of fluoroscopically guided diagnostic intra-articular injections of local anesthetic and effectiveness of intra-articular steroid injections in treating sacroiliac joint (SIJ) pain. Systematic review. Ten reviewers independently assessed 45 publications on diagnostic validity or effectiveness of fluoroscopically guided intra-articular SIJ injections. For diagnostic injections, the primary outcome was validity; for therapeutic injections, analgesia. Secondary outcomes were also described. Of 45 articles reviewed, 39 yielded diagnostic data on physical exam findings, provocation tests, and SIJ injections for diagnosing SIJ pain, and 15 addressed therapeutic effectiveness. When confirmed by comparative local anesthetic blocks with a high degree of pain relief, no single physical exam maneuver predicts response to diagnostic injections. When at least three physical exam findings are present, sensitivity, and specificity increases significantly. The prevalence of SIJ pain is likely 20-30% among patients that have suspected SIJ pain based on history and physical examination. This estimate may be higher in certain subgroups such as the elderly and fusion patients. Two randomized controlled trials and multiple observational studies supported the effectiveness of therapeutic sacroiliac joint injections. Based on this literature, it is unclear whether image-guided intra-articular diagnostic injections of local anesthetic predict positive responses to therapeutic agents. The overall quality of evidence is moderate for the effectiveness of therapeutic SIJ injections. Wiley Periodicals, Inc.
A Web-based Examination System Based on PHP+MySQL.
Wen, Ji; Zhang, Yang; Yan, Yong; Xia, Shunren
2005-01-01
The design and implementation of web-based examination system constructed by PHP and MySQL is presented in this paper. Three primary parts, including students',teachers' and administrators', are introduced and analyzed in detail. Initial application has demonstrated the system's feasibility and reasonability.*
An effective model for store and retrieve big health data in cloud computing.
Goli-Malekabadi, Zohreh; Sargolzaei-Javan, Morteza; Akbari, Mohammad Kazem
2016-08-01
The volume of healthcare data including different and variable text types, sounds, and images is increasing day to day. Therefore, the storage and processing of these data is a necessary and challenging issue. Generally, relational databases are used for storing health data which are not able to handle the massive and diverse nature of them. This study aimed at presenting the model based on NoSQL databases for the storage of healthcare data. Despite different types of NoSQL databases, document-based DBs were selected by a survey on the nature of health data. The presented model was implemented in the Cloud environment for accessing to the distribution properties. Then, the data were distributed on the database by applying the Shard property. The efficiency of the model was evaluated in comparison with the previous data model, Relational Database, considering query time, data preparation, flexibility, and extensibility parameters. The results showed that the presented model approximately performed the same as SQL Server for "read" query while it acted more efficiently than SQL Server for "write" query. Also, the performance of the presented model was better than SQL Server in the case of flexibility, data preparation and extensibility. Based on these observations, the proposed model was more effective than Relational Databases for handling health data. Copyright © 2016 Elsevier Ireland Ltd. All rights reserved.
Evaluation of NoSQL databases for DIRAC monitoring and beyond
NASA Astrophysics Data System (ADS)
Mathe, Z.; Casajus Ramo, A.; Stagni, F.; Tomassetti, L.
2015-12-01
Nowadays, many database systems are available but they may not be optimized for storing time series data. Monitoring DIRAC jobs would be better done using a database optimised for storing time series data. So far it was done using a MySQL database, which is not well suited for such an application. Therefore alternatives have been investigated. Choosing an appropriate database for storing huge amounts of time series data is not trivial as one must take into account different aspects such as manageability, scalability and extensibility. We compared the performance of Elasticsearch, OpenTSDB (based on HBase) and InfluxDB NoSQL databases, using the same set of machines and the same data. We also evaluated the effort required for maintaining them. Using the LHCb Workload Management System (WMS), based on DIRAC as a use case we set up a new monitoring system, in parallel with the current MySQL system, and we stored the same data into the databases under test. We evaluated Grafana (for OpenTSDB) and Kibana (for ElasticSearch) metrics and graph editors for creating dashboards, in order to have a clear picture on the usability of each candidate. In this paper we present the results of this study and the performance of the selected technology. We also give an outlook of other potential applications of NoSQL databases within the DIRAC project.
Fault Injection Validation of a Safety-Critical TMR Sysem
NASA Astrophysics Data System (ADS)
Irrera, Ivano; Madeira, Henrique; Zentai, Andras; Hergovics, Beata
2016-08-01
Digital systems and their software are the core technology for controlling and monitoring industrial systems in practically all activity domains. Functional safety standards such as the European standard EN 50128 for railway applications define the procedures and technical requirements for the development of software for railway control and protection systems. The validation of such systems is a highly demanding task. In this paper we discuss the use of fault injection techniques, which have been used extensively in several domains, particularly in the space domain, to complement the traditional procedures to validate a SIL (Safety Integrity Level) 4 system for railway signalling, implementing a TMR (Triple Modular Redundancy) architecture. The fault injection tool is based on JTAG technology. The results of our injection campaign showed a high degree of tolerance to most of the injected faults, but several cases of unexpected behaviour have also been observed, helping understanding worst-case scenarios.
DOE Office of Scientific and Technical Information (OSTI.GOV)
David Nix, Lisa Simirenko
2006-10-25
The Biolmaging Database (BID) is a relational database developed to store the data and meta-data for the 3D gene expression in early Drosophila embryo development on a cellular level. The schema was written to be used with the MySQL DBMS but with minor modifications can be used on any SQL compliant relational DBMS.
The SQL Server Database for Non Computer Professional Teaching Reform
ERIC Educational Resources Information Center
Liu, Xiangwei
2012-01-01
A summary of the teaching methods of the non-computer professional SQL Server database, analyzes the current situation of the teaching course. According to non computer professional curriculum teaching characteristic, put forward some teaching reform methods, and put it into practice, improve the students' analysis ability, practice ability and…
Implementing CBM: SQL-Tutor after Fifteen Years
ERIC Educational Resources Information Center
Mitrovic, Antonija; Ohlsson, Stellan
2016-01-01
SQL-Tutor is the first constraint-based tutor. The initial conference papers about the system were published in 1998 (Mitrovic 1998a, 1998b, 1998c), with an "IJAIED" paper published in 1999 (Mitrovic and Ohlsson, "International Journal Artificial Intelligence in Education," 10(3-4), 238-256, 1999). We published another…
CheD: chemical database compilation tool, Internet server, and client for SQL servers.
Trepalin, S V; Yarkov, A V
2001-01-01
An efficient program, which runs on a personal computer, for the storage, retrieval, and processing of chemical information, is presented, The program can work both as a stand-alone application or in conjunction with a specifically written Web server application or with some standard SQL servers, e.g., Oracle, Interbase, and MS SQL. New types of data fields are introduced, e.g., arrays for spectral information storage, HTML and database links, and user-defined functions. CheD has an open architecture; thus, custom data types, controls, and services may be added. A WWW server application for chemical data retrieval features an easy and user-friendly installation on Windows NT or 95 platforms.
Agentless Cloud-Wide Monitoring of Virtual Disk State
2015-10-01
packages include Apache, MySQL , PHP, Ruby on Rails, Java Application Servers, and many others. Figure 2.12 shows the results of a run of the Software...Linux, Apache, MySQL , PHP (LAMP) set of applications. Thus, many file-level update logs will contain the same versions of files repeated across many
Comparing IndexedHBase and Riak for Serving Truthy: Performance of Data Loading and Query Evaluation
2013-08-01
Research Triangle Park, NC 27709-2211 15. SUBJECT TERMS performance evaluation, distributed database, noSQL , HBase, indexing Xiaoming Gao, Judy Qiu...common hashtags created during a given time window. With the purpose of finding a solution for these challenges, we evaluate NoSQL databases such as
2012-11-27
with powerful analysis tools and an informatics approach leveraging best-of-breed NoSQL databases, in order to store, search and retrieve relevant...dictionaries, and JavaScript also has good support. The MongoDB project[15] was chosen as a scalable NoSQL data store for the cheminfor- matics components
An Experimental Investigation of Complexity in Database Query Formulation Tasks
ERIC Educational Resources Information Center
Casterella, Gretchen Irwin; Vijayasarathy, Leo
2013-01-01
Information Technology professionals and other knowledge workers rely on their ability to extract data from organizational databases to respond to business questions and support decision making. Structured query language (SQL) is the standard programming language for querying data in relational databases, and SQL skills are in high demand and are…
Knowledge Query Language (KQL)
2016-02-12
Lexington Massachusetts This page intentionally left blank. iii EXECUTIVE SUMMARY Currently, queries for data ...retrieval from non-Structured Query Language (NoSQL) data stores are tightly coupled to the specific implementation of the data store implementation...independent of the storage content and format for querying NoSQL or relational data stores. This approach uses address expressions (or A-Expressions
Research of GIS-services applicability for solution of spatial analysis tasks.
NASA Astrophysics Data System (ADS)
Terekhin, D. A.; Botygin, I. A.; Sherstneva, A. I.; Sherstnev, V. S.
2017-01-01
Experiments for working out the areas of applying various gis-services in the tasks of spatial analysis are discussed in this paper. Google Maps, Yandex Maps, Microsoft SQL Server are used as services of spatial analysis. All services have shown a comparable speed of analyzing the spatial data when carrying out elemental spatial requests (building up the buffer zone of a point object) as well as the preferences of Microsoft SQL Server in operating with more complicated spatial requests. When building up elemental spatial requests, internet-services show higher efficiency due to cliental data handling with JavaScript-subprograms. A weak point of public internet-services is an impossibility to handle data on a server side and a barren variety of spatial analysis functions. Microsoft SQL Server offers a large variety of functions needed for spatial analysis on the server side. The authors conclude that when solving practical problems, the capabilities of internet-services used in building up routes and completing other functions with spatial analysis with Microsoft SQL Server should be involved.
2MASS Catalog Server Kit Version 2.1
NASA Astrophysics Data System (ADS)
Yamauchi, C.
2013-10-01
The 2MASS Catalog Server Kit is open source software for use in easily constructing a high performance search server for important astronomical catalogs. This software utilizes the open source RDBMS PostgreSQL, therefore, any users can setup the database on their local computers by following step-by-step installation guide. The kit provides highly optimized stored functions for positional searchs similar to SDSS SkyServer. Together with these, the powerful SQL environment of PostgreSQL will meet various user's demands. We released 2MASS Catalog Server Kit version 2.1 in 2012 May, which supports the latest WISE All-Sky catalog (563,921,584 rows) and 9 major all-sky catalogs. Local databases are often indispensable for observatories with unstable or narrow-band networks or severe use, such as retrieving large numbers of records within a small period of time. This software is the best for such purposes, and increasing supported catalogs and improvements of version 2.1 can cover a wider range of applications including advanced calibration system, scientific studies using complicated SQL queries, etc. Official page: http://www.ir.isas.jaxa.jp/~cyamauch/2masskit/
Use of Graph Database for the Integration of Heterogeneous Biological Data.
Yoon, Byoung-Ha; Kim, Seon-Kyu; Kim, Seon-Young
2017-03-01
Understanding complex relationships among heterogeneous biological data is one of the fundamental goals in biology. In most cases, diverse biological data are stored in relational databases, such as MySQL and Oracle, which store data in multiple tables and then infer relationships by multiple-join statements. Recently, a new type of database, called the graph-based database, was developed to natively represent various kinds of complex relationships, and it is widely used among computer science communities and IT industries. Here, we demonstrate the feasibility of using a graph-based database for complex biological relationships by comparing the performance between MySQL and Neo4j, one of the most widely used graph databases. We collected various biological data (protein-protein interaction, drug-target, gene-disease, etc.) from several existing sources, removed duplicate and redundant data, and finally constructed a graph database containing 114,550 nodes and 82,674,321 relationships. When we tested the query execution performance of MySQL versus Neo4j, we found that Neo4j outperformed MySQL in all cases. While Neo4j exhibited a very fast response for various queries, MySQL exhibited latent or unfinished responses for complex queries with multiple-join statements. These results show that using graph-based databases, such as Neo4j, is an efficient way to store complex biological relationships. Moreover, querying a graph database in diverse ways has the potential to reveal novel relationships among heterogeneous biological data.
Use of Graph Database for the Integration of Heterogeneous Biological Data
Yoon, Byoung-Ha; Kim, Seon-Kyu
2017-01-01
Understanding complex relationships among heterogeneous biological data is one of the fundamental goals in biology. In most cases, diverse biological data are stored in relational databases, such as MySQL and Oracle, which store data in multiple tables and then infer relationships by multiple-join statements. Recently, a new type of database, called the graph-based database, was developed to natively represent various kinds of complex relationships, and it is widely used among computer science communities and IT industries. Here, we demonstrate the feasibility of using a graph-based database for complex biological relationships by comparing the performance between MySQL and Neo4j, one of the most widely used graph databases. We collected various biological data (protein-protein interaction, drug-target, gene-disease, etc.) from several existing sources, removed duplicate and redundant data, and finally constructed a graph database containing 114,550 nodes and 82,674,321 relationships. When we tested the query execution performance of MySQL versus Neo4j, we found that Neo4j outperformed MySQL in all cases. While Neo4j exhibited a very fast response for various queries, MySQL exhibited latent or unfinished responses for complex queries with multiple-join statements. These results show that using graph-based databases, such as Neo4j, is an efficient way to store complex biological relationships. Moreover, querying a graph database in diverse ways has the potential to reveal novel relationships among heterogeneous biological data. PMID:28416946
Farhan, Bilal; Soltani, Tandis; Do, Rebecca; Perez, Claudia; Choi, Hanul; Ghoniem, Gamal
2018-05-02
Endoscopic injection of urethral bulking agents is an office procedure that is used to treat stress urinary incontinence secondary to internal sphincteric deficiency. Validation studies important part of simulator evaluation and is considered important step to establish the effectiveness of simulation-based training. The endoscopic needle injection (ENI) simulator has not been formally validated, although it has been used widely at University of California, Irvine. We aimed to assess the face, content, and construct validity of the UC, Irvine ENI simulator. Dissected female porcine bladders were mounted in a modified Hysteroscopy Diagnostic Trainer. Using routine endoscopic equipment for this procedure with video monitoring, 6 urologists (experts group) and 6 urology trainee (novice group) completed urethral bulking agents injections on a total of 12 bladders using ENI simulator. Face and content validities were assessed by using structured quantitative survey which rating the realism. Construct validity was assessed by comparing the performance, time of the procedure, and the occlusive (anatomical and functional) evaluations between the experts and novices. Trainees also completed a postprocedure feedback survey. Effective injections were evaluated by measuring the retrograde urethral opening pressure, visual cystoscopic coaptation, and postprocedure gross anatomic examination. All 12 participants felt the simulator was a good training tool and should be used as essential part of urology training (face validity). ENI simulator showed good face and content validity with average score varies between the experts and the novices was 3.9/5 and 3.8/5, respectively. Content validity evaluation showed that most aspects of the simulator were adequately realistic (mean Likert scores 3.9-3.8/5). However, the bladder does not bleed, and sometimes thin. Experts significantly outperformed novices (p < 001) across all measure of performance therefore establishing construct validity. The ENI simulator shows face, content and construct validities, although few aspects of simulator were not very realistic (e.g., bleeding).This study provides a base for the future formal validation for this simulator and for continuing use of this simulator in endourology training. Copyright © 2018 Association of Program Directors in Surgery. Published by Elsevier Inc. All rights reserved.
Design and implementation of website information disclosure assessment system.
Cho, Ying-Chiang; Pan, Jen-Yi
2015-01-01
Internet application technologies, such as cloud computing and cloud storage, have increasingly changed people's lives. Websites contain vast amounts of personal privacy information. In order to protect this information, network security technologies, such as database protection and data encryption, attract many researchers. The most serious problems concerning web vulnerability are e-mail address and network database leakages. These leakages have many causes. For example, malicious users can steal database contents, taking advantage of mistakes made by programmers and administrators. In order to mitigate this type of abuse, a website information disclosure assessment system is proposed in this study. This system utilizes a series of technologies, such as web crawler algorithms, SQL injection attack detection, and web vulnerability mining, to assess a website's information disclosure. Thirty websites, randomly sampled from the top 50 world colleges, were used to collect leakage information. This testing showed the importance of increasing the security and privacy of website information for academic websites.
Model for Semantically Rich Point Cloud Data
NASA Astrophysics Data System (ADS)
Poux, F.; Neuville, R.; Hallot, P.; Billen, R.
2017-10-01
This paper proposes an interoperable model for managing high dimensional point clouds while integrating semantics. Point clouds from sensors are a direct source of information physically describing a 3D state of the recorded environment. As such, they are an exhaustive representation of the real world at every scale: 3D reality-based spatial data. Their generation is increasingly fast but processing routines and data models lack of knowledge to reason from information extraction rather than interpretation. The enhanced smart point cloud developed model allows to bring intelligence to point clouds via 3 connected meta-models while linking available knowledge and classification procedures that permits semantic injection. Interoperability drives the model adaptation to potentially many applications through specialized domain ontologies. A first prototype is implemented in Python and PostgreSQL database and allows to combine semantic and spatial concepts for basic hybrid queries on different point clouds.
Distributed Episodic Exploratory Planning (DEEP)
2008-12-01
API). For DEEP, Hibernate offered the following advantages: • Abstracts SQL by utilizing HQL so any database with a Java Database Connectivity... Hibernate SQL ICCRTS International Command and Control Research and Technology Symposium JDB Java Distributed Blackboard JDBC Java Database Connectivity...selected because of its opportunistic reasoning capabilities and implemented in Java for platform independence. Java was chosen for ease of
Investigating the Limitations of Advanced Design Methods through Real World Application
2016-03-31
36 War Room Laptop Display ( MySQL , JMP 9 Pro, 64-bit Windows) Georgia Tech Secure Collaborative Visualization Environment ( MySQL , JMP 9 Pro...investigate expanding the EA for VC3ATS • Would like to consider both an expansion of the use of current Java -based BPM approach and other potential EA
ERIC Educational Resources Information Center
Piyayodilokchai, Hongsiri; Panjaburee, Patcharin; Laosinchai, Parames; Ketpichainarong, Watcharee; Ruenwongsa, Pintip
2013-01-01
With the benefit of multimedia and the learning cycle approach in promoting effective active learning, this paper proposed a learning cycle approach-based, multimedia-supplemented instructional unit for Structured Query Language (SQL) for second-year undergraduate students with the aim of enhancing their basic knowledge of SQL and ability to apply…
Large Declarative Memories in ACT-R
2009-12-01
containing the persistent DM of interest PDM-user Username required by the PostgreSQL DBMS for DB access PDM- passwd Password required by the PostgreSQL...34model-v5-DM" :pdm-user "Scott" :pdm- passwd “Open_Seseme" :pdm-resets-clear-db T :pdm-add-dm-serializes T :pdm-active T ... Figure 1: Activating and
Software Application for Supporting the Education of Database Systems
ERIC Educational Resources Information Center
Vágner, Anikó
2015-01-01
The article introduces an application which supports the education of database systems, particularly the teaching of SQL and PL/SQL in Oracle Database Management System environment. The application has two parts, one is the database schema and its content, and the other is a C# application. The schema is to administrate and store the tasks and the…
2008-03-01
Fortunately, built into Excel is the capability to use ActiveX Data Objects (ADO), a software feature which uses VBA to interface with external...part of Excel’s ActiveX Direct Objects (ADO) functionality, Excel can execute SQL queries in Access with VBA. An SQL query statement can be written
ERIC Educational Resources Information Center
Mills, Robert J.; Dupin-Bryant, Pamela A.; Johnson, John D.; Beaulieu, Tanya Y.
2015-01-01
The demand for Information Systems (IS) graduates with expertise in Structured Query Language (SQL) and database management is vast and projected to increase as "big data" becomes ubiquitous. To prepare students to solve complex problems in a data-driven world, educators must explore instructional strategies to help link prior knowledge…
Knowledge Query Language (KQL)
2016-02-01
unlimited. This page intentionally left blank. iii EXECUTIVE SUMMARY Currently, queries for data ...retrieval from non-Structured Query Language (NoSQL) data stores are tightly coupled to the specific implementation of the data store implementation, making...of the storage content and format for querying NoSQL or relational data stores. This approach uses address expressions (or A-Expressions) embedded in
Evolution of the LBT Telemetry System
NASA Astrophysics Data System (ADS)
Summers, K.; Biddick, C.; De La Peña, M. D.; Summers, D.
2014-05-01
The Large Binocular Telescope (LBT) Telescope Control System (TCS) records about 10GB of telemetry data per night. Additionally, the vibration monitoring system records about 9GB of telemetry data per night. Through 2013, we have amassed over 6TB of Hierarchical Data Format (HDF5) files and almost 9TB in a MySQL database of TCS and vibration data. The LBT telemetry system, in its third major revision since 2004, provides the mechanism to capture and store this data. The telemetry system has evolved from a simple HDF file system with MySQL stream definitions within the TCS, to a separate system using a MySQL database system for the definitions and data, and finally to no database use at all, using HDF5 files.
Quantum displacement receiver for M-ary phase-shift-keyed coherent states
DOE Office of Scientific and Technical Information (OSTI.GOV)
Izumi, Shuro; Takeoka, Masahiro; Fujiwara, Mikio
2014-12-04
We propose quantum receivers for 3- and 4-ary phase-shift-keyed (PSK) coherent state signals to overcome the standard quantum limit (SQL). Our receiver, consisting of a displacement operation and on-off detectors with or without feedforward, provides an error probability performance beyond the SQL. We show feedforward operations can tolerate the requirement for the detector specifications.
Tomcat, Oracle & XML Web Archive
DOE Office of Scientific and Technical Information (OSTI.GOV)
Cothren, D. C.
2008-01-01
The TOX (Tomcat Oracle & XML) web archive is a foundation for development of HTTP-based applications using Tomcat (or some other servlet container) and an Oracle RDBMS. Use of TOX requires coding primarily in PL/SQL, JavaScript, and XSLT, but also in HTML, CSS and potentially Java. Coded in Java and PL/SQL itself, TOX provides the foundation for more complex applications to be built.
Architecture Knowledge for Evaluating Scalable Databases
2015-01-16
problems, arising from the proliferation of new data models and distributed technologies for building scalable, available data stores . Architects must...longer are relational databases the de facto standard for building data repositories. Highly distributed, scalable “ NoSQL ” databases [11] have emerged...This is especially challenging at the data storage layer. The multitude of competing NoSQL database technologies creates a complex and rapidly
DOE Office of Scientific and Technical Information (OSTI.GOV)
This software is a plug-in that interfaces between the Phoenix Integration's Model Center and the Base SAS 9.2 applications. The end use of the plug-in is to link input and output data that resides in SAS tables or MS SQL to and from "legacy" software programs without recoding. The potential end users are users who need to run legacy code and want data stored in a SQL database.
Comparison of the Frontier Distributed Database Caching System to NoSQL Databases
NASA Astrophysics Data System (ADS)
Dykstra, Dave
2012-12-01
One of the main attractions of non-relational “NoSQL” databases is their ability to scale to large numbers of readers, including readers spread over a wide area. The Frontier distributed database caching system, used in production by the Large Hadron Collider CMS and ATLAS detector projects for Conditions data, is based on traditional SQL databases but also adds high scalability and the ability to be distributed over a wide-area for an important subset of applications. This paper compares the major characteristics of the two different approaches and identifies the criteria for choosing which approach to prefer over the other. It also compares in some detail the NoSQL databases used by CMS and ATLAS: MongoDB, CouchDB, HBase, and Cassandra.
Classification of HTTP Attacks: A Study on the ECML/PKDD 2007 Discovery Challenge
DOE Office of Scientific and Technical Information (OSTI.GOV)
Gallagher, Brian; Eliassi-Rad, Tina
2009-07-08
As the world becomes more reliant on Web applications for commercial, financial, and medical transactions, cyber attacks on the World Wide Web are increasing in frequency and severity. Web applications provide an attractive alternative to traditional desktop applications due to their accessibility and ease of deployment. However, the accessibility of Web applications also makes them extremely vulnerable to attack. This inherent vulnerability is intensified by the distributed nature ofWeb applications and the complexity of configuring application servers. These factors have led to a proliferation of Web-based attacks, in which attackers surreptitiously inject code into HTTP requests, allowing them to executemore » arbitrary commands on remote systems and perform malicious activities such as reading, altering, or destroying sensitive data. One approach for dealing with HTTP-based attacks is to identify malicious code in incoming HTTP requests and eliminate bad requests before they are processed. Using machine learning techniques, we can build a classifier to automatically label requests as “Valid” or “Attack.” For this study, we develop a simple, but effective HTTP attack classifier, based on the vector space model used commonly for Information Retrieval. Our classifier not only separates attacks from valid requests, but can also identify specific attack types (e.g., “SQL Injection” or “Path Traversal”). We demonstrate the effectiveness of our approach through experiments on the ECML/PKDD 2007 Discovery Challenge data set. Specifically, we show that our approach achieves higher precision and recall than previous methods. In addition, our approach has a number of desirable characteristics, including robustness to missing contextual information, interpretability of models, and scalability.« less
A database for coconut crop improvement.
Rajagopal, Velamoor; Manimekalai, Ramaswamy; Devakumar, Krishnamurthy; Rajesh; Karun, Anitha; Niral, Vittal; Gopal, Murali; Aziz, Shamina; Gunasekaran, Marimuthu; Kumar, Mundappurathe Ramesh; Chandrasekar, Arumugam
2005-12-08
Coconut crop improvement requires a number of biotechnology and bioinformatics tools. A database containing information on CG (coconut germplasm), CCI (coconut cultivar identification), CD (coconut disease), MIFSPC (microbial information systems in plantation crops) and VO (vegetable oils) is described. The database was developed using MySQL and PostgreSQL running in Linux operating system. The database interface is developed in PHP, HTML and JAVA. http://www.bioinfcpcri.org.
Dynamic measurement of speed of sound in n-Heptane by ultrasonics during fuel injections.
Minnetti, Elisa; Pandarese, Giuseppe; Evangelisti, Piersavio; Verdugo, Francisco Rodriguez; Ungaro, Carmine; Bastari, Alessandro; Paone, Nicola
2017-11-01
The paper presents a technique to measure the speed of sound in fuels based on pulse-echo ultrasound. The method is applied inside the test chamber of a Zeuch-type instrument used for indirect measurement of the injection rate (Mexus). The paper outlines the pulse-echo method, considering probe installation, ultrasound beam propagation inside the test chamber, typical signals obtained, as well as different processing algorithms. The method is validated in static conditions by comparing the experimental results to the NIST database both for water and n-Heptane. The ultrasonic system is synchronized to the injector so that time resolved samples of speed of sound can be successfully acquired during a series of injections. Results at different operating conditions in n-Heptane are shown. An uncertainty analysis supports the analysis of results and allows to validate the method. Experimental results show that the speed of sound variation during an injection event is less than 1%, so the Mexus model assumption to consider it constant during the injection is valid. Copyright © 2017 Elsevier B.V. All rights reserved.
Aguirre-Gamboa, Raul; Gomez-Rueda, Hugo; Martínez-Ledesma, Emmanuel; Martínez-Torteya, Antonio; Chacolla-Huaringa, Rafael; Rodriguez-Barrientos, Alberto; Tamez-Peña, José G; Treviño, Victor
2013-01-01
Validation of multi-gene biomarkers for clinical outcomes is one of the most important issues for cancer prognosis. An important source of information for virtual validation is the high number of available cancer datasets. Nevertheless, assessing the prognostic performance of a gene expression signature along datasets is a difficult task for Biologists and Physicians and also time-consuming for Statisticians and Bioinformaticians. Therefore, to facilitate performance comparisons and validations of survival biomarkers for cancer outcomes, we developed SurvExpress, a cancer-wide gene expression database with clinical outcomes and a web-based tool that provides survival analysis and risk assessment of cancer datasets. The main input of SurvExpress is only the biomarker gene list. We generated a cancer database collecting more than 20,000 samples and 130 datasets with censored clinical information covering tumors over 20 tissues. We implemented a web interface to perform biomarker validation and comparisons in this database, where a multivariate survival analysis can be accomplished in about one minute. We show the utility and simplicity of SurvExpress in two biomarker applications for breast and lung cancer. Compared to other tools, SurvExpress is the largest, most versatile, and quickest free tool available. SurvExpress web can be accessed in http://bioinformatica.mty.itesm.mx/SurvExpress (a tutorial is included). The website was implemented in JSP, JavaScript, MySQL, and R.
Aguirre-Gamboa, Raul; Gomez-Rueda, Hugo; Martínez-Ledesma, Emmanuel; Martínez-Torteya, Antonio; Chacolla-Huaringa, Rafael; Rodriguez-Barrientos, Alberto; Tamez-Peña, José G.; Treviño, Victor
2013-01-01
Validation of multi-gene biomarkers for clinical outcomes is one of the most important issues for cancer prognosis. An important source of information for virtual validation is the high number of available cancer datasets. Nevertheless, assessing the prognostic performance of a gene expression signature along datasets is a difficult task for Biologists and Physicians and also time-consuming for Statisticians and Bioinformaticians. Therefore, to facilitate performance comparisons and validations of survival biomarkers for cancer outcomes, we developed SurvExpress, a cancer-wide gene expression database with clinical outcomes and a web-based tool that provides survival analysis and risk assessment of cancer datasets. The main input of SurvExpress is only the biomarker gene list. We generated a cancer database collecting more than 20,000 samples and 130 datasets with censored clinical information covering tumors over 20 tissues. We implemented a web interface to perform biomarker validation and comparisons in this database, where a multivariate survival analysis can be accomplished in about one minute. We show the utility and simplicity of SurvExpress in two biomarker applications for breast and lung cancer. Compared to other tools, SurvExpress is the largest, most versatile, and quickest free tool available. SurvExpress web can be accessed in http://bioinformatica.mty.itesm.mx/SurvExpress (a tutorial is included). The website was implemented in JSP, JavaScript, MySQL, and R. PMID:24066126
ExplorEnz: a MySQL database of the IUBMB enzyme nomenclature
McDonald, Andrew G; Boyce, Sinéad; Moss, Gerard P; Dixon, Henry BF; Tipton, Keith F
2007-01-01
Background We describe the database ExplorEnz, which is the primary repository for EC numbers and enzyme data that are being curated on behalf of the IUBMB. The enzyme nomenclature is incorporated into many other resources, including the ExPASy-ENZYME, BRENDA and KEGG bioinformatics databases. Description The data, which are stored in a MySQL database, preserve the formatting of chemical and enzyme names. A simple, easy to use, web-based query interface is provided, along with an advanced search engine for more complex queries. The database is publicly available at . The data are available for download as SQL and XML files via FTP. Conclusion ExplorEnz has powerful and flexible search capabilities and provides the scientific community with the most up-to-date version of the IUBMB Enzyme List. PMID:17662133
ExplorEnz: a MySQL database of the IUBMB enzyme nomenclature.
McDonald, Andrew G; Boyce, Sinéad; Moss, Gerard P; Dixon, Henry B F; Tipton, Keith F
2007-07-27
We describe the database ExplorEnz, which is the primary repository for EC numbers and enzyme data that are being curated on behalf of the IUBMB. The enzyme nomenclature is incorporated into many other resources, including the ExPASy-ENZYME, BRENDA and KEGG bioinformatics databases. The data, which are stored in a MySQL database, preserve the formatting of chemical and enzyme names. A simple, easy to use, web-based query interface is provided, along with an advanced search engine for more complex queries. The database is publicly available at http://www.enzyme-database.org. The data are available for download as SQL and XML files via FTP. ExplorEnz has powerful and flexible search capabilities and provides the scientific community with the most up-to-date version of the IUBMB Enzyme List.
Evaluating the Cassandra NoSQL Database Approach for Genomic Data Persistency.
Aniceto, Rodrigo; Xavier, Rene; Guimarães, Valeria; Hondo, Fernanda; Holanda, Maristela; Walter, Maria Emilia; Lifschitz, Sérgio
2015-01-01
Rapid advances in high-throughput sequencing techniques have created interesting computational challenges in bioinformatics. One of them refers to management of massive amounts of data generated by automatic sequencers. We need to deal with the persistency of genomic data, particularly storing and analyzing these large-scale processed data. To find an alternative to the frequently considered relational database model becomes a compelling task. Other data models may be more effective when dealing with a very large amount of nonconventional data, especially for writing and retrieving operations. In this paper, we discuss the Cassandra NoSQL database approach for storing genomic data. We perform an analysis of persistency and I/O operations with real data, using the Cassandra database system. We also compare the results obtained with a classical relational database system and another NoSQL database approach, MongoDB.
Engineering the ATLAS TAG Browser
NASA Astrophysics Data System (ADS)
Zhang, Qizhi; ATLAS Collaboration
2011-12-01
ELSSI is a web-based event metadata (TAG) browser and event-level selection service for ATLAS. In this paper, we describe some of the challenges encountered in the process of developing ELSSI, and the software engineering strategies adopted to address those challenges. Approaches to management of access to data, browsing, data rendering, query building, query validation, execution, connection management, and communication with auxiliary services are discussed. We also describe strategies for dealing with data that may vary over time, such as run-dependent trigger decision decoding. Along with examples, we illustrate how programming techniques in multiple languages (PHP, JAVASCRIPT, XML, AJAX, and PL/SQL) have been blended to achieve the required results. Finally, we evaluate features of the ELSSI service in terms of functionality, scalability, and performance.
Study of the method of water-injected meat identifying based on low-field nuclear magnetic resonance
NASA Astrophysics Data System (ADS)
Xu, Jianmei; Lin, Qing; Yang, Fang; Zheng, Zheng; Ai, Zhujun
2018-01-01
The aim of this study to apply low-field nuclear magnetic resonance technique was to study regular variation of the transverse relaxation spectral parameters of water-injected meat with the proportion of water injection. Based on this, the method of one-way ANOVA and discriminant analysis was used to analyse the differences between these parameters in the capacity of distinguishing water-injected proportion, and established a model for identifying water-injected meat. The results show that, except for T 21b, T 22e and T 23b, the other parameters of the T 2 relaxation spectrum changed regularly with the change of water-injected proportion. The ability of different parameters to distinguish water-injected proportion was different. Based on S, P 22 and T 23m as the prediction variable, the Fisher model and the Bayes model were established by discriminant analysis method, qualitative and quantitative classification of water-injected meat can be realized. The rate of correct discrimination of distinguished validation and cross validation were 88%, the model was stable.
Comparisons of Attacks on Honeypots With Those on Real Networks
2006-03-01
Oracle , MySQL , or PostgreSQL. Figure 2 shows an incoming packet and the process involved before and after the Snort engine detects the suspicious...stored on a separate, secured system.”[2]. Honeypots have several other uses besides monitoring attackers. They serve to protect real networks and...interaction vs . high-interaction. Although, both low-interaction and high-interaction honeypots are effective in soliciting attacks, high-interaction
Network Security Visualization
1999-09-27
performing SQL generation and result-set binding, inserting acquired security events into the database and gathering the requested data for Console scene...objects is also auto-generated by a VBA script. Built into the auto-generated table access objects are the preferred join paths between tables. This...much of the Server itself) never have to deal with SQL directly. This is one aspect of laying the groundwork for supporting RDBMSs from multiple vendors
A Data Warehouse to Support Condition Based Maintenance (CBM)
2005-05-01
Application ( VBA ) code sequence to import the original MAST-generated CSV and then create a single output table in DBASE IV format. The DBASE IV format...database architecture (Oracle, Sybase, MS- SQL , etc). This design includes table definitions, comments, specification of table attributes, primary and foreign...built queries and applications. Needs the application developers to construct data views. No SQL programming experience. b. Power Database User - knows
2016-03-01
science IT information technology JBOD just a bunch of disks JDBC java database connectivity xviii JPME Joint Professional Military Education JSO...Joint Service Officer JVM java virtual machine MPP massively parallel processing MPTE Manpower, Personnel, Training, and Education NAVMAC Navy...27 external database, whether it is MySQL , Oracle, DB2, or SQL Server (Teller, 2015). Connectors optimize the data transfer by obtaining metadata
A database for coconut crop improvement
Rajagopal, Velamoor; Manimekalai, Ramaswamy; Devakumar, Krishnamurthy; Rajesh; Karun, Anitha; Niral, Vittal; Gopal, Murali; Aziz, Shamina; Gunasekaran, Marimuthu; Kumar, Mundappurathe Ramesh; Chandrasekar, Arumugam
2005-01-01
Coconut crop improvement requires a number of biotechnology and bioinformatics tools. A database containing information on CG (coconut germplasm), CCI (coconut cultivar identification), CD (coconut disease), MIFSPC (microbial information systems in plantation crops) and VO (vegetable oils) is described. The database was developed using MySQL and PostgreSQL running in Linux operating system. The database interface is developed in PHP, HTML and JAVA. Availability http://www.bioinfcpcri.org PMID:17597858
Quality Attribute-Guided Evaluation of NoSQL Databases: An Experience Report
2014-10-18
detailed technical evaluations of NoSQL databases specifically, and big data systems in general, that have become apparent during our study... big data , software systems [Agarwal 2011]. Internet-born organizations such as Google and Amazon are at the cutting edge of this revolution...Chang 2008], along with those of numerous other big data innovators, have made a variety of open source and commercial data management technologies
Alignment of high-throughput sequencing data inside in-memory databases.
Firnkorn, Daniel; Knaup-Gregori, Petra; Lorenzo Bermejo, Justo; Ganzinger, Matthias
2014-01-01
In times of high-throughput DNA sequencing techniques, performance-capable analysis of DNA sequences is of high importance. Computer supported DNA analysis is still an intensive time-consuming task. In this paper we explore the potential of a new In-Memory database technology by using SAP's High Performance Analytic Appliance (HANA). We focus on read alignment as one of the first steps in DNA sequence analysis. In particular, we examined the widely used Burrows-Wheeler Aligner (BWA) and implemented stored procedures in both, HANA and the free database system MySQL, to compare execution time and memory management. To ensure that the results are comparable, MySQL has been running in memory as well, utilizing its integrated memory engine for database table creation. We implemented stored procedures, containing exact and inexact searching of DNA reads within the reference genome GRCh37. Due to technical restrictions in SAP HANA concerning recursion, the inexact matching problem could not be implemented on this platform. Hence, performance analysis between HANA and MySQL was made by comparing the execution time of the exact search procedures. Here, HANA was approximately 27 times faster than MySQL which means, that there is a high potential within the new In-Memory concepts, leading to further developments of DNA analysis procedures in the future.
Chevat, Catherine; Viala-Danten, Muriel; Dias-Barbosa, Carla; Nguyen, Van Hung
2009-01-01
Background Influenza is among the most common infectious diseases. The main protection against influenza is vaccination. A self-administered questionnaire was developed and validated for use in clinical trials to assess subjects' perception and acceptance of influenza vaccination and its subsequent injection site reactions (ISR). Methods The VAPI questionnaire was developed based on interviews with vaccinees. The initial version was administered to subjects in international clinical trials comparing intradermal with intramuscular influenza vaccination. Item reduction and scale construction were carried out using principal component and multitrait analyses (n = 549). Psychometric validation of the final version was conducted per country (n = 5,543) and included construct and clinical validity and internal consistency reliability. All subjects gave their written informed consent before being interviewed or included in the clinical studies. Results The final questionnaire comprised 4 dimensions ("bother from ISR"; "arm movement"; "sleep"; "acceptability") grouping 16 items, and 5 individual items (anxiety before vaccination; bother from pain during vaccination; satisfaction with injection system; willingness to be vaccinated next year; anxiety about vaccination next year). Construct validity was confirmed for all scales in most of the countries. Internal consistency reliability was good for all versions (Cronbach's alpha ranging from 0.68 to 0.94), as was clinical validity: scores were positively correlated with the severity of ISR and pain. Conclusion The VAPI questionnaire is a valid and reliable tool, assessing the acceptance of vaccine injection and reactions following vaccination. Trial registration NCT00258934, NCT00383526, NCT00383539. PMID:19261173
Domańska, Barbara; Stumpp, Oliver; Poon, Steven; Oray, Serkan; Mountian, Irina; Pichon, Clovis
2018-01-01
We incorporated patient feedback from human factors studies (HFS) in the patient-centric design and validation of ava ® , an electromechanical device (e-Device) for self-injecting the anti-tumor necrosis factor certolizumab pegol (CZP). Healthcare professionals, caregivers, healthy volunteers, and patients with rheumatoid arthritis, psoriatic arthritis, ankylosing spondylitis, or Crohn's disease participated in 11 formative HFS to optimize the e-Device design through intended user feedback; nine studies involved simulated injections. Formative participant questionnaire feedback was collected following e-Device prototype handling. Validation HFS (one EU study and one US study) assessed the safe and effective setup and use of the e-Device using 22 predefined critical tasks. Task outcomes were categorized as "failures" if participants did not succeed within three attempts. Two hundred eighty-three participants entered formative (163) and validation (120) HFS; 260 participants performed one or more simulated e-Device self-injections. Design changes following formative HFS included alterations to buttons and the graphical user interface screen. All validation HFS participants completed critical tasks necessary for CZP dose delivery, with minimal critical task failures (12 of 572 critical tasks, 2.1%, in the EU study, and 2 of 5310 critical tasks, less than 0.1%, in the US study). CZP e-Device development was guided by intended user feedback through HFS, ensuring the final design addressed patients' needs. In both validation studies, participants successfully performed all critical tasks, demonstrating safe and effective e-Device self-injections. UCB Pharma. Plain language summary available on the journal website.
Factors Leading to Effectiveness and Satisfaction in Civil Engineer Information Systems
2008-03-01
recently acquired MySQL in 2008 shortly after Oracle failed to acquire MySQL in 2007. For more information on policy implications concerning the use...individual level serves as the pertinent outcome variable and is used to evaluate and compare information systems in this study. Researchers have found...interim work information management system used by the Civil Engineer Operations Flight. The functions served by this system date back to the late
2004-06-01
remote databases, has seen little vendor acceptance. Each database ( Oracle , DB2, MySQL , etc.) has its own client- server protocol. Therefore each...existing standards – SQL , X.500/LDAP, FTP, etc. • View information dissemination as selective replication – State-oriented vs . message-oriented...allowing the 8 application to start. The resource management system would serve as a broker to the resources, making sure that resources are not
2003-03-01
PXSLServlet Paul A. Open Source Relational x X 23 Tchistopolskii sql2dtd David Mertz Public domain Relational x -- sql2xml Scott Hathaway Public...March 2003. [Hunter 2001] Hunter, David ; Cagle, Kurt; Dix, Chris; Kovack, Roger; Pinnock, Jonathan, Rafter, Jeff; Beginning XML (2nd Edition...Postgraduate School Monterey, California 4. Curt Blais Naval Postgraduate School Monterey, California 5 Erik Chaum NAVSEA Undersea
Large-scale Graph Computation on Just a PC
2014-05-01
edges for several vertices simultaneously). We compared the performance of GraphChi-DB to Neo4j using their Java API (we discuss MySQL comparison in the...75 4.7.6 Comparison to RDBMS ( MySQL ) . . . . . . . . . . . . . . . . . . . . . 75 4.7.7 Summary of the...Windows method, GraphChi. The C++ implementation has circa 8,000 lines of code. We have also de- veloped a Java -version of GraphChi, but it does not
Delayed Instantiation Bulk Operations for Management of Distributed, Object-Based Storage Systems
2009-08-01
source and destination object sets, while they have attribute pages to indicate that history . Fourth, we allow for operations to occur on any objects...client dialogue to the PostgreSQL database where server-side functions implement the service logic for the requests. The translation is done...to satisfy client requests, and performs delayed instantiation bulk operations. It is built around a PostgreSQL database with tables for storing
Wong, Wing Chung; Kim, Dewey; Carter, Hannah; Diekhans, Mark; Ryan, Michael C; Karchin, Rachel
2011-08-01
Thousands of cancer exomes are currently being sequenced, yielding millions of non-synonymous single nucleotide variants (SNVs) of possible relevance to disease etiology. Here, we provide a software toolkit to prioritize SNVs based on their predicted contribution to tumorigenesis. It includes a database of precomputed, predictive features covering all positions in the annotated human exome and can be used either stand-alone or as part of a larger variant discovery pipeline. MySQL database, source code and binaries freely available for academic/government use at http://wiki.chasmsoftware.org, Source in Python and C++. Requires 32 or 64-bit Linux system (tested on Fedora Core 8,10,11 and Ubuntu 10), 2.5*≤ Python <3.0*, MySQL server >5.0, 60 GB available hard disk space (50 MB for software and data files, 40 GB for MySQL database dump when uncompressed), 2 GB of RAM.
Motivational and metacognitive feedback in SQL-Tutor*
NASA Astrophysics Data System (ADS)
Hull, Alison; du Boulay, Benedict
2015-04-01
Motivation and metacognition are strongly intertwined, with learners high in self-efficacy more likely to use a variety of self-regulatory learning strategies, as well as to persist longer on challenging tasks. The aim of the research was to improve the learner's focus on the process and experience of problem-solving while using an Intelligent Tutoring System (ITS) and including motivational and metacognitive feedback based on the learner's past states and experiences. An existing ITS, SQL-Tutor, was used with first-year undergraduates studying a database module. The study used two versions of SQL-Tutor: the Control group used a base version providing domain feedback and the Study group used an extended version that also provided motivational and metacognitive feedback. This paper summarises the pre- and post-process results. Comparisons between groups showed some differing trends both in learning outcomes and behaviour in favour of the Study group.
0.75 atoms improve the clock signal of 10,000 atoms
NASA Astrophysics Data System (ADS)
Kruse, I.; Lange, K.; Peise, J.; Lücke, B.; Pezzè, L.; Arlt, J.; Ertmer, W.; Lisdat, C.; Santos, L.; Smerzi, A.; Klempt, C.
2017-02-01
Since the pioneering work of Ramsey, atom interferometers are employed for precision metrology, in particular to measure time and to realize the second. In a classical interferometer, an ensemble of atoms is prepared in one of the two input states, whereas the second one is left empty. In this case, the vacuum noise restricts the precision of the interferometer to the standard quantum limit (SQL). Here, we propose and experimentally demonstrate a novel clock configuration that surpasses the SQL by squeezing the vacuum in the empty input state. We create a squeezed vacuum state containing an average of 0.75 atoms to improve the clock sensitivity of 10,000 atoms by 2.05 dB. The SQL poses a significant limitation for today's microwave fountain clocks, which serve as the main time reference. We evaluate the major technical limitations and challenges for devising a next generation of fountain clocks based on atomic squeezed vacuum.
Improvement of an Atomic Clock using Squeezed Vacuum
NASA Astrophysics Data System (ADS)
Kruse, I.; Lange, K.; Peise, J.; Lücke, B.; Pezzè, L.; Arlt, J.; Ertmer, W.; Lisdat, C.; Santos, L.; Smerzi, A.; Klempt, C.
2016-09-01
Since the pioneering work of Ramsey, atom interferometers are employed for precision metrology, in particular to measure time and to realize the second. In a classical interferometer, an ensemble of atoms is prepared in one of the two input states, whereas the second one is left empty. In this case, the vacuum noise restricts the precision of the interferometer to the standard quantum limit (SQL). Here, we propose and experimentally demonstrate a novel clock configuration that surpasses the SQL by squeezing the vacuum in the empty input state. We create a squeezed vacuum state containing an average of 0.75 atoms to improve the clock sensitivity of 10000 atoms by 2.05-0.37 +0 .34 dB . The SQL poses a significant limitation for today's microwave fountain clocks, which serve as the main time reference. We evaluate the major technical limitations and challenges for devising a next generation of fountain clocks based on atomic squeezed vacuum.
Evaluating the Cassandra NoSQL Database Approach for Genomic Data Persistency
Aniceto, Rodrigo; Xavier, Rene; Guimarães, Valeria; Hondo, Fernanda; Holanda, Maristela; Walter, Maria Emilia; Lifschitz, Sérgio
2015-01-01
Rapid advances in high-throughput sequencing techniques have created interesting computational challenges in bioinformatics. One of them refers to management of massive amounts of data generated by automatic sequencers. We need to deal with the persistency of genomic data, particularly storing and analyzing these large-scale processed data. To find an alternative to the frequently considered relational database model becomes a compelling task. Other data models may be more effective when dealing with a very large amount of nonconventional data, especially for writing and retrieving operations. In this paper, we discuss the Cassandra NoSQL database approach for storing genomic data. We perform an analysis of persistency and I/O operations with real data, using the Cassandra database system. We also compare the results obtained with a classical relational database system and another NoSQL database approach, MongoDB. PMID:26558254
Managing Complex Interoperability Solutions using Model-Driven Architecture
2011-06-01
such as Oracle or MySQL . Each data model for a specific RDBMS is a distinct PSM. Or the system may want to exchange information with other C2...reduced number of transformations, e.g., from an RDBMS physical schema to the corresponding SQL script needed to instantiate the tables in a relational...tance of models. In engineering, a model serves several purposes: 1. It presents an abstract view of a complex system or of a complex information
Information Collection using Handheld Devices in Unreliable Networking Environments
2014-06-01
different types of mobile devices that connect wirelessly to a database 8 server. The actual backend database is not important to the mobile clients...Google’s infrastructure and local servers with MySQL and PostgreSQL on the backend (ODK 2014b). (2) Google Fusion Tables are used to do basic link...how we conduct business. Our requirements to share information do not change simply because there is little or no existing infrastructure in our
2016-03-01
Representational state transfer Java messaging service Java application programming interface (API) Internet relay chat (IRC)/extensible messaging and...JBoss application server or an Apache Tomcat servlet container instance. The relational database management system can be either PostgreSQL or MySQL ... Java library called direct web remoting. This library has been part of the core CACE architecture for quite some time; however, there have not been
MicroRNA Gene Regulatory Networks in Peripheral Nerve Sheath Tumors
2013-09-01
3.0 hierarchical clustering of both the X and the Y-axis using Centroid linkage. The resulting clustered matrixes were visualized using Java Treeview...To score potential ceRNA interactions, the 54979 human interactions were loaded into a mySQL database and when the user selects a given mRNA all...on the fly using PHP interactions with mySQL in a similar fashion as previously described in our publicly available databases such as sarcoma
Understanding and Capturing People’s Mobile App Privacy Preferences
2013-10-28
The entire apps’ metadata takes up about 500MB of storage space when stored in a MySQL database and all the binary files take approximately 300GB of...functionality that can de- compile Dalvik bytecodes to Java source code faster than other de-compilers. Given the scale of the app analysis we planned on... java libraries, such as parser, sql connectors, etc Targeted Ads 137 admob, adwhirl, greystripe… Provided by mobile behavioral ads company to
Shark: SQL and Analytics with Cost-Based Query Optimization on Coarse-Grained Distributed Memory
2014-01-13
RDBMS and contains a database (often MySQL or Derby) with a namespace for tables, table metadata and partition information. Table data is stored in an...serialization/deserialization) Java interface implementations with corresponding object inspectors. The Hive driver controls the processing of queries, coordinat...native API, RDD operations are invoked through a functional interface similar to DryadLINQ [32] in Scala, Java or Python. For example, the Scala code for
Implementing an Intrusion Detection System in the Mysea Architecture
2008-06-01
password for each user passwd <username> then follow the prompts 2. PostgreSQL 7.4.18 Installation Perform the following steps as root: 1. Copy...password changed Repeat for user snort. exit After making the groups and users the group and passwd file needs to be updated. Set security and...untrusted/bin/xtsmkgroup > /etc/group chmod 644 /etc/group /xts/untrusted/bin/xtsmkpasswd > /etc/ passwd chmod 644 /etc/ passwd 3. PostgreSQL 7.4.18
Astronomical Archive at Tartu Observatory
NASA Astrophysics Data System (ADS)
Annuk, K.
2007-10-01
Archiving astronomical data is important task not only at large observatories but also at small observatories. Here we describe the astronomical archive at Tartu Observatory. The archive consists of old photographic plate images, photographic spectrograms, CCD direct--images and CCD spectroscopic data. The photographic plate digitizing project was started in 2005. An on-line database (based on MySQL) was created. The database includes CCD data as well photographic data. A PHP-MySQL interface was written for access to all data.
Shark: SQL and Rich Analytics at Scale
2012-11-26
learning programs up to 100 faster than Hadoop. Unlike previous systems, Shark shows that it is possible to achieve these speedups while retaining a...Shark to run SQL queries up to 100× faster than Apache Hive, and machine learning programs up to 100× faster than Hadoop. Unlike previous systems, Shark...so using a runtime that is optimized for such workloads and a programming model that is designed to express machine learn - ing algorithms. 4.1
Nosql for Storage and Retrieval of Large LIDAR Data Collections
NASA Astrophysics Data System (ADS)
Boehm, J.; Liu, K.
2015-08-01
Developments in LiDAR technology over the past decades have made LiDAR to become a mature and widely accepted source of geospatial information. This in turn has led to an enormous growth in data volume. The central idea for a file-centric storage of LiDAR point clouds is the observation that large collections of LiDAR data are typically delivered as large collections of files, rather than single files of terabyte size. This split of the dataset, commonly referred to as tiling, was usually done to accommodate a specific processing pipeline. It makes therefore sense to preserve this split. A document oriented NoSQL database can easily emulate this data partitioning, by representing each tile (file) in a separate document. The document stores the metadata of the tile. The actual files are stored in a distributed file system emulated by the NoSQL database. We demonstrate the use of MongoDB a highly scalable document oriented NoSQL database for storing large LiDAR files. MongoDB like any NoSQL database allows for queries on the attributes of the document. As a specialty MongoDB also allows spatial queries. Hence we can perform spatial queries on the bounding boxes of the LiDAR tiles. Inserting and retrieving files on a cloud-based database is compared to native file system and cloud storage transfer speed.
Design and Implementation of Website Information Disclosure Assessment System
Cho, Ying-Chiang; Pan, Jen-Yi
2015-01-01
Internet application technologies, such as cloud computing and cloud storage, have increasingly changed people’s lives. Websites contain vast amounts of personal privacy information. In order to protect this information, network security technologies, such as database protection and data encryption, attract many researchers. The most serious problems concerning web vulnerability are e-mail address and network database leakages. These leakages have many causes. For example, malicious users can steal database contents, taking advantage of mistakes made by programmers and administrators. In order to mitigate this type of abuse, a website information disclosure assessment system is proposed in this study. This system utilizes a series of technologies, such as web crawler algorithms, SQL injection attack detection, and web vulnerability mining, to assess a website’s information disclosure. Thirty websites, randomly sampled from the top 50 world colleges, were used to collect leakage information. This testing showed the importance of increasing the security and privacy of website information for academic websites. PMID:25768434
Martinez, Luc; Consoli, Silla M; Monnier, Louis; Simon, Dominique; Wong, Olivier; Yomtov, Bernard; Guéron, Béatrice; Benmedjahed, Khadra; Guillemin, Isabelle; Arnould, Benoit
2007-01-01
Background Although insulin therapy is well-accepted by symptomatic diabetic patients, it is still often delayed in less severe patients, in whom injectable insulin remains under-used. A better understanding of patients' perception of insulin would eventually help physicians to adopt the most appropriate dialogue when having to motivate patients to initiate or to intensify insulin injection. Methods The 'Studying the Hurdles of Insulin Prescription' (SHIP) questionnaire was developed based on a list of concepts derived from three diabetic patients' focus groups, and was included into two cross-sectional studies with similar design: SHIP Oral study and SHIP Premix study. Diabetic patients treated with oral hypoglycaemic agents (OHA; n = 1,494) and patients already treated with insulin (n = 1,150) completed the questionnaire at baseline, 6- and 12 months. Psychometric properties were assessed: 1) structure analysis by Principal Component Analysis (PCA) with Varimax rotation, 2) internal consistency reliability (Cronbach's alpha), and 3) concurrent validity (Spearman correlation coefficients with the Fear of Self-Injecting (FSI) score of the Diabetes Fear of Injecting and Self-testing Questionnaire. Reluctance/motivation towards insulin was assessed. Scores' ability to predict patients' insulin injection reluctance/motivation and initiation/intensification was evaluated with the Area Under the Receiver Operating Characteristic (ROC) Curve (AUC). Results PCA analysis confirmed the structure of the 14 items grouped into 3 dimensions: 'acceptance and motivation', 'fear and constraints', and 'restraints and barriers' towards insulin injection. Internal consistency reliability was excellent (Cronbach's alpha > 0.70); concurrent validity was good. The three scores were significantly predictive of patients' reluctance/motivation towards insulin injection initiation, as they were of patients' actual switch, except for the 'restraints and barriers' dimension. 'Acceptance and motivation' and 'fears and constraints' dimensions were also significantly predictive of patients' reluctance/motivation towards insulin intensification. By the end of the 12-month study, 179 of the initially OHA-treated patients had started insulin injections; 186 of the patients already treated with insulin had increased their injections. Conclusion The SHIP questionnaire provides reliable and valid assessment of diabetic patients' attitude towards insulin and injections. The predictive power of scores for patients' reluctance/motivation and actual treatment decisions demonstrates encouraging potential for further application in clinical practice. PMID:17727695
Web-based application on employee performance assessment using exponential comparison method
NASA Astrophysics Data System (ADS)
Maryana, S.; Kurnia, E.; Ruyani, A.
2017-02-01
Employee performance assessment is also called a performance review, performance evaluation, or assessment of employees, is an effort to assess the achievements of staffing performance with the aim to increase productivity of employees and companies. This application helps in the assessment of employee performance using five criteria: Presence, Quality of Work, Quantity of Work, Discipline, and Teamwork. The system uses the Exponential Comparative Method and Weighting Eckenrode. Calculation results using graphs were provided to see the assessment of each employee. Programming language used in this system is written in Notepad++ and MySQL database. The testing result on the system can be concluded that this application is correspond with the design and running properly. The test conducted is structural test, functional test, and validation, sensitivity analysis, and SUMI testing.
Freire, Sergio Miranda; Teodoro, Douglas; Wei-Kleiner, Fang; Sundvall, Erik; Karlsson, Daniel; Lambrix, Patrick
2016-01-01
This study provides an experimental performance evaluation on population-based queries of NoSQL databases storing archetype-based Electronic Health Record (EHR) data. There are few published studies regarding the performance of persistence mechanisms for systems that use multilevel modelling approaches, especially when the focus is on population-based queries. A healthcare dataset with 4.2 million records stored in a relational database (MySQL) was used to generate XML and JSON documents based on the openEHR reference model. Six datasets with different sizes were created from these documents and imported into three single machine XML databases (BaseX, eXistdb and Berkeley DB XML) and into a distributed NoSQL database system based on the MapReduce approach, Couchbase, deployed in different cluster configurations of 1, 2, 4, 8 and 12 machines. Population-based queries were submitted to those databases and to the original relational database. Database size and query response times are presented. The XML databases were considerably slower and required much more space than Couchbase. Overall, Couchbase had better response times than MySQL, especially for larger datasets. However, Couchbase requires indexing for each differently formulated query and the indexing time increases with the size of the datasets. The performances of the clusters with 2, 4, 8 and 12 nodes were not better than the single node cluster in relation to the query response time, but the indexing time was reduced proportionally to the number of nodes. The tested XML databases had acceptable performance for openEHR-based data in some querying use cases and small datasets, but were generally much slower than Couchbase. Couchbase also outperformed the response times of the relational database, but required more disk space and had a much longer indexing time. Systems like Couchbase are thus interesting research targets for scalable storage and querying of archetype-based EHR data when population-based use cases are of interest. PMID:26958859
Freire, Sergio Miranda; Teodoro, Douglas; Wei-Kleiner, Fang; Sundvall, Erik; Karlsson, Daniel; Lambrix, Patrick
2016-01-01
This study provides an experimental performance evaluation on population-based queries of NoSQL databases storing archetype-based Electronic Health Record (EHR) data. There are few published studies regarding the performance of persistence mechanisms for systems that use multilevel modelling approaches, especially when the focus is on population-based queries. A healthcare dataset with 4.2 million records stored in a relational database (MySQL) was used to generate XML and JSON documents based on the openEHR reference model. Six datasets with different sizes were created from these documents and imported into three single machine XML databases (BaseX, eXistdb and Berkeley DB XML) and into a distributed NoSQL database system based on the MapReduce approach, Couchbase, deployed in different cluster configurations of 1, 2, 4, 8 and 12 machines. Population-based queries were submitted to those databases and to the original relational database. Database size and query response times are presented. The XML databases were considerably slower and required much more space than Couchbase. Overall, Couchbase had better response times than MySQL, especially for larger datasets. However, Couchbase requires indexing for each differently formulated query and the indexing time increases with the size of the datasets. The performances of the clusters with 2, 4, 8 and 12 nodes were not better than the single node cluster in relation to the query response time, but the indexing time was reduced proportionally to the number of nodes. The tested XML databases had acceptable performance for openEHR-based data in some querying use cases and small datasets, but were generally much slower than Couchbase. Couchbase also outperformed the response times of the relational database, but required more disk space and had a much longer indexing time. Systems like Couchbase are thus interesting research targets for scalable storage and querying of archetype-based EHR data when population-based use cases are of interest.
CGDM: collaborative genomic data model for molecular profiling data using NoSQL.
Wang, Shicai; Mares, Mihaela A; Guo, Yi-Ke
2016-12-01
High-throughput molecular profiling has greatly improved patient stratification and mechanistic understanding of diseases. With the increasing amount of data used in translational medicine studies in recent years, there is a need to improve the performance of data warehouses in terms of data retrieval and statistical processing. Both relational and Key Value models have been used for managing molecular profiling data. Key Value models such as SeqWare have been shown to be particularly advantageous in terms of query processing speed for large datasets. However, more improvement can be achieved, particularly through better indexing techniques of the Key Value models, taking advantage of the types of queries which are specific for the high-throughput molecular profiling data. In this article, we introduce a Collaborative Genomic Data Model (CGDM), aimed at significantly increasing the query processing speed for the main classes of queries on genomic databases. CGDM creates three Collaborative Global Clustering Index Tables (CGCITs) to solve the velocity and variety issues at the cost of limited extra volume. Several benchmarking experiments were carried out, comparing CGDM implemented on HBase to the traditional SQL data model (TDM) implemented on both HBase and MySQL Cluster, using large publicly available molecular profiling datasets taken from NCBI and HapMap. In the microarray case, CGDM on HBase performed up to 246 times faster than TDM on HBase and 7 times faster than TDM on MySQL Cluster. In single nucleotide polymorphism case, CGDM on HBase outperformed TDM on HBase by up to 351 times and TDM on MySQL Cluster by up to 9 times. The CGDM source code is available at https://github.com/evanswang/CGDM. y.guo@imperial.ac.uk. © The Author 2016. Published by Oxford University Press. All rights reserved. For Permissions, please e-mail: journals.permissions@oup.com.
High dimensional biological data retrieval optimization with NoSQL technology.
Wang, Shicai; Pandis, Ioannis; Wu, Chao; He, Sijin; Johnson, David; Emam, Ibrahim; Guitton, Florian; Guo, Yike
2014-01-01
High-throughput transcriptomic data generated by microarray experiments is the most abundant and frequently stored kind of data currently used in translational medicine studies. Although microarray data is supported in data warehouses such as tranSMART, when querying relational databases for hundreds of different patient gene expression records queries are slow due to poor performance. Non-relational data models, such as the key-value model implemented in NoSQL databases, hold promise to be more performant solutions. Our motivation is to improve the performance of the tranSMART data warehouse with a view to supporting Next Generation Sequencing data. In this paper we introduce a new data model better suited for high-dimensional data storage and querying, optimized for database scalability and performance. We have designed a key-value pair data model to support faster queries over large-scale microarray data and implemented the model using HBase, an implementation of Google's BigTable storage system. An experimental performance comparison was carried out against the traditional relational data model implemented in both MySQL Cluster and MongoDB, using a large publicly available transcriptomic data set taken from NCBI GEO concerning Multiple Myeloma. Our new key-value data model implemented on HBase exhibits an average 5.24-fold increase in high-dimensional biological data query performance compared to the relational model implemented on MySQL Cluster, and an average 6.47-fold increase on query performance on MongoDB. The performance evaluation found that the new key-value data model, in particular its implementation in HBase, outperforms the relational model currently implemented in tranSMART. We propose that NoSQL technology holds great promise for large-scale data management, in particular for high-dimensional biological data such as that demonstrated in the performance evaluation described in this paper. We aim to use this new data model as a basis for migrating tranSMART's implementation to a more scalable solution for Big Data.
High dimensional biological data retrieval optimization with NoSQL technology
2014-01-01
Background High-throughput transcriptomic data generated by microarray experiments is the most abundant and frequently stored kind of data currently used in translational medicine studies. Although microarray data is supported in data warehouses such as tranSMART, when querying relational databases for hundreds of different patient gene expression records queries are slow due to poor performance. Non-relational data models, such as the key-value model implemented in NoSQL databases, hold promise to be more performant solutions. Our motivation is to improve the performance of the tranSMART data warehouse with a view to supporting Next Generation Sequencing data. Results In this paper we introduce a new data model better suited for high-dimensional data storage and querying, optimized for database scalability and performance. We have designed a key-value pair data model to support faster queries over large-scale microarray data and implemented the model using HBase, an implementation of Google's BigTable storage system. An experimental performance comparison was carried out against the traditional relational data model implemented in both MySQL Cluster and MongoDB, using a large publicly available transcriptomic data set taken from NCBI GEO concerning Multiple Myeloma. Our new key-value data model implemented on HBase exhibits an average 5.24-fold increase in high-dimensional biological data query performance compared to the relational model implemented on MySQL Cluster, and an average 6.47-fold increase on query performance on MongoDB. Conclusions The performance evaluation found that the new key-value data model, in particular its implementation in HBase, outperforms the relational model currently implemented in tranSMART. We propose that NoSQL technology holds great promise for large-scale data management, in particular for high-dimensional biological data such as that demonstrated in the performance evaluation described in this paper. We aim to use this new data model as a basis for migrating tranSMART's implementation to a more scalable solution for Big Data. PMID:25435347
NASA Astrophysics Data System (ADS)
Knosp, B.; Gangl, M.; Hristova-Veleva, S. M.; Kim, R. M.; Li, P.; Turk, J.; Vu, Q. A.
2015-12-01
The JPL Tropical Cyclone Information System (TCIS) brings together satellite, aircraft, and model forecast data from several NASA, NOAA, and other data centers to assist researchers in comparing and analyzing data and model forecast related to tropical cyclones. The TCIS has been running a near-real time (NRT) data portal during North Atlantic hurricane season that typically runs from June through October each year, since 2010. Data collected by the TCIS varies by type, format, contents, and frequency and is served to the user in two ways: (1) as image overlays on a virtual globe and (2) as derived output from a suite of analysis tools. In order to support these two functions, the data must be collected and then made searchable by criteria such as date, mission, product, pressure level, and geospatial region. Creating a database architecture that is flexible enough to manage, intelligently interrogate, and ultimately present this disparate data to the user in a meaningful way has been the primary challenge. The database solution for the TCIS has been to use a hybrid MySQL + Solr implementation. After testing other relational database and NoSQL solutions, such as PostgreSQL and MongoDB respectively, this solution has given the TCIS the best offerings in terms of query speed and result reliability. This database solution also supports the challenging (and memory overwhelming) geospatial queries that are necessary to support analysis tools requested by users. Though hardly new technologies on their own, our implementation of MySQL + Solr had to be customized and tuned to be able to accurately store, index, and search the TCIS data holdings. In this presentation, we will discuss how we arrived on our MySQL + Solr database architecture, why it offers us the most consistent fast and reliable results, and how it supports our front end so that we can offer users a look into our "big data" holdings.
NASA Astrophysics Data System (ADS)
Velazquez, Enrique Israel
Improvements in medical and genomic technologies have dramatically increased the production of electronic data over the last decade. As a result, data management is rapidly becoming a major determinant, and urgent challenge, for the development of Precision Medicine. Although successful data management is achievable using Relational Database Management Systems (RDBMS), exponential data growth is a significant contributor to failure scenarios. Growing amounts of data can also be observed in other sectors, such as economics and business, which, together with the previous facts, suggests that alternate database approaches (NoSQL) may soon be required for efficient storage and management of big databases. However, this hypothesis has been difficult to test in the Precision Medicine field since alternate database architectures are complex to assess and means to integrate heterogeneous electronic health records (EHR) with dynamic genomic data are not easily available. In this dissertation, we present a novel set of experiments for identifying NoSQL database approaches that enable effective data storage and management in Precision Medicine using patients' clinical and genomic information from the cancer genome atlas (TCGA). The first experiment draws on performance and scalability from biologically meaningful queries with differing complexity and database sizes. The second experiment measures performance and scalability in database updates without schema changes. The third experiment assesses performance and scalability in database updates with schema modifications due dynamic data. We have identified two NoSQL approach, based on Cassandra and Redis, which seems to be the ideal database management systems for our precision medicine queries in terms of performance and scalability. We present NoSQL approaches and show how they can be used to manage clinical and genomic big data. Our research is relevant to the public health since we are focusing on one of the main challenges to the development of Precision Medicine and, consequently, investigating a potential solution to the progressively increasing demands on health care.
NASA Astrophysics Data System (ADS)
Yamanishi, Manabu
A combined experimental and computational investigation was performed in order to evaluate the effects of various design parameters of an in-line injection pump on the nozzle exit characteristics for DI diesel engines. Measurements of the pump chamber pressure and the delivery valve lift were included for validation by using specially designed transducers installed inside the pump. The results confirm that the simulation model is capable of predicting the pump operation for all the different designs investigated pump operating conditions. Following the successful validation of this model, parametric studies were performed which allow for improved fuel injection system design.
2014-06-01
central location. Each of the SQLite databases are converted and stored in one MySQL database and the pcap files are parsed to extract call information...from the specific communications applications used during the experiment. This extracted data is then stored in the same MySQL database. With all...rhythm of the event. Figure 3 demonstrates the application usage over the course of the experiment for the EXDIR. As seen, the EXDIR spent the majority
An Analysis Platform for Mobile Ad Hoc Network (MANET) Scenario Execution Log Data
2016-01-01
these technologies. 4.1 Backend Technologies • Java 1.8 • my-sql-connector- java -5.0.8.jar • Tomcat • VirtualBox • Kali MANET Virtual Machine 4.2...Frontend Technologies • LAMPP 4.3 Database • MySQL Server 5. Database The SEDAP database settings and structure are described in this section...contains all the backend java functionality including the web services, should be placed in the webapps directory inside the Tomcat installation
Design and Implementation of 3D Model Data Management System Based on SQL
NASA Astrophysics Data System (ADS)
Li, Shitao; Zhang, Shixin; Zhang, Zhanling; Li, Shiming; Jia, Kun; Hu, Zhongxu; Ping, Liang; Hu, Youming; Li, Yanlei
CAD/CAM technology plays an increasingly important role in the machinery manufacturing industry. As an important means of production, the accumulated three-dimensional models in many years of design work are valuable. Thus the management of these three-dimensional models is of great significance. This paper gives detailed explanation for a method to design three-dimensional model databases based on SQL and to implement the functions such as insertion, modification, inquiry, preview and so on.
Zhang, Mengliang; Zhao, Yang; Harrington, Peter de B; Chen, Pei
2016-03-01
Two simple fingerprinting methods, flow-injection coupled to ultraviolet spectroscopy and proton nuclear magnetic resonance, were used for discriminating between Aurantii fructus immaturus and Fructus poniciri trifoliatae immaturus . Both methods were combined with partial least-squares discriminant analysis. In the flow-injection method, four data representations were evaluated: total ultraviolet absorbance chromatograms, averaged ultraviolet spectra, absorbance at 193, 205, 225, and 283 nm, and absorbance at 225 and 283 nm. Prediction rates of 100% were achieved for all data representations by partial least-squares discriminant analysis using leave-one-sample-out cross-validation. The prediction rate for the proton nuclear magnetic resonance data by partial least-squares discriminant analysis with leave-one-sample-out cross-validation was also 100%. A new validation set of data was collected by flow-injection with ultraviolet spectroscopic detection two weeks later and predicted by partial least-squares discriminant analysis models constructed by the initial data representations with no parameter changes. The classification rates were 95% with the total ultraviolet absorbance chromatograms datasets and 100% with the other three datasets. Flow-injection with ultraviolet detection and proton nuclear magnetic resonance are simple, high throughput, and low-cost methods for discrimination studies.
NASA Astrophysics Data System (ADS)
Catanzarite, Joseph; Burke, Christopher J.; Li, Jie; Seader, Shawn; Haas, Michael R.; Batalha, Natalie; Henze, Christopher; Christiansen, Jessie; Kepler Project, NASA Advanced Supercomputing Division
2016-06-01
The Kepler Mission is developing an Analytic Completeness Model (ACM) to estimate detection completeness contours as a function of exoplanet radius and period for each target star. Accurate completeness contours are necessary for robust estimation of exoplanet occurrence rates.The main components of the ACM for a target star are: detection efficiency as a function of SNR, the window function (WF) and the one-sigma depth function (OSDF). (Ref. Burke et al. 2015). The WF captures the falloff in transit detection probability at long periods that is determined by the observation window (the duration over which the target star has been observed). The OSDF is the transit depth (in parts per million) that yields SNR of unity for the full transit train. It is a function of period, and accounts for the time-varying properties of the noise and for missing or deweighted data.We are performing flux-level transit injection (FLTI) experiments on selected Kepler target stars with the goal of refining and validating the ACM. “Flux-level” injection machinery inserts exoplanet transit signatures directly into the flux time series, as opposed to “pixel-level” injection, which inserts transit signatures into the individual pixels using the pixel response function. See Jie Li's poster: ID #2493668, "Flux-level transit injection experiments with the NASA Pleiades Supercomputer" for details, including performance statistics.Since FLTI is affordable for only a small subset of the Kepler targets, the ACM is designed to apply to most Kepler target stars. We validate this model using “deep” FLTI experiments, with ~500,000 injection realizations on each of a small number of targets and “shallow” FLTI experiments with ~2000 injection realizations on each of many targets. From the results of these experiments, we identify anomalous targets, model their behavior and refine the ACM accordingly.In this presentation, we discuss progress in validating and refining the ACM, and we compare our detection efficiency curves with those derived from the associated pixel-level transit injection experiments.Kepler was selected as the 10th mission of the Discovery Program. Funding for this mission is provided by NASA, Science Mission Directorate.
CHASM and SNVBox: toolkit for detecting biologically important single nucleotide mutations in cancer
Carter, Hannah; Diekhans, Mark; Ryan, Michael C.; Karchin, Rachel
2011-01-01
Summary: Thousands of cancer exomes are currently being sequenced, yielding millions of non-synonymous single nucleotide variants (SNVs) of possible relevance to disease etiology. Here, we provide a software toolkit to prioritize SNVs based on their predicted contribution to tumorigenesis. It includes a database of precomputed, predictive features covering all positions in the annotated human exome and can be used either stand-alone or as part of a larger variant discovery pipeline. Availability and Implementation: MySQL database, source code and binaries freely available for academic/government use at http://wiki.chasmsoftware.org, Source in Python and C++. Requires 32 or 64-bit Linux system (tested on Fedora Core 8,10,11 and Ubuntu 10), 2.5*≤ Python <3.0*, MySQL server >5.0, 60 GB available hard disk space (50 MB for software and data files, 40 GB for MySQL database dump when uncompressed), 2 GB of RAM. Contact: karchin@jhu.edu Supplementary Information: Supplementary data are available at Bioinformatics online. PMID:21685053
Datacube Services in Action, Using Open Source and Open Standards
NASA Astrophysics Data System (ADS)
Baumann, P.; Misev, D.
2016-12-01
Array Databases comprise novel, promising technology for massive spatio-temporal datacubes, extending the SQL paradigm of "any query, anytime" to n-D arrays. On server side, such queries can be optimized, parallelized, and distributed based on partitioned array storage. The rasdaman ("raster data manager") system, which has pioneered Array Databases, is available in open source on www.rasdaman.org. Its declarative query language extends SQL with array operators which are optimized and parallelized on server side. The rasdaman engine, which is part of OSGeo Live, is mature and in operational use databases individually holding dozens of Terabytes. Further, the rasdaman concepts have strongly impacted international Big Data standards in the field, including the forthcoming MDA ("Multi-Dimensional Array") extension to ISO SQL, the OGC Web Coverage Service (WCS) and Web Coverage Processing Service (WCPS) standards, and the forthcoming INSPIRE WCS/WCPS; in both OGC and INSPIRE, OGC is WCS Core Reference Implementation. In our talk we present concepts, architecture, operational services, and standardization impact of open-source rasdaman, as well as experiences made.
Providing R-Tree Support for Mongodb
NASA Astrophysics Data System (ADS)
Xiang, Longgang; Shao, Xiaotian; Wang, Dehao
2016-06-01
Supporting large amounts of spatial data is a significant characteristic of modern databases. However, unlike some mature relational databases, such as Oracle and PostgreSQL, most of current burgeoning NoSQL databases are not well designed for storing geospatial data, which is becoming increasingly important in various fields. In this paper, we propose a novel method to provide R-tree index, as well as corresponding spatial range query and nearest neighbour query functions, for MongoDB, one of the most prevalent NoSQL databases. First, after in-depth analysis of MongoDB's features, we devise an efficient tabular document structure which flattens R-tree index into MongoDB collections. Further, relevant mechanisms of R-tree operations are issued, and then we discuss in detail how to integrate R-tree into MongoDB. Finally, we present the experimental results which show that our proposed method out-performs the built-in spatial index of MongoDB. Our research will greatly facilitate big data management issues with MongoDB in a variety of geospatial information applications.
Photo-z-SQL: Integrated, flexible photometric redshift computation in a database
NASA Astrophysics Data System (ADS)
Beck, R.; Dobos, L.; Budavári, T.; Szalay, A. S.; Csabai, I.
2017-04-01
We present a flexible template-based photometric redshift estimation framework, implemented in C#, that can be seamlessly integrated into a SQL database (or DB) server and executed on-demand in SQL. The DB integration eliminates the need to move large photometric datasets outside a database for redshift estimation, and utilizes the computational capabilities of DB hardware. The code is able to perform both maximum likelihood and Bayesian estimation, and can handle inputs of variable photometric filter sets and corresponding broad-band magnitudes. It is possible to take into account the full covariance matrix between filters, and filter zero points can be empirically calibrated using measurements with given redshifts. The list of spectral templates and the prior can be specified flexibly, and the expensive synthetic magnitude computations are done via lazy evaluation, coupled with a caching of results. Parallel execution is fully supported. For large upcoming photometric surveys such as the LSST, the ability to perform in-place photo-z calculation would be a significant advantage. Also, the efficient handling of variable filter sets is a necessity for heterogeneous databases, for example the Hubble Source Catalog, and for cross-match services such as SkyQuery. We illustrate the performance of our code on two reference photo-z estimation testing datasets, and provide an analysis of execution time and scalability with respect to different configurations. The code is available for download at https://github.com/beckrob/Photo-z-SQL.
NoSQL technologies for the CMS Conditions Database
NASA Astrophysics Data System (ADS)
Sipos, Roland
2015-12-01
With the restart of the LHC in 2015, the growth of the CMS Conditions dataset will continue, therefore the need of consistent and highly available access to the Conditions makes a great cause to revisit different aspects of the current data storage solutions. We present a study of alternative data storage backends for the Conditions Databases, by evaluating some of the most popular NoSQL databases to support a key-value representation of the CMS Conditions. The definition of the database infrastructure is based on the need of storing the conditions as BLOBs. Because of this, each condition can reach the size that may require special treatment (splitting) in these NoSQL databases. As big binary objects may be problematic in several database systems, and also to give an accurate baseline, a testing framework extension was implemented to measure the characteristics of the handling of arbitrary binary data in these databases. Based on the evaluation, prototypes of a document store, using a column-oriented and plain key-value store, are deployed. An adaption layer to access the backends in the CMS Offline software was developed to provide transparent support for these NoSQL databases in the CMS context. Additional data modelling approaches and considerations in the software layer, deployment and automatization of the databases are also covered in the research. In this paper we present the results of the evaluation as well as a performance comparison of the prototypes studied.
Pesteie, Mehran; Abolmaesumi, Purang; Ashab, Hussam Al-Deen; Lessoway, Victoria A; Massey, Simon; Gunka, Vit; Rohling, Robert N
2015-06-01
Injection therapy is a commonly used solution for back pain management. This procedure typically involves percutaneous insertion of a needle between or around the vertebrae, to deliver anesthetics near nerve bundles. Most frequently, spinal injections are performed either blindly using palpation or under the guidance of fluoroscopy or computed tomography. Recently, due to the drawbacks of the ionizing radiation of such imaging modalities, there has been a growing interest in using ultrasound imaging as an alternative. However, the complex spinal anatomy with different wave-like structures, affected by speckle noise, makes the accurate identification of the appropriate injection plane difficult. The aim of this study was to propose an automated system that can identify the optimal plane for epidural steroid injections and facet joint injections. A multi-scale and multi-directional feature extraction system to provide automated identification of the appropriate plane is proposed. Local Hadamard coefficients are obtained using the sequency-ordered Hadamard transform at multiple scales. Directional features are extracted from local coefficients which correspond to different regions in the ultrasound images. An artificial neural network is trained based on the local directional Hadamard features for classification. The proposed method yields distinctive features for classification which successfully classified 1032 images out of 1090 for epidural steroid injection and 990 images out of 1052 for facet joint injection. In order to validate the proposed method, a leave-one-out cross-validation was performed. The average classification accuracy for leave-one-out validation was 94 % for epidural and 90 % for facet joint targets. Also, the feature extraction time for the proposed method was 20 ms for a native 2D ultrasound image. A real-time machine learning system based on the local directional Hadamard features extracted by the sequency-ordered Hadamard transform for detecting the laminae and facet joints in ultrasound images has been proposed. The system has the potential to assist the anesthesiologists in quickly finding the target plane for epidural steroid injections and facet joint injections.
Field Validation of Supercritical CO 2 Reactivity with Basalts
DOE Office of Scientific and Technical Information (OSTI.GOV)
McGrail, B. Peter; Schaef, Herbert T.; Spane, Frank A.
2017-01-10
Continued global use of fossil fuels places a premium on developing technology solutions to minimize increases in atmospheric CO 2 levels. CO 2 storage in reactive basalts might be one of these solutions by permanently converting injected gaseous CO 2 into solid carbonates. Herein we report results from a field demonstration where ~1000 MT of CO 2 was injected into a natural basalt formation in Eastern Washington State. Following two years of post-injection monitoring, cores were obtained from within the injection zone and subjected to detailed physical and chemical analysis. Nodules found in vesicles throughout the cores were identified asmore » the carbonate mineral, ankerite Ca[Fe, Mg, Mn](CO 3) 2. Carbon isotope analysis showed the nodules are chemically distinct as compared with natural carbonates present in the basalt and clear correlation with the isotopic signature of the injected CO 2. These findings provide field validation of rapid mineralization rates observed from years of laboratory testing with basalts.« less
Validity of injecting drug users' self report of hepatitis A, B, and C.
Schlicting, Erin G; Johnson, Mark E; Brems, Christiane; Wells, Rebecca S; Fisher, Dennis G; Reynolds, Grace
2003-01-01
To test the validity of drug users self-reports of diseases associated with drug use, in this case hepatitis A, B, and C. Injecting drug users (n = 653) were recruited and asked whether they had been diagnosed previously with hepatitis A, B, and/or C. These self-report data were compared to total hepatitis A antibody, hepatitis B core antibody, and hepatitis C antibody seromarkers as a means of determining the validity of the self-reported information. Anchorage, Alaska. Criteria for inclusion included being at least 18-years old; testing positive on urinalysis for cocaine metabolites, amphetamine, or morphine; having visible signs of injection (track marks). Serological testing for hepatitis A, B, and C. Findings indicate high specificity, low sensitivity, and low kappa coefficients for all three self-report measures. Subgroup analyses revealed significant differences in sensitivity associated with previous substance abuse treatment experience for hepatitis B self-report and with gender for hepatitis C self-report. Given the low sensitivity, the validity of drug users, self-reported information on hepatitis should be considered with caution.
Techniques for Efficiently Managing Large Geosciences Data Sets
NASA Astrophysics Data System (ADS)
Kruger, A.; Krajewski, W. F.; Bradley, A. A.; Smith, J. A.; Baeck, M. L.; Steiner, M.; Lawrence, R. E.; Ramamurthy, M. K.; Weber, J.; Delgreco, S. A.; Domaszczynski, P.; Seo, B.; Gunyon, C. A.
2007-12-01
We have developed techniques and software tools for efficiently managing large geosciences data sets. While the techniques were developed as part of an NSF-Funded ITR project that focuses on making NEXRAD weather data and rainfall products available to hydrologists and other scientists, they are relevant to other geosciences disciplines that deal with large data sets. Metadata, relational databases, data compression, and networking are central to our methodology. Data and derived products are stored on file servers in a compressed format. URLs to, and metadata about the data and derived products are managed in a PostgreSQL database. Virtually all access to the data and products is through this database. Geosciences data normally require a number of processing steps to transform the raw data into useful products: data quality assurance, coordinate transformations and georeferencing, applying calibration information, and many more. We have developed the concept of crawlers that manage this scientific workflow. Crawlers are unattended processes that run indefinitely, and at set intervals query the database for their next assignment. A database table functions as a roster for the crawlers. Crawlers perform well-defined tasks that are, except for perhaps sequencing, largely independent from other crawlers. Once a crawler is done with its current assignment, it updates the database roster table, and gets its next assignment by querying the database. We have developed a library that enables one to quickly add crawlers. The library provides hooks to external (i.e., C-language) compiled codes, so that developers can work and contribute independently. Processes called ingesters inject data into the system. The bulk of the data are from a real-time feed using UCAR/Unidata's IDD/LDM software. An exciting recent development is the establishment of a Unidata HYDRO feed that feeds value-added metadata over the IDD/LDM. Ingesters grab the metadata and populate the PostgreSQL tables. These and other concepts we have developed have enabled us to efficiently manage a 70 Tb (and growing) data weather radar data set.
Park, Dong-Jin; Kang, Ji-Hyoun; Lee, Jeong-Won; Lee, Kyung-Eun; Wen, Lihui; Kim, Tae-Jong; Park, Yong-Wook; Nam, Tai-Seung; Kim, Myung-Sun; Lee, Shin-Seok
2013-07-01
The aim of this study was to assess and validate the Korean version of the Boston Carpal Tunnel Questionnaire (K-BCTQ) in patients with carpal tunnel syndrome (CTS). After translation and cultural adaptation of the BCTQ to a Korean version, the K-BCTQ was administered to 54 patients with CTS; it was administered again after 2 weeks to assess reliability. Additionally, we administered K-DASH and EQ-5D to assess construct-validity. In a prospective study of responsiveness to clinical change, 29 of 54 patients were treated by ultrasonography-guided local corticosteroid injection therapy. The internal consistency of the K-BCTQ was high (Cronbach's alpha: 0.915) and the intra-class correlation coefficients were 0.931 for the symptom severity scale (P<0.001) and 0.844 for the functional severity scale (P<0.001). The construct-validity between the symptom severity scale and the K-DASH, and between the functional severity scale and the K-DASH were significantly correlated (both P<0.001). Clinical improvement was noted in 29 patients with injection therapy. The effect size of symptom severity was 0.67, and that of functional severity was 0.58. In conclusion, the K-BCTQ shows good reliability, construct-validity, and acceptable responsiveness after local corticosteroid injection therapy (Clinical trial number, KCT0000050).
NASA Astrophysics Data System (ADS)
Song, Enzhe; Fan, Liyun; Chen, Chao; Dong, Quan; Ma, Xiuzhen; Bai, Yun
2013-09-01
A simulation model of an electronically controlled two solenoid valve fuel injection system for a diesel engine is established in the AMESim environment. The accuracy of the model is validated through comparison with experimental data. The influence of pre-injection control parameters on main-injection quantity under different control modes is analyzed. In the spill control valve mode, main-injection fuel quantity decreases gradually and then reaches a stable level because of the increase in multi-injection dwell time. In the needle control valve mode, main-injection fuel quantity increases with rising multi-injection dwell time; this effect becomes more obvious at high-speed revolutions and large main-injection pulse widths. Pre-injection pulse width has no obvious influence on main-injection quantity under the two control modes; the variation in main-injection quantity is in the range of 1 mm3.
Re-Engineering the United States Marine Corps’ Enlisted Assignment Model (EAM)
1998-06-01
and Execution button opens the switchboard in Figure 12. This form accesses all of the VBA code that is associated with this form and the code that...the prototype to prompt the user and to inform him of what he is about to do. Each of the buttons that are on these forms is connected to an SQL ...into the field for building a rule are the values 2;3;4;5;6;7. The user can enter an SQL statement that would determine the values or the user could
Virtual file system on NoSQL for processing high volumes of HL7 messages.
Kimura, Eizen; Ishihara, Ken
2015-01-01
The Standardized Structured Medical Information Exchange (SS-MIX) is intended to be the standard repository for HL7 messages that depend on a local file system. However, its scalability is limited. We implemented a virtual file system using NoSQL to incorporate modern computing technology into SS-MIX and allow the system to integrate local patient IDs from different healthcare systems into a universal system. We discuss its implementation using the database MongoDB and describe its performance in a case study.
2013-01-01
Background Quality of life (QOL) is an important outcome measure in the treatment of heroin addiction. The Taiwan version of the World Health Organization Quality of Life assessment (WHOQOL-BREF [TW]) has been developed and studied in various groups, but not specifically in a population of injection drug users. The aim of this study was to analyze the psychometric properties of the WHOQOL-BREF (TW) in a sample of injection drug users undergoing methadone maintenance treatment. Methods A total of 553 participants were interviewed and completed the instrument. Item-response distributions, internal consistency, corrected item-domain correlation, criterion-related validity, and construct validity through confirmatory factor analysis were evaluated. Results The frequency distribution of the 4 domains of the WHOQOL-BREF (TW) showed no floor or ceiling effects. The instrument demonstrated adequate internal consistency (Cronbach’s alpha coefficients were higher than 0.7 across the 4 domains) and all items had acceptable correlation with the corresponding domain scores (r = 0.32-0.73). Correlations (p < 0.01) of the 4 domains with the 2 benchmark items assessing overall QOL and general health were supportive of criterion-related validity. Confirmatory factor analysis yielded marginal goodness-of-fit between the 4-domain model and the sample data. Conclusions The hypothesized WHOQOL-BREF measurement model was appropriate for the injection drug users after some adjustments. Despite different patterns found in the confirmatory factor analysis, the findings overall suggest that the WHOQOL-BREF (TW) is a reliable and valid measure of QOL among injection drug users and can be utilized in future treatment outcome studies. The factor structure provided by the study also helps to understand the QOL characteristics of the injection drug users in Taiwan. However, more research is needed to examine its test-retest reliability and sensitivity to changes due to treatment. PMID:24325611
Hatt, Charles R.; Jain, Ameet K.; Parthasarathy, Vijay; Lang, Andrew; Raval, Amish N.
2014-01-01
Myocardial infarction (MI) is one of the leading causes of death in the world. Small animal studies have shown that stem-cell therapy offers dramatic functional improvement post-MI. An endomyocardial catheter injection approach to therapeutic agent delivery has been proposed to improve efficacy through increased cell retention. Accurate targeting is critical for reaching areas of greatest therapeutic potential while avoiding a life-threatening myocardial perforation. Multimodal image fusion has been proposed as a way to improve these procedures by augmenting traditional intra-operative imaging modalities with high resolution pre-procedural images. Previous approaches have suffered from a lack of real-time tissue imaging and dependence on X-ray imaging to track devices, leading to increased ionizing radiation dose. In this paper, we present a new image fusion system for catheter-based targeted delivery of therapeutic agents. The system registers real-time 3D echocardiography, magnetic resonance, X-ray, and electromagnetic sensor tracking within a single flexible framework. All system calibrations and registrations were validated and found to have target registration errors less than 5 mm in the worst case. Injection accuracy was validated in a motion enabled cardiac injection phantom, where targeting accuracy ranged from 0.57 to 3.81 mm. Clinical feasibility was demonstrated with in-vivo swine experiments, where injections were successfully made into targeted regions of the heart. PMID:23561056
Mahony, Mary C; Patterson, Patricia; Hayward, Brooke; North, Robert; Green, Dawne
2015-05-01
To demonstrate, using human factors engineering (HFE), that a redesigned, pre-filled, ready-to-use, pre-asembled follitropin alfa pen can be used to administer prescribed follitropin alfa doses safely and accurately. A failure modes and effects analysis identified hazards and harms potentially caused by use errors; risk-control measures were implemented to ensure acceptable device use risk management. Participants were women with infertility, their significant others, and fertility nurse (FN) professionals. Preliminary testing included 'Instructions for Use' (IFU) and pre-validation studies. Validation studies used simulated injections in a representative use environment; participants received prior training on pen use. User performance in preliminary testing led to IFU revisions and a change to outer needle cap design to mitigate needle stick potential. In the first validation study (49 users, 343 simulated injections), in the FN group, one observed critical use error resulted in a device design modification and another in an IFU change. A second validation study tested the mitigation strategies; previously reported use errors were not repeated. Through an iterative process involving a series of studies, modifications were made to the pen design and IFU. Simulated-use testing demonstrated that the redesigned pen can be used to administer follitropin alfa effectively and safely.
Dweep, Harsh; Sticht, Carsten; Pandey, Priyanka; Gretz, Norbert
2011-10-01
MicroRNAs are small, non-coding RNA molecules that can complementarily bind to the mRNA 3'-UTR region to regulate the gene expression by transcriptional repression or induction of mRNA degradation. Increasing evidence suggests a new mechanism by which miRNAs may regulate target gene expression by binding in promoter and amino acid coding regions. Most of the existing databases on miRNAs are restricted to mRNA 3'-UTR region. To address this issue, we present miRWalk, a comprehensive database on miRNAs, which hosts predicted as well as validated miRNA binding sites, information on all known genes of human, mouse and rat. All mRNAs, mitochondrial genes and 10 kb upstream flanking regions of all known genes of human, mouse and rat were analyzed by using a newly developed algorithm named 'miRWalk' as well as with eight already established programs for putative miRNA binding sites. An automated and extensive text-mining search was performed on PubMed database to extract validated information on miRNAs. Combined information was put into a MySQL database. miRWalk presents predicted and validated information on miRNA-target interaction. Such a resource enables researchers to validate new targets of miRNA not only on 3'-UTR, but also on the other regions of all known genes. The 'Validated Target module' is updated every month and the 'Predicted Target module' is updated every 6 months. miRWalk is freely available at http://mirwalk.uni-hd.de/. Copyright © 2011 Elsevier Inc. All rights reserved.
Validation of the NCC Code for Staged Transverse Injection and Computations for a RBCC Combustor
NASA Technical Reports Server (NTRS)
Ajmani, Kumud; Liu, Nan-Suey
2005-01-01
The NCC code was validated for a case involving staged transverse injection into Mach 2 flow behind a rearward facing step. Comparisons with experimental data and with solutions from the FPVortex code was then used to perform computations to study fuel-air mixing for the combustor of a candidate rocket based combined cycle engine geometry. Comparisons with a one-dimensional analysis and a three-dimensional code (VULCAN) were performed to assess the qualitative and quantitative performance of the NCC solver.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Nguyen, Ba Nghiep; Jin, Xiaoshi; Wang, Jin
2012-02-23
This report describes the work conducted under the CRADA Nr. PNNL/304 between Battelle PNNL and Autodesk whose objective is to validate the new process models developed under the previous CRADA for large injection-molded LFT composite structures. To this end, the ARD-RSC and fiber length attrition models implemented in the 2013 research version of Moldflow was used to simulate the injection molding of 600-mm x 600-mm x 3-mm plaques from 40% glass/polypropylene (Dow Chemical DLGF9411.00) and 40% glass/polyamide 6,6 (DuPont Zytel 75LG40HSL BK031) materials. The injection molding was performed by Injection Technologies, Inc. at Windsor, Ontario (under a subcontract by Oakmore » Ridge National Laboratory, ORNL) using the mold offered by the Automotive Composite Consortium (ACC). Two fill speeds under the same back pressure were used to produce plaques under slow-fill and fast-fill conditions. Also, two gating options were used to achieve the following desired flow patterns: flows in edge-gated plaques and in center-gated plaques. After molding, ORNL performed measurements of fiber orientation and length distributions for process model validations. The structure of this report is as follows. After the Introduction (Section 1), Section 2 provides a summary of the ARD-RSC and fiber length attrition models. A summary of model implementations in the latest research version of Moldflow is given in Section 3. Section 4 provides the key processing conditions and parameters for molding of the ACC plaques. The validations of the ARD-RSC and fiber length attrition models are presented and discussed in Section 5. The conclusions will be drawn in Section 6.« less
Heterogeneity, pore pressure, and injectate chemistry: Control measures for geologic carbon storage
Dewers, Thomas; Eichhubl, Peter; Ganis, Ben; ...
2017-11-28
Desirable outcomes for geologic carbon storage include maximizing storage efficiency, preserving injectivity, and avoiding unwanted consequences such as caprock or wellbore leakage or induced seismicity during and post injection. Here, to achieve these outcomes, three control measures are evident including pore pressure, injectate chemistry, and knowledge and prudent use of geologic heterogeneity. Field, experimental, and modeling examples are presented that demonstrate controllable GCS via these three measures. Observed changes in reservoir response accompanying CO 2 injection at the Cranfield (Mississippi, USA) site, along with lab testing, show potential for use of injectate chemistry as a means to alter fracture permeabilitymore » (with concomitant improvements for sweep and storage efficiency). Further control of reservoir sweep attends brine extraction from reservoirs, with benefit for pressure control, mitigation of reservoir and wellbore damage, and water use. State-of-the-art validated models predict the extent of damage and deformation associated with pore pressure hazards in reservoirs, timing and location of networks of fractures, and development of localized leakage pathways. Experimentally validated geomechanics models show where wellbore failure is likely to occur during injection, and efficiency of repair methods. Use of heterogeneity as a control measure includes where best to inject, and where to avoid attempts at storage. Lastly, an example is use of waste zones or leaky seals to both reduce pore pressure hazards and enhance residual CO 2 trapping.« less
Heterogeneity, pore pressure, and injectate chemistry: Control measures for geologic carbon storage
DOE Office of Scientific and Technical Information (OSTI.GOV)
Dewers, Thomas; Eichhubl, Peter; Ganis, Ben
Desirable outcomes for geologic carbon storage include maximizing storage efficiency, preserving injectivity, and avoiding unwanted consequences such as caprock or wellbore leakage or induced seismicity during and post injection. Here, to achieve these outcomes, three control measures are evident including pore pressure, injectate chemistry, and knowledge and prudent use of geologic heterogeneity. Field, experimental, and modeling examples are presented that demonstrate controllable GCS via these three measures. Observed changes in reservoir response accompanying CO 2 injection at the Cranfield (Mississippi, USA) site, along with lab testing, show potential for use of injectate chemistry as a means to alter fracture permeabilitymore » (with concomitant improvements for sweep and storage efficiency). Further control of reservoir sweep attends brine extraction from reservoirs, with benefit for pressure control, mitigation of reservoir and wellbore damage, and water use. State-of-the-art validated models predict the extent of damage and deformation associated with pore pressure hazards in reservoirs, timing and location of networks of fractures, and development of localized leakage pathways. Experimentally validated geomechanics models show where wellbore failure is likely to occur during injection, and efficiency of repair methods. Use of heterogeneity as a control measure includes where best to inject, and where to avoid attempts at storage. Lastly, an example is use of waste zones or leaky seals to both reduce pore pressure hazards and enhance residual CO 2 trapping.« less
Monitoring of IaaS and scientific applications on the Cloud using the Elasticsearch ecosystem
NASA Astrophysics Data System (ADS)
Bagnasco, S.; Berzano, D.; Guarise, A.; Lusso, S.; Masera, M.; Vallero, S.
2015-05-01
The private Cloud at the Torino INFN computing centre offers IaaS services to different scientific computing applications. The infrastructure is managed with the OpenNebula cloud controller. The main stakeholders of the facility are a grid Tier-2 site for the ALICE collaboration at LHC, an interactive analysis facility for the same experiment and a grid Tier-2 site for the BES-III collaboration, plus an increasing number of other small tenants. Besides keeping track of the usage, the automation of dynamic allocation of resources to tenants requires detailed monitoring and accounting of the resource usage. As a first investigation towards this, we set up a monitoring system to inspect the site activities both in terms of IaaS and applications running on the hosted virtual instances. For this purpose we used the Elasticsearch, Logstash and Kibana stack. In the current implementation, the heterogeneous accounting information is fed to different MySQL databases and sent to Elasticsearch via a custom Logstash plugin. For the IaaS metering, we developed sensors for the OpenNebula API. The IaaS level information gathered through the API is sent to the MySQL database through an ad-hoc developed RESTful web service, which is also used for other accounting purposes. Concerning the application level, we used the Root plugin TProofMonSenderSQL to collect accounting data from the interactive analysis facility. The BES-III virtual instances used to be monitored with Zabbix, as a proof of concept we also retrieve the information contained in the Zabbix database. Each of these three cases is indexed separately in Elasticsearch. We are now starting to consider dismissing the intermediate level provided by the SQL database and evaluating a NoSQL option as a unique central database for all the monitoring information. We setup a set of Kibana dashboards with pre-defined queries in order to monitor the relevant information in each case. In this way we have achieved a uniform monitoring interface for both the IaaS and the scientific applications, mostly leveraging off-the-shelf tools.
"Science SQL" as a Building Block for Flexible, Standards-based Data Infrastructures
NASA Astrophysics Data System (ADS)
Baumann, Peter
2016-04-01
We have learnt to live with the pain of separating data and metadata into non-interoperable silos. For metadata, we enjoy the flexibility of databases, be they relational, graph, or some other NoSQL. Contrasting this, users still "drown in files" as an unstructured, low-level archiving paradigm. It is time to bridge this chasm which once was technologically induced, but today can be overcome. One building block towards a common re-integrated information space is to support massive multi-dimensional spatio-temporal arrays. These "datacubes" appear as sensor, image, simulation, and statistics data in all science and engineering domains, and beyond. For example, 2-D satellilte imagery, 2-D x/y/t image timeseries and x/y/z geophysical voxel data, and 4-D x/y/z/t climate data contribute to today's data deluge in the Earth sciences. Virtual observatories in the Space sciences routinely generate Petabytes of such data. Life sciences deal with microarray data, confocal microscopy, human brain data, which all fall into the same category. The ISO SQL/MDA (Multi-Dimensional Arrays) candidate standard is extending SQL with modelling and query support for n-D arrays ("datacubes") in a flexible, domain-neutral way. This heralds a new generation of services with new quality parameters, such as flexibility, ease of access, embedding into well-known user tools, and scalability mechanisms that remain completely transparent to users. Technology like the EU rasdaman ("raster data manager") Array Database system can support all of the above examples simultaneously, with one technology. This is practically proven: As of today, rasdaman is in operational use on hundreds of Terabytes of satellite image timeseries datacubes, with transparent query distribution across more than 1,000 nodes. Therefore, Array Databases offering SQL/MDA constitute a natural common building block for next-generation data infrastructures. Being initiator and editor of the standard we present principles, implementation facets, and application examples as a basis for further discussion. Further, we highlight recent implementation progress in parallelization, data distribution, and query optimization showing their effects on real-life use cases.
Gharibi Loron, Ali; Sardari, Soroush; Narenjkar, Jamshid; Sayyah, Mohammad
2017-01-01
Resistance to antiepileptic drugs and the intolerability in 20-30% of the patients raises demand for developing new drugs with improved efficacy and safety. Acceptable anticonvulsant activity, good tolerability, and inexpensiveness of docosahexaenoic acid (DHA) make it as a good candidate for designing and development of the new anticonvulsant medications. Ten DHA-based molecules were screened based on in silico screening of DHA-like molecules by root-mean-square deviation of atomic positions, the biological activity score of Professional Association for SQL Server, and structural requirements suggested by pharmacophore design. Anticonvulsant activity was tested against clonic seizures induced by pentylenetetrazole (PTZ, 60 mg/kg, i.p.) and tonic seizures induced by maximal electroshock (MES, 50 mA, 50 Hz, 1 ms duration) by intracerebroventricular (i.c.v.) injection of the screened compounds to mice. Among screened compounds, 4-Phenylbutyric acid, 4-Biphenylacetic acid, phenylacetic acid, and 2-Phenylbutyric acid showed significant protective activity in pentylenetetrazole test with ED50 values of 4, 5, 78, and 70 mM, respectively. In MES test, shikimic acid and 4-tert-Butylcyclo-hexanecarboxylic acid showed significant activity with ED50 values 29 and 637 mM, respectively. Effective compounds had no mortality in mice up to the maximum i.c.v. injectable dose of 1 mM. Common electrochemical features and three-dimensional spatial structures of the effective compounds suggest the involvement of the anticonvulsant mechanisms similar to the parent compound DHA.
Loron, Ali Gharibi; Sardari, Soroush; Narenjkar, Jamshid; Sayyah, Mohammad
2017-01-01
Background: Resistance to antiepileptic drugs and the intolerability in 20-30% of the patients raises demand for developing new drugs with improved efficacy and safety. Acceptable anticonvulsant activity, good tolerability, and inexpensiveness of docosahexaenoic acid (DHA) make it as a good candidate for designing and development of the new anticonvulsant medications. Methods: Ten DHA-based molecules were screened based on in silico screening of DHA-like molecules by root-mean-square deviation of atomic positions, the biological activity score of Professional Association for SQL Server, and structural requirements suggested by pharmacophore design. Anticonvulsant activity was tested against clonic seizures induced by pentylenetetrazole (PTZ, 60 mg/kg, i.p.) and tonic seizures induced by maximal electroshock (MES, 50 mA, 50 Hz, 1 ms duration) by intracerebroventricular (i.c.v.) injection of the screened compounds to mice. Results: Among screened compounds, 4-Phenylbutyric acid, 4-Biphenylacetic acid, phenylacetic acid, and 2-Phenylbutyric acid showed significant protective activity in pentylenetetrazole test with ED50 values of 4, 5, 78, and 70 mM, respectively. In MES test, shikimic acid and 4-tert-Butylcyclo-hexanecarboxylic acid showed significant activity with ED50 values 29 and 637 mM, respectively. Effective compounds had no mortality in mice up to the maximum i.c.v. injectable dose of 1 mM. Conclusion: Common electrochemical features and three-dimensional spatial structures of the effective compounds suggest the involvement of the anticonvulsant mechanisms similar to the parent compound DHA. PMID:27592363
Agile Datacube Analytics (not just) for the Earth Sciences
NASA Astrophysics Data System (ADS)
Misev, Dimitar; Merticariu, Vlad; Baumann, Peter
2017-04-01
Metadata are considered small, smart, and queryable; data, on the other hand, are known as big, clumsy, hard to analyze. Consequently, gridded data - such as images, image timeseries, and climate datacubes - are managed separately from the metadata, and with different, restricted retrieval capabilities. One reason for this silo approach is that databases, while good at tables, XML hierarchies, RDF graphs, etc., traditionally do not support multi-dimensional arrays well. This gap is being closed by Array Databases which extend the SQL paradigm of "any query, anytime" to NoSQL arrays. They introduce semantically rich modelling combined with declarative, high-level query languages on n-D arrays. On Server side, such queries can be optimized, parallelized, and distributed based on partitioned array storage. This way, they offer new vistas in flexibility, scalability, performance, and data integration. In this respect, the forthcoming ISO SQL extension MDA ("Multi-dimensional Arrays") will be a game changer in Big Data Analytics. We introduce concepts and opportunities through the example of rasdaman ("raster data manager") which in fact has pioneered the field of Array Databases and forms the blueprint for ISO SQL/MDA and further Big Data standards, such as OGC WCPS for querying spatio-temporal Earth datacubes. With operational installations exceeding 140 TB queries have been split across more than one thousand cloud nodes, using CPUs as well as GPUs. Installations can easily be mashed up securely, enabling large-scale location-transparent query processing in federations. Federation queries have been demonstrated live at EGU 2016 spanning Europe and Australia in the context of the intercontinental EarthServer initiative, visualized through NASA WorldWind.
Agile Datacube Analytics (not just) for the Earth Sciences
NASA Astrophysics Data System (ADS)
Baumann, P.
2016-12-01
Metadata are considered small, smart, and queryable; data, on the other hand, are known as big, clumsy, hard to analyze. Consequently, gridded data - such as images, image timeseries, and climate datacubes - are managed separately from the metadata, and with different, restricted retrieval capabilities. One reason for this silo approach is that databases, while good at tables, XML hierarchies, RDF graphs, etc., traditionally do not support multi-dimensional arrays well.This gap is being closed by Array Databases which extend the SQL paradigm of "any query, anytime" to NoSQL arrays. They introduce semantically rich modelling combined with declarative, high-level query languages on n-D arrays. On Server side, such queries can be optimized, parallelized, and distributed based on partitioned array storage. This way, they offer new vistas in flexibility, scalability, performance, and data integration. In this respect, the forthcoming ISO SQL extension MDA ("Multi-dimensional Arrays") will be a game changer in Big Data Analytics.We introduce concepts and opportunities through the example of rasdaman ("raster data manager") which in fact has pioneered the field of Array Databases and forms the blueprint for ISO SQL/MDA and further Big Data standards, such as OGC WCPS for querying spatio-temporal Earth datacubes. With operational installations exceeding 140 TB queries have been split across more than one thousand cloud nodes, using CPUs as well as GPUs. Installations can easily be mashed up securely, enabling large-scale location-transparent query processing in federations. Federation queries have been demonstrated live at EGU 2016 spanning Europe and Australia in the context of the intercontinental EarthServer initiative, visualized through NASA WorldWind.
Improving Broadband Displacement Detection with Quantum Correlations
NASA Astrophysics Data System (ADS)
Kampel, N. S.; Peterson, R. W.; Fischer, R.; Yu, P.-L.; Cicak, K.; Simmonds, R. W.; Lehnert, K. W.; Regal, C. A.
2017-04-01
Interferometers enable ultrasensitive measurement in a wide array of applications from gravitational wave searches to force microscopes. The role of quantum mechanics in the metrological limits of interferometers has a rich history, and a large number of techniques to surpass conventional limits have been proposed. In a typical measurement configuration, the trade-off between the probe's shot noise (imprecision) and its quantum backaction results in what is known as the standard quantum limit (SQL). In this work, we investigate how quantum correlations accessed by modifying the readout of the interferometer can access physics beyond the SQL and improve displacement sensitivity. Specifically, we use an optical cavity to probe the motion of a silicon nitride membrane off mechanical resonance, as one would do in a broadband displacement or force measurement, and observe sensitivity better than the SQL dictates for our quantum efficiency. Our measurement illustrates the core idea behind a technique known as variational readout, in which the optical readout quadrature is changed as a function of frequency to improve broadband displacement detection. And, more generally, our result is a salient example of how correlations can aid sensing in the presence of backaction.
Experience with ATLAS MySQL PanDA database service
NASA Astrophysics Data System (ADS)
Smirnov, Y.; Wlodek, T.; De, K.; Hover, J.; Ozturk, N.; Smith, J.; Wenaus, T.; Yu, D.
2010-04-01
The PanDA distributed production and analysis system has been in production use for ATLAS data processing and analysis since late 2005 in the US, and globally throughout ATLAS since early 2008. Its core architecture is based on a set of stateless web services served by Apache and backed by a suite of MySQL databases that are the repository for all PanDA information: active and archival job queues, dataset and file catalogs, site configuration information, monitoring information, system control parameters, and so on. This database system is one of the most critical components of PanDA, and has successfully delivered the functional and scaling performance required by PanDA, currently operating at a scale of half a million jobs per week, with much growth still to come. In this paper we describe the design and implementation of the PanDA database system, its architecture of MySQL servers deployed at BNL and CERN, backup strategy and monitoring tools. The system has been developed, thoroughly tested, and brought to production to provide highly reliable, scalable, flexible and available database services for ATLAS Monte Carlo production, reconstruction and physics analysis.
Survey data and metadata modelling using document-oriented NoSQL
NASA Astrophysics Data System (ADS)
Rahmatuti Maghfiroh, Lutfi; Gusti Bagus Baskara Nugraha, I.
2018-03-01
Survey data that are collected from year to year have metadata change. However it need to be stored integratedly to get statistical data faster and easier. Data warehouse (DW) can be used to solve this limitation. However there is a change of variables in every period that can not be accommodated by DW. Traditional DW can not handle variable change via Slowly Changing Dimension (SCD). Previous research handle the change of variables in DW to manage metadata by using multiversion DW (MVDW). MVDW is designed using relational model. Some researches also found that developing nonrelational model in NoSQL database has reading time faster than the relational model. Therefore, we propose changes to metadata management by using NoSQL. This study proposes a model DW to manage change and algorithms to retrieve data with metadata changes. Evaluation of the proposed models and algorithms result in that database with the proposed design can retrieve data with metadata changes properly. This paper has contribution in comprehensive data analysis with metadata changes (especially data survey) in integrated storage.
Efficient data management tools for the heterogeneous big data warehouse
NASA Astrophysics Data System (ADS)
Alekseev, A. A.; Osipova, V. V.; Ivanov, M. A.; Klimentov, A.; Grigorieva, N. V.; Nalamwar, H. S.
2016-09-01
The traditional RDBMS has been consistent for the normalized data structures. RDBMS served well for decades, but the technology is not optimal for data processing and analysis in data intensive fields like social networks, oil-gas industry, experiments at the Large Hadron Collider, etc. Several challenges have been raised recently on the scalability of data warehouse like workload against the transactional schema, in particular for the analysis of archived data or the aggregation of data for summary and accounting purposes. The paper evaluates new database technologies like HBase, Cassandra, and MongoDB commonly referred as NoSQL databases for handling messy, varied and large amount of data. The evaluation depends upon the performance, throughput and scalability of the above technologies for several scientific and industrial use-cases. This paper outlines the technologies and architectures needed for processing Big Data, as well as the description of the back-end application that implements data migration from RDBMS to NoSQL data warehouse, NoSQL database organization and how it could be useful for further data analytics.
Advanced information processing system: Fault injection study and results
NASA Technical Reports Server (NTRS)
Burkhardt, Laura F.; Masotto, Thomas K.; Lala, Jaynarayan H.
1992-01-01
The objective of the AIPS program is to achieve a validated fault tolerant distributed computer system. The goals of the AIPS fault injection study were: (1) to present the fault injection study components addressing the AIPS validation objective; (2) to obtain feedback for fault removal from the design implementation; (3) to obtain statistical data regarding fault detection, isolation, and reconfiguration responses; and (4) to obtain data regarding the effects of faults on system performance. The parameters are described that must be varied to create a comprehensive set of fault injection tests, the subset of test cases selected, the test case measurements, and the test case execution. Both pin level hardware faults using a hardware fault injector and software injected memory mutations were used to test the system. An overview is provided of the hardware fault injector and the associated software used to carry out the experiments. Detailed specifications are given of fault and test results for the I/O Network and the AIPS Fault Tolerant Processor, respectively. The results are summarized and conclusions are given.
A Quality-Control-Oriented Database for a Mesoscale Meteorological Observation Network
NASA Astrophysics Data System (ADS)
Lussana, C.; Ranci, M.; Uboldi, F.
2012-04-01
In the operational context of a local weather service, data accessibility and quality related issues must be managed by taking into account a wide set of user needs. This work describes the structure and the operational choices made for the operational implementation of a database system storing data from highly automated observing stations, metadata and information on data quality. Lombardy's environmental protection agency, ARPA Lombardia, manages a highly automated mesoscale meteorological network. A Quality Assurance System (QAS) ensures that reliable observational information is collected and disseminated to the users. The weather unit in ARPA Lombardia, at the same time an important QAS component and an intensive data user, has developed a database specifically aimed to: 1) providing quick access to data for operational activities and 2) ensuring data quality for real-time applications, by means of an Automatic Data Quality Control (ADQC) procedure. Quantities stored in the archive include hourly aggregated observations of: precipitation amount, temperature, wind, relative humidity, pressure, global and net solar radiation. The ADQC performs several independent tests on raw data and compares their results in a decision-making procedure. An important ADQC component is the Spatial Consistency Test based on Optimal Interpolation. Interpolated and Cross-Validation analysis values are also stored in the database, providing further information to human operators and useful estimates in case of missing data. The technical solution adopted is based on a LAMP (Linux, Apache, MySQL and Php) system, constituting an open source environment suitable for both development and operational practice. The ADQC procedure itself is performed by R scripts directly interacting with the MySQL database. Users and network managers can access the database by using a set of web-based Php applications.
Lowering the Barrier for Standards-Compliant and Discoverable Hydrological Data Publication
NASA Astrophysics Data System (ADS)
Kadlec, J.
2013-12-01
The growing need for sharing and integration of hydrological and climate data across multiple organizations has resulted in the development of distributed, services-based, standards-compliant hydrological data management and data hosting systems. The problem with these systems is complicated set-up and deployment. Many existing systems assume that the data publisher has remote-desktop access to a locally managed server and experience with computer network setup. For corporate websites, shared web hosting services with limited root access provide an inexpensive, dynamic web presence solution using the Linux, Apache, MySQL and PHP (LAMP) software stack. In this paper, we hypothesize that a webhosting service provides an optimal, low-cost solution for hydrological data hosting. We propose a software architecture of a standards-compliant, lightweight and easy-to-deploy hydrological data management system that can be deployed on the majority of existing shared internet webhosting services. The architecture and design is validated by developing Hydroserver Lite: a PHP and MySQL-based hydrological data hosting package that is fully standards-compliant and compatible with the Consortium of Universities for Advancement of Hydrologic Sciences (CUAHSI) hydrologic information system. It is already being used for management of field data collection by students of the McCall Outdoor Science School in Idaho. For testing, the Hydroserver Lite software has been installed on multiple different free and low-cost webhosting sites including Godaddy, Bluehost and 000webhost. The number of steps required to set-up the server is compared with the number of steps required to set-up other standards-compliant hydrologic data hosting systems including THREDDS, IstSOS and MapServer SOS.
Injection Drug User Quality of Life Scale (IDUQOL): findings from a content validation study.
Hubley, Anita M; Palepu, Anita
2007-07-30
Quality of life studies among injection drug users have primarily focused on health-related measures. The chaotic life-style of many injection drug users (IDUs), however, extends far beyond their health, and impacts upon social relationships, employment opportunities, housing, and day to day survival. Most current quality of life instruments do not capture the realities of people living with addictions. The Injection Drug Users' Quality of Life Scale (IDUQOL) was developed to reflect the life areas of relevance to IDUs. The present study examined the content validity of the IDUQOL using judgmental methods based on subject matter experts' (SMEs) ratings of various elements of this measure (e.g., appropriateness of life areas or items, names and descriptions of life areas, instructions for administration and scoring). Six SMEs were provided with a copy of the IDUQOL and its administration and scoring manual and a detailed content validation questionnaire. Two commonly used judgmental measures of inter-rater agreement, the Content Validity Index (CVI) and the Average Deviation Mean Index (ADM), were used to evaluate SMEs' agreement on ratings of IDUQOL elements. A total of 75 elements of the IDUQOL were examined. The CVI results showed that all elements were endorsed by the required number of SMEs or more. The ADM results showed that acceptable agreement (i.e., practical significance) was obtained for all elements but statistically significant agreement was missed for nine elements. For these elements, SMEs' feedback was examined for ways to improve the elements. Open-ended feedback also provided suggestions for other revisions to the IDUQOL. The results of the study provided strong evidence in support of the content validity of the IDUQOL and direction for the revision of some IDUQOL elements.
Injection Drug User Quality of Life Scale (IDUQOL): Findings from a content validation study
Hubley, Anita M; Palepu, Anita
2007-01-01
Background Quality of life studies among injection drug users have primarily focused on health-related measures. The chaotic life-style of many injection drug users (IDUs), however, extends far beyond their health, and impacts upon social relationships, employment opportunities, housing, and day to day survival. Most current quality of life instruments do not capture the realities of people living with addictions. The Injection Drug Users' Quality of Life Scale (IDUQOL) was developed to reflect the life areas of relevance to IDUs. The present study examined the content validity of the IDUQOL using judgmental methods based on subject matter experts' (SMEs) ratings of various elements of this measure (e.g., appropriateness of life areas or items, names and descriptions of life areas, instructions for administration and scoring). Methods Six SMEs were provided with a copy of the IDUQOL and its administration and scoring manual and a detailed content validation questionnaire. Two commonly used judgmental measures of inter-rater agreement, the Content Validity Index (CVI) and the Average Deviation Mean Index (ADM), were used to evaluate SMEs' agreement on ratings of IDUQOL elements. Results A total of 75 elements of the IDUQOL were examined. The CVI results showed that all elements were endorsed by the required number of SMEs or more. The ADM results showed that acceptable agreement (i.e., practical significance) was obtained for all elements but statistically significant agreement was missed for nine elements. For these elements, SMEs' feedback was examined for ways to improve the elements. Open-ended feedback also provided suggestions for other revisions to the IDUQOL. Conclusion The results of the study provided strong evidence in support of the content validity of the IDUQOL and direction for the revision of some IDUQOL elements. PMID:17663783
Celik, Selda; Pinar, Rukiye
2016-09-01
To examine the psychometric properties of a Turkish version of the Diabetes Fear of Injecting and Self-testing Questionnaire (D-FISQ). Forward-backward translation of the D-FISQ from English into Turkish was conducted. Original English and translated forms were examined by a panel group. Validity was investigated using content, confirmatory factor analysis, and divergent validity. Reliability was assessed using Cronbach α values, item-total correlations, and intraclass correlations. The sample comprised 350 patients with diabetes. Data were analyzed using SPSS 15.0 for Windows and LISREL 8. The content validity index for the panel members was .90, which indicated perfect content validity; items in D-FISQ were clear, concise, readable, and distinct. Confirmatory factor analysis confirmed the original construct of the D-FISQ. All items had factor loadings higher than the recommended level of .40. The D-FISQ scores were discriminated by the level of anxiety. Reliability results were also satisfactory. Cronbach α values were within ideal limits. Item-total correlation coefficient ranged from .72 to .86. In terms of test-retest reliability, intraclass correlation coefficient was found to be over .90. D-FISQ is a valid and reliable questionnaire in assessing needle-prick fear among Turkish patients with diabetes. We recommend performing the Turkish D-FISQ in determining and screening patients with diabetes who have fear related to self-insulin injection and finger-prick test. Thus, health care professionals should be aware of the potential consequences of injection fear such as insulin misuse and poor self-monitoring of blood glucose, which may have unfavorable effects on optimal diabetes management. Copyright © 2016. Published by Elsevier B.V.
A System for Web-based Access to the HSOS Database
NASA Astrophysics Data System (ADS)
Lin, G.
Huairou Solar Observing Station's (HSOS) magnetogram and dopplergram are world-class instruments. Access to their data has opened to the world. Web-based access to the data will provide a powerful, convenient tool for data searching and solar physics. It is necessary that our data be provided to users via the Web when it is opened to the world. In this presentation, the author describes general design and programming construction of the system. The system will be generated by PHP and MySQL. The author also introduces basic feature of PHP and MySQL.
Optical levitation of a mirror for reaching the standard quantum limit.
Michimura, Yuta; Kuwahara, Yuya; Ushiba, Takafumi; Matsumoto, Nobuyuki; Ando, Masaki
2017-06-12
We propose a new method to optically levitate a macroscopic mirror with two vertical Fabry-Pérot cavities linearly aligned. This configuration gives the simplest possible optical levitation in which the number of laser beams used is the minimum of two. We demonstrate that reaching the standard quantum limit (SQL) of a displacement measurement with our system is feasible with current technology. The cavity geometry and the levitated mirror parameters are designed to ensure that the Brownian vibration of the mirror surface is smaller than the SQL. Our scheme provides a promising tool for testing macroscopic quantum mechanics.
Optical levitation of a mirror for reaching the standard quantum limit
NASA Astrophysics Data System (ADS)
Michimura, Yuta; Kuwahara, Yuya; Ushiba, Takafumi; Matsumoto, Nobuyuki; Ando, Masaki
2017-06-01
We propose a new method to optically levitate a macroscopic mirror with two vertical Fabry-P{\\'e}rot cavities linearly aligned. This configuration gives the simplest possible optical levitation in which the number of laser beams used is the minimum of two. We demonstrate that reaching the standard quantum limit (SQL) of a displacement measurement with our system is feasible with current technology. The cavity geometry and the levitated mirror parameters are designed to ensure that the Brownian vibration of the mirror surface is smaller than the SQL. Our scheme provides a promising tool for testing macroscopic quantum mechanics.
Orthos, an alarm system for the ALICE DAQ operations
NASA Astrophysics Data System (ADS)
Chapeland, Sylvain; Carena, Franco; Carena, Wisla; Chibante Barroso, Vasco; Costa, Filippo; Denes, Ervin; Divia, Roberto; Fuchs, Ulrich; Grigore, Alexandru; Simonetti, Giuseppe; Soos, Csaba; Telesca, Adriana; Vande Vyvre, Pierre; von Haller, Barthelemy
2012-12-01
ALICE (A Large Ion Collider Experiment) is the heavy-ion detector studying the physics of strongly interacting matter and the quark-gluon plasma at the CERN LHC (Large Hadron Collider). The DAQ (Data Acquisition System) facilities handle the data flow from the detectors electronics up to the mass storage. The DAQ system is based on a large farm of commodity hardware consisting of more than 600 devices (Linux PCs, storage, network switches), and controls hundreds of distributed hardware and software components interacting together. This paper presents Orthos, the alarm system used to detect, log, report, and follow-up abnormal situations on the DAQ machines at the experimental area. The main objective of this package is to integrate alarm detection and notification mechanisms with a full-featured issues tracker, in order to prioritize, assign, and fix system failures optimally. This tool relies on a database repository with a logic engine, SQL interfaces to inject or query metrics, and dynamic web pages for user interaction. We describe the system architecture, the technologies used for the implementation, and the integration with existing monitoring tools.
A high-performance spatial database based approach for pathology imaging algorithm evaluation
Wang, Fusheng; Kong, Jun; Gao, Jingjing; Cooper, Lee A.D.; Kurc, Tahsin; Zhou, Zhengwen; Adler, David; Vergara-Niedermayr, Cristobal; Katigbak, Bryan; Brat, Daniel J.; Saltz, Joel H.
2013-01-01
Background: Algorithm evaluation provides a means to characterize variability across image analysis algorithms, validate algorithms by comparison with human annotations, combine results from multiple algorithms for performance improvement, and facilitate algorithm sensitivity studies. The sizes of images and image analysis results in pathology image analysis pose significant challenges in algorithm evaluation. We present an efficient parallel spatial database approach to model, normalize, manage, and query large volumes of analytical image result data. This provides an efficient platform for algorithm evaluation. Our experiments with a set of brain tumor images demonstrate the application, scalability, and effectiveness of the platform. Context: The paper describes an approach and platform for evaluation of pathology image analysis algorithms. The platform facilitates algorithm evaluation through a high-performance database built on the Pathology Analytic Imaging Standards (PAIS) data model. Aims: (1) Develop a framework to support algorithm evaluation by modeling and managing analytical results and human annotations from pathology images; (2) Create a robust data normalization tool for converting, validating, and fixing spatial data from algorithm or human annotations; (3) Develop a set of queries to support data sampling and result comparisons; (4) Achieve high performance computation capacity via a parallel data management infrastructure, parallel data loading and spatial indexing optimizations in this infrastructure. Materials and Methods: We have considered two scenarios for algorithm evaluation: (1) algorithm comparison where multiple result sets from different methods are compared and consolidated; and (2) algorithm validation where algorithm results are compared with human annotations. We have developed a spatial normalization toolkit to validate and normalize spatial boundaries produced by image analysis algorithms or human annotations. The validated data were formatted based on the PAIS data model and loaded into a spatial database. To support efficient data loading, we have implemented a parallel data loading tool that takes advantage of multi-core CPUs to accelerate data injection. The spatial database manages both geometric shapes and image features or classifications, and enables spatial sampling, result comparison, and result aggregation through expressive structured query language (SQL) queries with spatial extensions. To provide scalable and efficient query support, we have employed a shared nothing parallel database architecture, which distributes data homogenously across multiple database partitions to take advantage of parallel computation power and implements spatial indexing to achieve high I/O throughput. Results: Our work proposes a high performance, parallel spatial database platform for algorithm validation and comparison. This platform was evaluated by storing, managing, and comparing analysis results from a set of brain tumor whole slide images. The tools we develop are open source and available to download. Conclusions: Pathology image algorithm validation and comparison are essential to iterative algorithm development and refinement. One critical component is the support for queries involving spatial predicates and comparisons. In our work, we develop an efficient data model and parallel database approach to model, normalize, manage and query large volumes of analytical image result data. Our experiments demonstrate that the data partitioning strategy and the grid-based indexing result in good data distribution across database nodes and reduce I/O overhead in spatial join queries through parallel retrieval of relevant data and quick subsetting of datasets. The set of tools in the framework provide a full pipeline to normalize, load, manage and query analytical results for algorithm evaluation. PMID:23599905
Cooley, Mary E; Nayak, Manan M; Abrahm, Janet L; Braun, Ilana M; Rabin, Michael S; Brzozowski, Jane; Lathan, Christopher; Berry, Donna L
2017-08-01
Adequate symptom and quality-of-life (SQL) management is a priority during cancer treatment. eHealth is a timely way to enhance patient-engagement, facilitate communication, and improve health outcomes. The objectives of this study were to describe patient and caregivers' perspectives for providing, processing, and managing SQL data to enhance communication and identify desired components for decision support. Data were collected from 64 participants through questionnaires and focus groups. Analysis was conducted using NVivo. Open and axial coding was completed, grouping commonalities and large constructs into nodes to identify and synthesize themes. Face-to-face meetings with clinicians were the prime time to communicate, and patients strove to understand treatment options and the effect on SQL by bringing caregivers to their visits, taking notes, tracking symptoms, and creating portable health records. Patients/caregivers struggled to self-manage their symptoms and were uncertain when to contact clinicians when experiencing uncontrolled symptoms. Most participants identified eHealth solutions for decision support. However, 38% of participants (n = 24) rarely used computers and identified non-eHealth options for decision support. Core components for both eHealth and non-eHealth systems were access to (1) cancer information, (2) medical records, (3) peer support, and (4) improved support and understanding on when to contact clinicians. Patients were faced with an overwhelming amount of information and relied on their caregivers to help navigate the complexities of cancer care and self-manage SQL. Health technologies can provide informational support; however, decision support needs to span multiple venues to avoid increasing disparities caused by a digital divide. Copyright © 2017 John Wiley & Sons, Ltd.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Roberts, D
Purpose: A unified database system was developed to allow accumulation, review and analysis of quality assurance (QA) data for measurement, treatment, imaging and simulation equipment in our department. Recording these data in a database allows a unified and structured approach to review and analysis of data gathered using commercial database tools. Methods: A clinical database was developed to track records of quality assurance operations on linear accelerators, a computed tomography (CT) scanner, high dose rate (HDR) afterloader and imaging systems such as on-board imaging (OBI) and Calypso in our department. The database was developed using Microsoft Access database and visualmore » basic for applications (VBA) programming interface. Separate modules were written for accumulation, review and analysis of daily, monthly and annual QA data. All modules were designed to use structured query language (SQL) as the basis of data accumulation and review. The SQL strings are dynamically re-written at run time. The database also features embedded documentation, storage of documents produced during QA activities and the ability to annotate all data within the database. Tests are defined in a set of tables that define test type, specific value, and schedule. Results: Daily, Monthly and Annual QA data has been taken in parallel with established procedures to test MQA. The database has been used to aggregate data across machines to examine the consistency of machine parameters and operations within the clinic for several months. Conclusion: The MQA application has been developed as an interface to a commercially available SQL engine (JET 5.0) and a standard database back-end. The MQA system has been used for several months for routine data collection.. The system is robust, relatively simple to extend and can be migrated to a commercial SQL server.« less
An SQL query generator for CLIPS
NASA Technical Reports Server (NTRS)
Snyder, James; Chirica, Laurian
1990-01-01
As expert systems become more widely used, their access to large amounts of external information becomes increasingly important. This information exists in several forms such as statistical, tabular data, knowledge gained by experts and large databases of information maintained by companies. Because many expert systems, including CLIPS, do not provide access to this external information, much of the usefulness of expert systems is left untapped. The scope of this paper is to describe a database extension for the CLIPS expert system shell. The current industry standard database language is SQL. Due to SQL standardization, large amounts of information stored on various computers, potentially at different locations, will be more easily accessible. Expert systems should be able to directly access these existing databases rather than requiring information to be re-entered into the expert system environment. The ORACLE relational database management system (RDBMS) was used to provide a database connection within the CLIPS environment. To facilitate relational database access a query generation system was developed as a CLIPS user function. The queries are entered in a CLlPS-like syntax and are passed to the query generator, which constructs and submits for execution, an SQL query to the ORACLE RDBMS. The query results are asserted as CLIPS facts. The query generator was developed primarily for use within the ICADS project (Intelligent Computer Aided Design System) currently being developed by the CAD Research Unit in the California Polytechnic State University (Cal Poly). In ICADS, there are several parallel or distributed expert systems accessing a common knowledge base of facts. Expert system has a narrow domain of interest and therefore needs only certain portions of the information. The query generator provides a common method of accessing this information and allows the expert system to specify what data is needed without specifying how to retrieve it.
TargetCompare: A web interface to compare simultaneous miRNAs targets
Moreira, Fabiano Cordeiro; Dustan, Bruno; Hamoy, Igor G; Ribeiro-dos-Santos, André M; dos Santos, Ândrea Ribeiro
2014-01-01
MicroRNAs (miRNAs) are small non-coding nucleotide sequences between 17 and 25 nucleotides in length that primarily function in the regulation of gene expression. A since miRNA has thousand of predict targets in a complex, regulatory cell signaling network. Therefore, it is of interest to study multiple target genes simultaneously. Hence, we describe a web tool (developed using Java programming language and MySQL database server) to analyse multiple targets of pre-selected miRNAs. We cross validated the tool in eight most highly expressed miRNAs in the antrum region of stomach. This helped to identify 43 potential genes that are target of at least six of the referred miRNAs. The developed tool aims to reduce the randomness and increase the chance of selecting strong candidate target genes and miRNAs responsible for playing important roles in the studied tissue. Availability http://lghm.ufpa.br/targetcompare PMID:25352731
TargetCompare: A web interface to compare simultaneous miRNAs targets.
Moreira, Fabiano Cordeiro; Dustan, Bruno; Hamoy, Igor G; Ribeiro-Dos-Santos, André M; Dos Santos, Andrea Ribeiro
2014-01-01
MicroRNAs (miRNAs) are small non-coding nucleotide sequences between 17 and 25 nucleotides in length that primarily function in the regulation of gene expression. A since miRNA has thousand of predict targets in a complex, regulatory cell signaling network. Therefore, it is of interest to study multiple target genes simultaneously. Hence, we describe a web tool (developed using Java programming language and MySQL database server) to analyse multiple targets of pre-selected miRNAs. We cross validated the tool in eight most highly expressed miRNAs in the antrum region of stomach. This helped to identify 43 potential genes that are target of at least six of the referred miRNAs. The developed tool aims to reduce the randomness and increase the chance of selecting strong candidate target genes and miRNAs responsible for playing important roles in the studied tissue. http://lghm.ufpa.br/targetcompare.
A Relational Encoding of a Conceptual Model with Multiple Temporal Dimensions
NASA Astrophysics Data System (ADS)
Gubiani, Donatella; Montanari, Angelo
The theoretical interest and the practical relevance of a systematic treatment of multiple temporal dimensions is widely recognized in the database and information system communities. Nevertheless, most relational databases have no temporal support at all. A few of them provide a limited support, in terms of temporal data types and predicates, constructors, and functions for the management of time values (borrowed from the SQL standard). One (resp., two) temporal dimensions are supported by historical and transaction-time (resp., bitemporal) databases only. In this paper, we provide a relational encoding of a conceptual model featuring four temporal dimensions, namely, the classical valid and transaction times, plus the event and availability times. We focus our attention on the distinctive technical features of the proposed temporal extension of the relation model. In the last part of the paper, we briefly show how to implement it in a standard DBMS.
Headspace profiling of cocaine samples for intelligence purposes.
Dujourdy, Laurence; Besacier, Fabrice
2008-08-06
A method for determination of residual solvents in illicit hydrochloride cocaine samples using static headspace-gas chromatography (HS-GC) associated with a storage computerized procedure is described for the profiling and comparison of seizures. The system involves a gas chromatographic separation of 18 occluded solvents followed by fully automatic data analysis and transfer to a PHP/MySQL database. First, a fractional factorial design was used to evaluate the main effects of some critical method parameters (salt choice, vial agitation intensity, oven temperature, pressurization and loop equilibration) on the results with a minimum of experiments. The method was then validated for tactical intelligence purposes (batch comparison) via several studies: selection of solvents and mathematical comparison tool, reproducibility and "cutting" influence studies. The decision threshold to determine the similarity of two samples was set and false positives and negatives evaluated. Finally, application of the method to distinguish geographical origins is discussed.
BIRS - Bioterrorism Information Retrieval System.
Tewari, Ashish Kumar; Rashi; Wadhwa, Gulshan; Sharma, Sanjeev Kumar; Jain, Chakresh Kumar
2013-01-01
Bioterrorism is the intended use of pathogenic strains of microbes to widen terror in a population. There is a definite need to promote research for development of vaccines, therapeutics and diagnostic methods as a part of preparedness to any bioterror attack in the future. BIRS is an open-access database of collective information on the organisms related to bioterrorism. The architecture of database utilizes the current open-source technology viz PHP ver 5.3.19, MySQL and IIS server under windows platform for database designing. Database stores information on literature, generic- information and unique pathways of about 10 microorganisms involved in bioterrorism. This may serve as a collective repository to accelerate the drug discovery and vaccines designing process against such bioterrorist agents (microbes). The available data has been validated from various online resources and literature mining in order to provide the user with a comprehensive information system. The database is freely available at http://www.bioterrorism.biowaves.org.
Automated ocean color product validation for the Southern California Bight
NASA Astrophysics Data System (ADS)
Davis, Curtiss O.; Tufillaro, Nicholas; Jones, Burt; Arnone, Robert
2012-06-01
Automated match ups allow us to maintain and improve the products of current satellite ocean color sensors (MODIS, MERIS), and new sensors (VIIRS). As part of the VIIRS mission preparation, we have created a web based automated match up tool that provides access to searchable fields for date, site, and products, and creates match-ups between satellite (MODIS, MERIS, VIIRS), and in-situ measurements (HyperPRO and SeaPRISM). The back end of the system is a 'mySQL' database, and the front end is a `php' web portal with pull down menus for searchable fields. Based on selections, graphics are generated showing match-ups and statistics, and ascii files are created for downloads for the matchup data. Examples are shown for matching the satellite data with the data from Platform Eureka SeaPRISM off L.A. Harbor in the Southern California Bight.
Prahs, Philipp; Radeck, Viola; Mayer, Christian; Cvetkov, Yordan; Cvetkova, Nadezhda; Helbig, Horst; Märker, David
2018-01-01
Intravitreal injections with anti-vascular endothelial growth factor (anti-VEGF) medications have become the standard of care for their respective indications. Optical coherence tomography (OCT) scans of the central retina provide detailed anatomical data and are widely used by clinicians in the decision-making process of anti-VEGF indication. In recent years, significant progress has been made in artificial intelligence and computer vision research. We trained a deep convolutional artificial neural network to predict treatment indication based on central retinal OCT scans without human intervention. A total of 183,402 retinal OCT B-scans acquired between 2008 and 2016 were exported from the institutional image archive of a university hospital. OCT images were cross-referenced with the electronic institutional intravitreal injection records. OCT images with a following intravitreal injection during the first 21 days after image acquisition were assigned into the 'injection' group, while the same amount of random OCT images without intravitreal injections was labeled as 'no injection'. After image preprocessing, OCT images were split in a 9:1 ratio to training and test datasets. We trained a GoogLeNet inception deep convolutional neural network and assessed its performance on the validation dataset. We calculated prediction accuracy, sensitivity, specificity, and receiver operating characteristics. The deep convolutional neural network was successfully trained on the extracted clinical data. The trained neural network classifier reached a prediction accuracy of 95.5% on the images in the validation dataset. For single retinal B-scans in the validation dataset, a sensitivity of 90.1% and a specificity of 96.2% were achieved. The area under the receiver operating characteristic curve was 0.968 on a per B-scan image basis, and 0.988 by averaging over six B-scans per examination on the validation dataset. Deep artificial neural networks show impressive performance on classification of retinal OCT scans. After training on historical clinical data, machine learning methods can offer the clinician support in the decision-making process. Care should be taken not to mistake neural network output as treatment recommendation and to ensure a final thorough evaluation by the treating physician.
Acoustic emission of rock mass under the constant-rate fluid injection
NASA Astrophysics Data System (ADS)
Shadrin Klishin, AV, VI
2018-03-01
The authors study acoustic emission in coal bed and difficult-to-cave roof under injection of fluid by pumps at a constant rate. The functional connection between the roof hydrofracture length and the total number of AE pulses is validated, it is also found that the coal bed hydroloosening time, injection rate and time behavior of acoustic emission activity depend on the fluid injection volume required until the fluid breakout in a roadway through growing fractures. In the formulas offered for the practical application, integral parameters that characterize permeability and porosity of rock mass and process parameters of the technology are found during test injection.
Induced seismicity provides insight into why earthquake ruptures stop.
Galis, Martin; Ampuero, Jean Paul; Mai, P Martin; Cappa, Frédéric
2017-12-01
Injection-induced earthquakes pose a serious seismic hazard but also offer an opportunity to gain insight into earthquake physics. Currently used models relating the maximum magnitude of injection-induced earthquakes to injection parameters do not incorporate rupture physics. We develop theoretical estimates, validated by simulations, of the size of ruptures induced by localized pore-pressure perturbations and propagating on prestressed faults. Our model accounts for ruptures growing beyond the perturbed area and distinguishes self-arrested from runaway ruptures. We develop a theoretical scaling relation between the largest magnitude of self-arrested earthquakes and the injected volume and find it consistent with observed maximum magnitudes of injection-induced earthquakes over a broad range of injected volumes, suggesting that, although runaway ruptures are possible, most injection-induced events so far have been self-arrested ruptures.
Nadkarni, P M
1997-08-01
Concept Locator (CL) is a client-server application that accesses a Sybase relational database server containing a subset of the UMLS Metathesaurus for the purpose of retrieval of concepts corresponding to one or more query expressions supplied to it. CL's query grammar permits complex Boolean expressions, wildcard patterns, and parenthesized (nested) subexpressions. CL translates the query expressions supplied to it into one or more SQL statements that actually perform the retrieval. The generated SQL is optimized by the client to take advantage of the strengths of the server's query optimizer, and sidesteps its weaknesses, so that execution is reasonably efficient.
Astronomical Data Processing Using SciQL, an SQL Based Query Language for Array Data
NASA Astrophysics Data System (ADS)
Zhang, Y.; Scheers, B.; Kersten, M.; Ivanova, M.; Nes, N.
2012-09-01
SciQL (pronounced as ‘cycle’) is a novel SQL-based array query language for scientific applications with both tables and arrays as first class citizens. SciQL lowers the entrance fee of adopting relational DBMS (RDBMS) in scientific domains, because it includes functionality often only found in mathematics software packages. In this paper, we demonstrate the usefulness of SciQL for astronomical data processing using examples from the Transient Key Project of the LOFAR radio telescope. In particular, how the LOFAR light-curve database of all detected sources can be constructed, by correlating sources across the spatial, frequency, time and polarisation domains.
GEOMAGIA50: An archeointensity database with PHP and MySQL
NASA Astrophysics Data System (ADS)
Korhonen, K.; Donadini, F.; Riisager, P.; Pesonen, L. J.
2008-04-01
The GEOMAGIA50 database stores 3798 archeomagnetic and paleomagnetic intensity determinations dated to the past 50,000 years. It also stores details of the measurement setup for each determination, which are used for ranking the data according to prescribed reliability criteria. The ranking system aims to alleviate the data reliability problem inherent in this kind of data. GEOMAGIA50 is based on two popular open source technologies. The MySQL database management system is used for storing the data, whereas the functionality and user interface are provided by server-side PHP scripts. This technical brief gives a detailed description of GEOMAGIA50 from a technical viewpoint.
The MAO NASU Plate Archive Database. Current Status and Perspectives
NASA Astrophysics Data System (ADS)
Pakuliak, L. K.; Sergeeva, T. P.
2006-04-01
The preliminary online version of the database of the MAO NASU plate archive is constructed on the basis of the relational database management system MySQL and permits an easy supplement of database with new collections of astronegatives, provides a high flexibility in constructing SQL-queries for data search optimization, PHP Basic Authorization protected access to administrative interface and wide range of search parameters. The current status of the database will be reported and the brief description of the search engine and means of the database integrity support will be given. Methods and means of the data verification and tasks for the further development will be discussed.
Using career ladders to motivate and retain employees: an implementation success story.
Garletts, Joseph A
2002-01-01
In October 2000, Phoenix-based Sonora Quest Laboratories, LLC (SQL), commissioned The Gelfond Group to survey SQL employees. Responding to negative survey scores, SQL developed and implemented an entry-level career ladder for line staff of the specimen management/referral testing department. The program was piloted in February 2001, and was implemented fully shortly thereafter. The ladder was designed to provide job enrichment opportunities through company-conducted training and advancement provisions. It contained requirements for productivity and quality of work performed in addition to increasingly rigorous training and competency documentation. Employees were accountable for their own advancement and for ensuring that all documentation was complete. Advancement was automatic once requirements were completed. Pay increases accompanied each advancement on a predetermined scale. At the end of 12 months, employee turnover dropped from 39% to less than 20% annually. Both productivity and morale improved, and results on a second employee survey indicated dramatic improvement in five key areas. The career ladder concept has been replicated successfully in several other departments, including phlebotomy, and a six-tiered ladder is under development for the clinical laboratory. It will encompass CLA, MLT, and MT positions from entry level to technical coordinator.
Mining the SDSS SkyServer SQL queries log
NASA Astrophysics Data System (ADS)
Hirota, Vitor M.; Santos, Rafael; Raddick, Jordan; Thakar, Ani
2016-05-01
SkyServer, the Internet portal for the Sloan Digital Sky Survey (SDSS) astronomic catalog, provides a set of tools that allows data access for astronomers and scientific education. One of SkyServer data access interfaces allows users to enter ad-hoc SQL statements to query the catalog. SkyServer also presents some template queries that can be used as basis for more complex queries. This interface has logged over 330 million queries submitted since 2001. It is expected that analysis of this data can be used to investigate usage patterns, identify potential new classes of queries, find similar queries, etc. and to shed some light on how users interact with the Sloan Digital Sky Survey data and how scientists have adopted the new paradigm of e-Science, which could in turn lead to enhancements on the user interfaces and experience in general. In this paper we review some approaches to SQL query mining, apply the traditional techniques used in the literature and present lessons learned, namely, that the general text mining approach for feature extraction and clustering does not seem to be adequate for this type of data, and, most importantly, we find that this type of analysis can result in very different queries being clustered together.
NVST Data Archiving System Based On FastBit NoSQL Database
NASA Astrophysics Data System (ADS)
Liu, Ying-bo; Wang, Feng; Ji, Kai-fan; Deng, Hui; Dai, Wei; Liang, Bo
2014-06-01
The New Vacuum Solar Telescope (NVST) is a 1-meter vacuum solar telescope that aims to observe the fine structures of active regions on the Sun. The main tasks of the NVST are high resolution imaging and spectral observations, including the measurements of the solar magnetic field. The NVST has been collecting more than 20 million FITS files since it began routine observations in 2012 and produces a maximum observational records of 120 thousand files in a day. Given the large amount of files, the effective archiving and retrieval of files becomes a critical and urgent problem. In this study, we implement a new data archiving system for the NVST based on the Fastbit Not Only Structured Query Language (NoSQL) database. Comparing to the relational database (i.e., MySQL; My Structured Query Language), the Fastbit database manifests distinctive advantages on indexing and querying performance. In a large scale database of 40 million records, the multi-field combined query response time of Fastbit database is about 15 times faster and fully meets the requirements of the NVST. Our study brings a new idea for massive astronomical data archiving and would contribute to the design of data management systems for other astronomical telescopes.
Integration in PACS of DICOM with TCP/IP, SQL, and X Windows
NASA Astrophysics Data System (ADS)
Reijns, Gerard L.
1994-05-01
The DICOM standard (Digital Imaging and Communications in Medicine) has been developed in order to obtain compatibility at the higher OSI levels. This paper describes the implementation of DICOM into our developed low cost PACS, which uses as much as possible standard software and standard protocols such as SQL, X Windows and TCP/IP. We adopted the requirement that all messages on the communication network have to be DICOM compatible. The translation between DICOM messages and SQL commands, which drive the relational data base, has been accommodated in our PACS supervisor. The translation takes only between 10 and 20 milliseconds. Images, that will be used the current day are stored in a distributed, parallel operating image base for reasons of performance. Extensive use has been made of X Windows to visualize images. A maximum of 12 images can simultaneously be displayed, of which one selected image can be manipulated (e.g., magnified, rotated, etc.), without affecting the other displayed images. The emphasis of the future work will be on performance measurements and modeling of our PACS and bringing the results of both methods in agreement with each other.
Technical Aspects of Interfacing MUMPS to an External SQL Relational Database Management System
Kuzmak, Peter M.; Walters, Richard F.; Penrod, Gail
1988-01-01
This paper describes an interface connecting InterSystems MUMPS (M/VX) to an external relational DBMS, the SYBASE Database Management System. The interface enables MUMPS to operate in a relational environment and gives the MUMPS language full access to a complete set of SQL commands. MUMPS generates SQL statements as ASCII text and sends them to the RDBMS. The RDBMS executes the statements and returns ASCII results to MUMPS. The interface suggests that the language features of MUMPS make it an attractive tool for use in the relational database environment. The approach described in this paper separates MUMPS from the relational database. Positioning the relational database outside of MUMPS promotes data sharing and permits a number of different options to be used for working with the data. Other languages like C, FORTRAN, and COBOL can access the RDBMS database. Advanced tools provided by the relational database vendor can also be used. SYBASE is an advanced high-performance transaction-oriented relational database management system for the VAX/VMS and UNIX operating systems. SYBASE is designed using a distributed open-systems architecture, and is relatively easy to interface with MUMPS.
Influence of Powder Injection Parameters in High-Pressure Cold Spray
NASA Astrophysics Data System (ADS)
Ozdemir, Ozan C.; Widener, Christian A.
2017-10-01
High-pressure cold spray systems are becoming widely accepted for use in the structural repair of surface defects of expensive machinery parts used in industrial and military equipment. The deposition quality of cold spray repairs is typically validated using coupon testing and through destructive analysis of mock-ups or first articles for a defined set of parameters. In order to provide a reliable repair, it is important to not only maintain the same processing parameters, but also to have optimum fixed parameters, such as the particle injection location. This study is intended to provide insight into the sensitivity of the way that the powder is injected upstream of supersonic nozzles in high-pressure cold spray systems and the effects of variations in injection parameters on the nature of the powder particle kinetics. Experimentally validated three-dimensional computational fluid dynamics (3D CFD) models are implemented to study the particle impact conditions for varying powder feeder tube size, powder feeder tube axial misalignment, and radial powder feeder injection location on the particle velocity and the deposition shape of aluminum alloy 6061. Outputs of the models are statistically analyzed to explore the shape of the spray plume distribution and resulting coating buildup.
Optimization of post-run corrections for water stable isotope measurements by laser spectroscopy
NASA Astrophysics Data System (ADS)
van Geldern, Robert; Barth, Johannes A. C.
2013-04-01
Light stable isotope analyses of hydrogen and oxygen of water are used in numerous aquatic studies from various scientific fields. The advantage of using stable isotope ratios is that water molecules serve as ubiquitous and already present natural tracers. Traditionally, the samples were analyzed in the laboratory by isotope ratio mass spectrometry (IRMS). Within recent years these analyses have been revolutionized by the development of new isotope ratio laser spectroscopy (IRIS) systems that are said to be cheaper, more robust and mobile compared to IRMS. Although easier to operate, laser systems also need thorough calibration with international reference materials and raw data need correction for analytical effects. A major issue in systems that use liquid injection via a vaporizer module is the memory effect, i.e. the carry-over from the previous analyzed sample in a sequence. This study presents an optimized and simple post-run correction procedure for liquid water injection developed for a Picarro water analyzer. The Excel(TM) template will rely exclusively on standard features implemented in MS Office without the need to run macros, additional code written in Visual Basic for Applications (VBA) or to use a database-related software such as MS Access or SQL Server. These protocols will maximize precision, accuracy and sample throughput via an efficient memory correction. The number of injections per unknown sample can be reduced to 4 or less. This procedure meets the demands of faster throughput with reduced costs per analysis. Procedures were verified by an international proficiency test and traditional IRMS techniques. The template is available free for scientific use from the corresponding author or the journals web site (van Geldern and Barth, 2012). References van Geldern, R. and Barth, J.A.C. (2012) Limnol. Oceanogr. Methods 10:1024-1036 [doi: 10.4319/lom.2012.10.1024
Validation of a Preclinical Spinal Safety Model: Effects of Intrathecal Morphine in the Neonatal Rat
Westin, B. David; Walker, Suellen M.; Deumens, Ronald; Grafe, Marjorie; Yaksh, Tony L.
2010-01-01
Background Preclinical studies demonstrate increased neuroapoptosis after general anesthesia in early life. Neuraxial techniques may minimize potential risks, but there has been no systematic evaluation of spinal analgesic safety in developmental models. We aimed to validate a preclinical model for evaluating dose-dependent efficacy, spinal cord toxicity, and long term function following intrathecal morphine in the neonatal rat. Methods Lumbar intrathecal injections were performed in anesthetized rats aged postnatal day (P)3, 10 and 21. The relationship between injectate volume and segmental spread was assessed post mortem and by in-vivo imaging. To determine the antinociceptive dose, mechanical withdrawal thresholds were measured at baseline and 30 minutes following intrathecal morphine. To evaluate toxicity, doses up to the maximum tolerated were administered, and spinal cord histopathology, apoptosis and glial response were evaluated 1 and 7 days following P3 or P21 injection. Sensory thresholds and gait analysis were evaluated at P35. Results Intrathecal injection can be reliably performed at all postnatal ages and injectate volume influences segmental spread. Intrathecal morphine produced spinally-mediated analgesia at all ages with lower dose requirements in younger pups. High dose intrathecal morphine did not produce signs of spinal cord toxicity or alter long-term function. Conclusions The therapeutic ratio for intrathecal morphine (toxic dose / antinociceptive dose) was at least 300 at P3, and at least 20 at P21 (latter doses limited by side effects). This data provides relative efficacy and safety data for comparison with other analgesic preparations and contributes supporting evidence for the validity of this preclinical neonatal safety model. PMID:20526189
Westin, B David; Walker, Suellen M; Deumens, Ronald; Grafe, Marjorie; Yaksh, Tony L
2010-07-01
Preclinical studies demonstrate increased neuroapoptosis after general anesthesia in early life. Neuraxial techniques may minimize potential risks, but there has been no systematic evaluation of spinal analgesic safety in developmental models. We aimed to validate a preclinical model for evaluating dose-dependent efficacy, spinal cord toxicity, and long-term function after intrathecal morphine in the neonatal rat. Lumbar intrathecal injections were performed in anesthetized rats aged postnatal day (P) 3, 10, and 21. The relationship between injectate volume and segmental spread was assessed postmortem and by in vivo imaging. To determine the antinociceptive dose, mechanical withdrawal thresholds were measured at baseline and 30 min after intrathecal morphine. To evaluate toxicity, doses up to the maximum tolerated were administered, and spinal cord histopathology, apoptosis, and glial response were evaluated 1 and 7 days after P3 or P21 injection. Sensory thresholds and gait analysis were evaluated at P35. Intrathecal injection can be reliably performed at all postnatal ages and injectate volume influences segmental spread. Intrathecal morphine produced spinally mediated analgesia at all ages with lower dose requirements in younger pups. High-dose intrathecal morphine did not produce signs of spinal cord toxicity or alter long-term function. The therapeutic ratio for intrathecal morphine (toxic dose/antinociceptive dose) was at least 300 at P3 and at least 20 at P21 (latter doses limited by side effects). These data provide relative efficacy and safety for comparison with other analgesic preparations and contribute supporting evidence for the validity of this preclinical neonatal safety model.
Lu, Jiaxi; Wang, Pengli; Wang, Qiuying; Wang, Yanan; Jiang, Miaomiao
2018-05-15
In the current study, we employed high-resolution proton and carbon nuclear magnetic resonance spectroscopy (¹H and 13 C NMR) for quantitative analysis of glycerol in drug injections without any complex pre-treatment or derivatization on samples. The established methods were validated with good specificity, linearity, accuracy, precision, stability, and repeatability. Our results revealed that the contents of glycerol were convenient to calculate directly via the integration ratios of peak areas with an internal standard in ¹H NMR spectra, while the integration of peak heights were proper for 13 C NMR in combination with an external calibration of glycerol. The developed methods were both successfully applied in drug injections. Quantitative NMR methods showed an extensive prospect for glycerol determination in various liquid samples.
Hadad, Ghada M; Abdel-Salam, Randa A; Emara, Samy
2011-12-01
Application of a sensitive and rapid flow injection analysis (FIA) method for determination of topiramate, piracetam, and levetiracetam in pharmaceutical formulations has been investigated. The method is based on the reaction with ortho-phtalaldehyde and 2-mercaptoethanol in a basic buffer and measurement of absorbance at 295 nm under flow conditions. Variables affecting the determination such as sample injection volume, pH, ionic strength, reagent concentrations, flow rate of reagent and other FIA parameters were optimized to produce the most sensitive and reproducible results using a quarter-fraction factorial design, for five factors at two levels. Also, the method has been optimized and fully validated in terms of linearity and range, limit of detection and quantitation, precision, selectivity and accuracy. The method was successfully applied to the analysis of pharmaceutical preparations.
A CFD Model for High Pressure Liquid Poison Injection for CANDU-6 Shutdown System No. 2
DOE Office of Scientific and Technical Information (OSTI.GOV)
Bo Wook Rhee; Chang Jun Jeong; Hye Jeong Yun
2002-07-01
In CANDU reactor one of the two reactor shutdown systems is the liquid poison injection system which injects the highly pressurized liquid neutron poison into the moderator tank via small holes on the nozzle pipes. To ensure the safe shutdown of a reactor it is necessary for the poison curtains generated by jets provide quick, and enough negative reactivity to the reactor during the early stage of the accident. In order to produce the neutron cross section necessary to perform this work, the poison concentration distribution during the transient is necessary. In this study, a set of models for analyzingmore » the transient poison concentration induced by this high pressure poison injection jet activated upon the reactor trip in a CANDU-6 reactor moderator tank has been developed and used to generate the poison concentration distribution of the poison curtains induced by the high pressure jets injected into the vacant region between the pressure tube banks. The poison injection rate through the jet holes drilled on the nozzle pipes is obtained by a 1-D transient hydrodynamic code called, ALITRIG, and this injection rate is used to provide the inlet boundary condition to a 3-D CFD model of the moderator tank based on CFX4.3, a CFD code, to simulate the formation of the poison jet curtain inside the moderator tank. For validation, an attempt was made to validate this model against a poison injection experiment performed at BARC. As conclusion this set of models is judged to be appropriate. (authors)« less
Beck, William; Kabiche, Sofiane; Balde, Issa-Bella; Carret, Sandra; Fontan, Jean-Eudes; Cisternino, Salvatore; Schlatter, Joël
2016-12-01
To assess the stability of pharmaceutical suxamethonium (succinylcholine) solution for injection by validated stability-indicating chromatographic method in vials stored at room temperature. The chromatographic assay was achieved by using a detector wavelength set at 218 nm, a C18 column, and an isocratic mobile phase (100% of water) at a flow rate of 0.6 mL/min for 5 minutes. The method was validated according to the International Conference on Harmonization guidelines with respect to the stability-indicating capacity of the method including linearity, limits of detection and quantitation, precision, accuracy, system suitability, robustness, and forced degradations. Linearity was achieved in the concentration range of 5 to 40 mg/mL with a correlation coefficient higher than 0.999. The limits of detection and quantification were 0.8 and 0.9 mg/mL, respectively. The percentage relative standard deviation for intraday (1.3-1.7) and interday (0.1-2.0) precision was found to be less than 2.1%. Accuracy was assessed by the recovery test of suxamethonium from solution for injection (99.5%-101.2%). Storage of suxamethonium solution for injection vials at ambient temperature (22°C-26°C) for 17 days demonstrated that at least 95% of original suxamethonium concentration remained stable. Copyright © 2016 Elsevier Inc. All rights reserved.
Induced seismicity provides insight into why earthquake ruptures stop
Galis, Martin; Ampuero, Jean Paul; Mai, P. Martin; Cappa, Frédéric
2017-01-01
Injection-induced earthquakes pose a serious seismic hazard but also offer an opportunity to gain insight into earthquake physics. Currently used models relating the maximum magnitude of injection-induced earthquakes to injection parameters do not incorporate rupture physics. We develop theoretical estimates, validated by simulations, of the size of ruptures induced by localized pore-pressure perturbations and propagating on prestressed faults. Our model accounts for ruptures growing beyond the perturbed area and distinguishes self-arrested from runaway ruptures. We develop a theoretical scaling relation between the largest magnitude of self-arrested earthquakes and the injected volume and find it consistent with observed maximum magnitudes of injection-induced earthquakes over a broad range of injected volumes, suggesting that, although runaway ruptures are possible, most injection-induced events so far have been self-arrested ruptures. PMID:29291250
Evaluation of Sub Query Performance in SQL Server
NASA Astrophysics Data System (ADS)
Oktavia, Tanty; Sujarwo, Surya
2014-03-01
The paper explores several sub query methods used in a query and their impact on the query performance. The study uses experimental approach to evaluate the performance of each sub query methods combined with indexing strategy. The sub query methods consist of in, exists, relational operator and relational operator combined with top operator. The experimental shows that using relational operator combined with indexing strategy in sub query has greater performance compared with using same method without indexing strategy and also other methods. In summary, for application that emphasized on the performance of retrieving data from database, it better to use relational operator combined with indexing strategy. This study is done on Microsoft SQL Server 2012.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Femec, D.A.
This report describes two code-generating tools used to speed design and implementation of relational databases and user interfaces: CREATE-SCHEMA and BUILD-SCREEN. CREATE-SCHEMA produces the SQL commands that actually create and define the database. BUILD-SCREEN takes templates for data entry screens and generates the screen management system routine calls to display the desired screen. Both tools also generate the related FORTRAN declaration statements and precompiled SQL calls. Included with this report is the source code for a number of FORTRAN routines and functions used by the user interface. This code is broadly applicable to a number of different databases.
NASA Technical Reports Server (NTRS)
Papazian, Peter B.; Perala, Rodney A.; Curry, John D.; Lankford, Alan B.; Keller, J. David
1988-01-01
Using three different current injection methods and a simple voltage probe, transfer impedances for Solid Rocket Motor (SRM) joints, wire meshes, aluminum foil, Thorstrand and a graphite composite motor case were measured. In all cases, the surface current distribution for the particular current injection device was calculated analytically or by finite difference methods. The results of these calculations were used to generate a geometric factor which was the ratio of total injected current to surface current density. The results were validated in several ways. For wire mesh measurements, results showed good agreement with calculated results for a 14 by 18 Al screen. SRM joint impedances were independently verified. The filiment wound case measurement results were validated only to the extent that their curve shape agrees with the expected form of transfer impedance for a homogeneous slab excited by a plane wave source.
Michałowski, Jarosław M; Droździel, Dawid; Matuszewski, Jacek; Koziejowski, Wojtek; Jednoróg, Katarzyna; Marchewka, Artur
2017-08-01
Emotionally charged pictorial materials are frequently used in phobia research, but no existing standardized picture database is dedicated to the study of different phobias. The present work describes the results of two independent studies through which we sought to develop and validate this type of database-a Set of Fear Inducing Pictures (SFIP). In Study 1, 270 fear-relevant and 130 neutral stimuli were rated for fear, arousal, and valence by four groups of participants; small-animal (N = 34), blood/injection (N = 26), social-fearful (N = 35), and nonfearful participants (N = 22). The results from Study 1 were employed to develop the final version of the SFIP, which includes fear-relevant images of social exposure (N = 40), blood/injection (N = 80), spiders/bugs (N = 80), and angry faces (N = 30), as well as 726 neutral photographs. In Study 2, we aimed to validate the SFIP in a sample of spider, blood/injection, social-fearful, and control individuals (N = 66). The fear-relevant images were rated as being more unpleasant and led to greater fear and arousal in fearful than in nonfearful individuals. The fear images differentiated between the three fear groups in the expected directions. Overall, the present findings provide evidence for the high validity of the SFIP and confirm that the set may be successfully used in phobia research.
2012-01-01
Purpose To examine the psychometric properties of the Injection Pen Assessment Questionnaire (IPAQ) including the following: 1) item and scale characteristics (e.g., frequencies, item distributions, and factor structure), 2) reliability, and 3) validity. Methods Focus groups and one-on-one dyad interviews guided the development of the IPAQ. The IPAQ was subsequently tested in 136 parent–child dyads in a Phase 3, 2-month, open-label, multicenter trial for a new Genotropin® disposable pen. Factor analysis was performed to inform the development of a scoring algorithm, and reliability and validity of the IPAQ were evaluated using the data from this two months study. Psychometric analyses were conducted separately for each injection pen. Results Confirmatory factor analysis provides evidence supporting a second order factor solution for four subscales and a total IPAQ score. These factor analysis results support the conceptual framework developed from previous qualitative research in patient dyads using the reusable pen. However, the IPAQ subscales did not consistently meet acceptable internal consistency reliability for some group level comparisons. Cronbach’s alphas for the total IPAQ score for both pens were 0.85, exceeding acceptable levels of reliability for group comparisons. Conclusions The total IPAQ score is a useful measure for evaluating ease of use and preference for injection pens in clinical trials among patient dyads receiving hGH. The psychometric properties of the individual subscales, mainly the lower internal consistency reliability of some of the subscales and the predictive validity findings, do not support the use of subscale scores alone as a primary endpoint. PMID:23046797
NASA Astrophysics Data System (ADS)
Kumar, Ashok; Nunley, Hayden; Marino, Alberto
2016-05-01
Quantum noise reduction (QNR) below the standard quantum limit (SQL) has been a subject of interest for the past two to three decades due to its wide range of applications in quantum metrology and quantum information processing. To date, most of the attention has focused on the study of QNR in the temporal domain. However, many areas in quantum optics, specifically in quantum imaging, could benefit from QNR not only in the temporal domain but also in the spatial domain. With the use of a high quantum efficiency electron multiplier charge coupled device (EMCCD) camera, we have observed spatial QNR below the SQL in bright narrowband twin light beams generated through a four-wave mixing (FWM) process in hot rubidium atoms. Owing to momentum conservation in this process, the twin beams are momentum correlated. This leads to spatial quantum correlations and spatial QNR. Our preliminary results show a spatial QNR of over 2 dB with respect to the SQL. Unlike previous results on spatial QNR with faint and broadband photon pairs from parametric down conversion (PDC), we demonstrate spatial QNR with spectrally and spatially narrowband bright light beams. The results obtained will be useful for atom light interaction based quantum protocols and quantum imaging. Work supported by the W.M. Keck Foundation.
Find the fish: using PROC SQL to build a relational database
Fabrizio, Mary C.; Nelson, Scott N.
1995-01-01
Reliable estimates of abundance and survival, gained through mark-recapture studies, are necessary to better understand how to manage and restore lake trout populations in the Great Lakes. Working with a 24-year data set from a mark-recapture study conducted in Lake Superior, we attempted to disclose information on tag shedding by examining recaptures of double-tagged fish. The data set consisted of 64,288 observations on fish which had been marked with one or more tags; a subset of these fish had been marked with two tags at initial capture. Although DATA and PROC statements could be used to obtain some of the information we sought, these statements could not be used to extract a complete set of results from the double-tagging experiments. We therefore used SQL processing to create three tables representing the same information but in a fully normalized relational structure. In addition, we created indices to efficiently examine complex relationships among the individual capture records. This approach allowed us to obtain all the information necessary to estimate tag retention through subsequent modeling. We believe that our success with SQL was due in large part to its ability to simultaneosly scan the same table more than once and to permit consideration of other tables in sub-queries.
Database constraints applied to metabolic pathway reconstruction tools.
Vilaplana, Jordi; Solsona, Francesc; Teixido, Ivan; Usié, Anabel; Karathia, Hiren; Alves, Rui; Mateo, Jordi
2014-01-01
Our group developed two biological applications, Biblio-MetReS and Homol-MetReS, accessing the same database of organisms with annotated genes. Biblio-MetReS is a data-mining application that facilitates the reconstruction of molecular networks based on automated text-mining analysis of published scientific literature. Homol-MetReS allows functional (re)annotation of proteomes, to properly identify both the individual proteins involved in the process(es) of interest and their function. It also enables the sets of proteins involved in the process(es) in different organisms to be compared directly. The efficiency of these biological applications is directly related to the design of the shared database. We classified and analyzed the different kinds of access to the database. Based on this study, we tried to adjust and tune the configurable parameters of the database server to reach the best performance of the communication data link to/from the database system. Different database technologies were analyzed. We started the study with a public relational SQL database, MySQL. Then, the same database was implemented by a MapReduce-based database named HBase. The results indicated that the standard configuration of MySQL gives an acceptable performance for low or medium size databases. Nevertheless, tuning database parameters can greatly improve the performance and lead to very competitive runtimes.
Integrating Radar Image Data with Google Maps
NASA Technical Reports Server (NTRS)
Chapman, Bruce D.; Gibas, Sarah
2010-01-01
A public Web site has been developed as a method for displaying the multitude of radar imagery collected by NASA s Airborne Synthetic Aperture Radar (AIRSAR) instrument during its 16-year mission. Utilizing NASA s internal AIRSAR site, the new Web site features more sophisticated visualization tools that enable the general public to have access to these images. The site was originally maintained at NASA on six computers: one that held the Oracle database, two that took care of the software for the interactive map, and three that were for the Web site itself. Several tasks were involved in moving this complicated setup to just one computer. First, the AIRSAR database was migrated from Oracle to MySQL. Then the back-end of the AIRSAR Web site was updated in order to access the MySQL database. To do this, a few of the scripts needed to be modified; specifically three Perl scripts that query that database. The database connections were then updated from Oracle to MySQL, numerous syntax errors were corrected, and a query was implemented that replaced one of the stored Oracle procedures. Lastly, the interactive map was designed, implemented, and tested so that users could easily browse and access the radar imagery through the Google Maps interface.
Extending SQL to Support Privacy Policies
NASA Astrophysics Data System (ADS)
Ghazinour, Kambiz; Pun, Sampson; Majedi, Maryam; Chinaci, Amir H.; Barker, Ken
Increasing concerns over Internet applications that violate user privacy by exploiting (back-end) database vulnerabilities must be addressed to protect both customer privacy and to ensure corporate strategic assets remain trustworthy. This chapter describes an extension onto database catalogues and Structured Query Language (SQL) for supporting privacy in Internet applications, such as in social networks, e-health, e-governmcnt, etc. The idea is to introduce new predicates to SQL commands to capture common privacy requirements, such as purpose, visibility, generalization, and retention for both mandatory and discretionary access control policies. The contribution is that corporations, when creating the underlying databases, will be able to define what their mandatory privacy policies arc with which all application users have to comply. Furthermore, each application user, when providing their own data, will be able to define their own privacy policies with which other users have to comply. The extension is supported with underlying catalogues and algorithms. The experiments demonstrate a very reasonable overhead for the extension. The result is a low-cost mechanism to create new systems that arc privacy aware and also to transform legacy databases to their privacy-preserving equivalents. Although the examples arc from social networks, one can apply the results to data security and user privacy of other enterprises as well.
NASA Technical Reports Server (NTRS)
Somsel, James P.
1998-01-01
The development of a water injected Orbital Combustion Process (OCP) engine was conducted to assess the viability of using the powerplant for high altitude NASA aircraft and General Aviation (GA) applications. An OCP direct fuel injected, 1.2 liter, three cylinder, two-stroke engine has been enhanced to independently inject water directly into the combustion chamber. The engine currently demonstrates low brake specific fuel consumption capability and an excellent power to weight ratio. With direct water injection, significant improvements can be made to engine power, to knock limits/ignition advance timing, and to engine NO(x) emissions. The principal aim of the testing was to validate a cyclic model developed by the Systems Analysis Branch at NASA Ames Research Center. The work is a continuation of Ames' investigations into a Phased Direct Fuel Injection Engine with High Pressure Charging (PDFI-ITPC).
An Array Library for Microsoft SQL Server with Astrophysical Applications
NASA Astrophysics Data System (ADS)
Dobos, L.; Szalay, A. S.; Blakeley, J.; Falck, B.; Budavári, T.; Csabai, I.
2012-09-01
Today's scientific simulations produce output on the 10-100 TB scale. This unprecedented amount of data requires data handling techniques that are beyond what is used for ordinary files. Relational database systems have been successfully used to store and process scientific data, but the new requirements constantly generate new challenges. Moving terabytes of data among servers on a timely basis is a tough problem, even with the newest high-throughput networks. Thus, moving the computations as close to the data as possible and minimizing the client-server overhead are absolutely necessary. At least data subsetting and preprocessing have to be done inside the server process. Out of the box commercial database systems perform very well in scientific applications from the prospective of data storage optimization, data retrieval, and memory management but lack basic functionality like handling scientific data structures or enabling advanced math inside the database server. The most important gap in Microsoft SQL Server is the lack of a native array data type. Fortunately, the technology exists to extend the database server with custom-written code that enables us to address these problems. We present the prototype of a custom-built extension to Microsoft SQL Server that adds array handling functionality to the database system. With our Array Library, fix-sized arrays of all basic numeric data types can be created and manipulated efficiently. Also, the library is designed to be able to be seamlessly integrated with the most common math libraries, such as BLAS, LAPACK, FFTW, etc. With the help of these libraries, complex operations, such as matrix inversions or Fourier transformations, can be done on-the-fly, from SQL code, inside the database server process. We are currently testing the prototype with two different scientific data sets: The Indra cosmological simulation will use it to store particle and density data from N-body simulations, and the Milky Way Laboratory project will use it to store galaxy simulation data.
A Customizable Dashboarding System for Watershed Model Interpretation
NASA Astrophysics Data System (ADS)
Easton, Z. M.; Collick, A.; Wagena, M. B.; Sommerlot, A.; Fuka, D.
2017-12-01
Stakeholders, including policymakers, agricultural water managers, and small farm managers, can benefit from the outputs of commonly run watershed models. However, the information that each stakeholder needs is be different. While policy makers are often interested in the broader effects that small farm management may have on a watershed during extreme events or over long periods, farmers are often interested in field specific effects at daily or seasonal period. To provide stakeholders with the ability to analyze and interpret data from large scale watershed models, we have developed a framework that can support custom exploration of the large datasets produced. For the volume of data produced by these models, SQL-based data queries are not efficient; thus, we employ a "Not Only SQL" (NO-SQL) query language, which allows data to scale in both quantity and query volumes. We demonstrate a stakeholder customizable Dashboarding system that allows stakeholders to create custom `dashboards' to summarize model output specific to their needs. Dashboarding is a dynamic and purpose-based visual interface needed to display one-to-many database linkages so that the information can be presented for a single time period or dynamically monitored over time and allows a user to quickly define focus areas of interest for their analysis. We utilize a single watershed model that is run four times daily with a combined set of climate projections, which are then indexed, and added to an ElasticSearch datastore. ElasticSearch is a NO-SQL search engine built on top of Apache Lucene, a free and open-source information retrieval software library. Aligned with the ElasticSearch project is the open source visualization and analysis system, Kibana, which we utilize for custom stakeholder dashboarding. The dashboards create a visualization of the stakeholder selected analysis and can be extended to recommend robust strategies to support decision-making.
How to maintain blood supply during computer network breakdown: a manual backup system.
Zeiler, T; Slonka, J; Bürgi, H R; Kretschmer, V
2000-12-01
Electronic data management systems using computer network systems and client/server architecture are increasingly used in laboratories and transfusion services. Severe problems arise if there is no network access to the database server and critical functions are not available. We describe a manual backup system (MBS) developed to maintain the delivery of blood products to patients in a hospital transfusion service in case of a computer network breakdown. All data are kept on a central SQL database connected to peripheral workstations in a local area network (LAN). Request entry from wards is performed via machine-readable request forms containing self-adhesive specimen labels with barcodes for test tubes. Data entry occurs on-line by bidirectional automated systems or off-line manually. One of the workstations in the laboratory contains a second SQL database which is frequently and incrementally updated. This workstation is run as a stand-alone, read-only database if the central SQL database is not available. In case of a network breakdown, the time-graded MBS is launched. Patient data, requesting ward and ordered tests/requests, are photocopied through a template from the request forms on special MBS worksheets serving as laboratory journal for manual processing and result report (a copy is left in the laboratory). As soon as the network is running again the data from the off-line period are entered into the primary SQL server. The MBS was successfully used at several occasions. The documentation of a 90-min breakdown period is presented in detail. Additional work resulted from the copy work and the belated manual data entry after restoration of the system. There was no delay in issue of blood products or result reporting. The backup system described has been proven to be simple, quick and safe to maintain urgent blood supply and distribution of laboratory results in case of unexpected network breakdown.
A generic minimization random allocation and blinding system on web.
Cai, Hongwei; Xia, Jielai; Xu, Dezhong; Gao, Donghuai; Yan, Yongping
2006-12-01
Minimization is a dynamic randomization method for clinical trials. Although recommended by many researchers, the utilization of minimization has been seldom reported in randomized trials mainly because of the controversy surrounding the validity of conventional analyses and its complexity in implementation. However, both the statistical and clinical validity of minimization were demonstrated in recent studies. Minimization random allocation system integrated with blinding function that could facilitate the implementation of this method in general clinical trials has not been reported. SYSTEM OVERVIEW: The system is a web-based random allocation system using Pocock and Simon minimization method. It also supports multiple treatment arms within a trial, multiple simultaneous trials, and blinding without further programming. This system was constructed with generic database schema design method, Pocock and Simon minimization method and blinding method. It was coded with Microsoft Visual Basic and Active Server Pages (ASP) programming languages. And all dataset were managed with a Microsoft SQL Server database. Some critical programming codes were also provided. SIMULATIONS AND RESULTS: Two clinical trials were simulated simultaneously to test the system's applicability. Not only balanced groups but also blinded allocation results were achieved in both trials. Practical considerations for minimization method, the benefits, general applicability and drawbacks of the technique implemented in this system are discussed. Promising features of the proposed system are also summarized.
NASA Astrophysics Data System (ADS)
Miza, A. T. N. A.; Shayfull, Z.; Nasir, S. M.; Fathullah, M.; Hazwan, M. H. M.
2017-09-01
In this study, Computer Aided Engineering was used for injection moulding simulation. The method of Design of experiment (DOE) was utilize according to the Latin Square orthogonal array. The relationship between the injection moulding parameters and warpage were identify based on the experimental data that used. Response Surface Methodology (RSM) was used as to validate the model accuracy. Then, the RSM and GA method were combine as to examine the optimum injection moulding process parameter. Therefore the optimisation of injection moulding is largely improve and the result shown an increasing accuracy and also reliability. The propose method by combining RSM and GA method also contribute in minimising the warpage from occur.
Ghanem, Mashhour M; Abu-Lafi, Saleh A; Hallak, Hussein O
2013-01-01
A simple, specific, accurate, and stability-indicating method was developed and validated for the quantitative determination of menadione sodium bisulfite in the injectable solution formulation. The method is based on zwitterionic hydrophilic interaction liquid chromatography (ZIC-HILIC) coupled with a photodiode array detector. The desired separation was achieved on the ZIC-HILIC column (250 mm × 4.6 mm, 5 μm) at 25°C temperature. The optimized mobile phase consisted of an isocratic solvent mixture of 200mM ammonium acetate (NH4AC) solution and acetonitrile (ACN) (20:80; v/v) pH-adjusted to 5.7 by glacial acetic acid. The mobile phase was fixed at 0.5 ml/min and the analytes were monitored at 261 nm using a photodiode array detector. The effects of the chromatographic conditions on the peak retention, peak USP tailing factor, and column efficiency were systematically optimized. Forced degradation experiments were carried out by exposing menadione sodium bisulfite standard and the injectable solution formulation to thermal, photolytic, oxidative, and acid-base hydrolytic stress conditions. The degradation products were well-resolved from the main peak and the excipients, thus proving that the method is a reliable, stability-indicating tool. The method was validated as per ICH and USP guidelines (USP34/NF29) and found to be adequate for the routine quantitative estimation of menadione sodium bisulfite in commercially available menadione sodium bisulfite injectable solution dosage forms.
Footprint Representation of Planetary Remote Sensing Data
NASA Astrophysics Data System (ADS)
Walter, S. H. G.; Gasselt, S. V.; Michael, G.; Neukum, G.
The geometric outline of remote sensing image data, the so called footprint, can be represented as a number of coordinate tuples. These polygons are associated with according attribute information such as orbit name, ground- and image resolution, solar longitude and illumination conditions to generate a powerful base for classification of planetary experiment data. Speed, handling and extended capabilites are the reasons for using geodatabases to store and access these data types. Techniques for such a spatial database of footprint data are demonstrated using the Relational Database Management System (RDBMS) PostgreSQL, spatially enabled by the PostGIS extension. Exemplary, footprints of the HRSC and OMEGA instruments, both onboard ESA's Mars Express Orbiter, are generated and connected to attribute information. The aim is to provide high-resolution footprints of the OMEGA instrument to the science community for the first time and make them available for web-based mapping applications like the "Planetary Interactive GIS-on-the-Web Analyzable Database" (PIG- WAD), produced by the USGS. Map overlays with HRSC or other instruments like MOC and THEMIS (footprint maps are already available for these instruments and can be integrated into the database) allow on-the-fly intersection and comparison as well as extended statistics of the data. Footprint polygons are generated one by one using standard software provided by the instrument teams. Attribute data is calculated and stored together with the geometric information. In the case of HRSC, the coordinates of the footprints are already available in the VICAR label of each image file. Using the VICAR RTL and PostgreSQL's libpq C library they are loaded into the database using the Well-Known Text (WKT) notation by the Open Geospatial Consortium, Inc. (OGC). For the OMEGA instrument, image data is read using IDL routines developed and distributed by the OMEGA team. Image outlines are exported together with relevant attribute data to the industry standard Shapefile format. These files are translated to a Structured Query Language (SQL) command sequence suitable for insertion into the PostGIS/PostgrSQL database using the shp2pgsql data loader provided by the PostGIS software. PostgreSQL's advanced features such as geometry types, rules, operators and functions allow complex spatial queries and on-the-fly processing of data on DBMS level e.g. generalisation of the outlines. Processing done by the DBMS, visualisation via GIS systems and utilisation for web-based applications like mapservers will be demonstrated.
An investigation of the performance of an electronic in-line pump system for diesel engines
NASA Astrophysics Data System (ADS)
Fan, Li-Yun; Zhu, Yuan-Xian; Long, Wu-Qiang; Ma, Xiu-Zhen; Xue, Ying-Ying
2008-12-01
WIT Electronic Fuel System Co., Ltd. has developed a new fuel injector, the Electronic In-line Pump (EIP) system, designed to meet China’s diesel engine emission and fuel economy regulations. It can be used on marine diesel engines and commercial vehicle engines through different EIP systems. A numerical model of the EIP system was built in the AMESim environment for the purpose of creating a design tool for engine application and system optimization. The model was used to predict key injection characteristics under different operating conditions, such as injection pressure, injection rate, and injection duration. To validate these predictions, experimental tests were conducted under the conditions that were modeled. The results were quite encouraging and in agreement with model predictions. Additional experiments were conducted to study the injection characteristics of the EIP system. These results show that injection pressure and injection quantity are insensitive to injection timing variations, this is due to the design of the constant velocity cam profile. Finally, injection quantity and pressure vs. pulse width at different cam speeds are presented, an important injection characteristic for EIP system calibration.
Gamé, Xavier; Khan, Shahid; Panicker, Jalesh N; Kalsi, Vinay; Dalton, Catherine; Elneil, Sohier; Hamid, Rizwan; Dasgupta, Prokar; Fowler, Clare J
2011-06-01
Therapy (case series). 4. What's known on the subject? and What does the study add? We know that repeated injections of botulinum toxin A are effective in treating refractory detrusor overactivity particularly in NDO. This study shows that in both NDO and IDO repeated injections of the toxin improve quality of life as assessed by three validated questionnaires. The effect is most marked after the first injection in NDO patients but thereafter similar in both groups. To compare the effect of repeated detrusor injections of botulinum toxin (BoNT-A) on health-related quality of life (HRQL) in patients with idiopathic (IDO) or neurogenic detrusor overactivity (NDO). Between 2003 and 2009, 151 patients (109 with NDO and 42 with IDO) were treated by BoNT-A (Botox®, Allergan Inc., Irvine, CA, USA). Changes in HRQL were assessed using the validated short forms of Urogenital Distress Inventory (UDI-6), the Incontinence Impact Questionnaire (IIQ-7) and EuroQOL-5D (EQ-5D) before and 4 weeks after BoNT-A. The maximum number of repeated injections was five (mean±sd, 2.8±1.05). Mean±sd follow-up was 27.49±17.01 months. The UDI-6 and IIQ-7 questionnaires showed a consistent improvement after repeated injections in both groups with detrusor overactivity. The EQ-5D was not statistically different before and after each injection in either the NDO or IDO population. After repeated injections, no statistical differences in the change on the UDI-6 and IIQ-7 scores were found between NDO and IDO, except after the first treatment, when the decrease in UDI-6 was higher in NDO than in IDO. The EQ-5D anxiety and depression subscore improved in both groups after each injection and with the number of injections. In IDO, after the second injection, no patient reported extreme anxiety or depression and, after the fourth injection, none had anxiety or depression. The inter-injection interval was shorter after the first injection in those with NDO than in IDO but was similar thereafter. Intradetrusor injections of BoNT-A improved the HRQL of both NDO and IDO patients. Although improvement in HRQL was greater and the duration of efficacy shorter in NDO patients after the first injection, there was no significant difference after subsequent injections. Mean inter-injection interval in IDO and in NDO patients was similar from the second injection onwards and improvements in HRQL score were the same. © 2010 THE AUTHORS. BJU INTERNATIONAL © 2010 BJU INTERNATIONAL.
The design and implementation of GML data management information system based on PostgreSQL
NASA Astrophysics Data System (ADS)
Zhang, Aiguo; Wu, Qunyong; Xu, Qifeng
2008-10-01
GML expresses geographic information in text, and it provides an extensible and standard way of spatial information encoding. At the present time, the management of GML data is in terms of document. By this way, the inquiry and update of GML data is inefficient, and it demands high memory when the document is comparatively large. In this respect, the paper put forward a data management of GML based on PostgreSQL. It designs four kinds of inquiries, which are inquiry of metadata, inquiry of geometry based on property, inquiry of property based on spatial information, and inquiry of spatial data based on location. At the same time, it designs and implements the visualization of the inquired WKT data.
A Medical Image Backup Architecture Based on a NoSQL Database and Cloud Computing Services.
Santos Simões de Almeida, Luan Henrique; Costa Oliveira, Marcelo
2015-01-01
The use of digital systems for storing medical images generates a huge volume of data. Digital images are commonly stored and managed on a Picture Archiving and Communication System (PACS), under the DICOM standard. However, PACS is limited because it is strongly dependent on the server's physical space. Alternatively, Cloud Computing arises as an extensive, low cost, and reconfigurable resource. However, medical images contain patient information that can not be made available in a public cloud. Therefore, a mechanism to anonymize these images is needed. This poster presents a solution for this issue by taking digital images from PACS, converting the information contained in each image file to a NoSQL database, and using cloud computing to store digital images.
Monitoring SLAC High Performance UNIX Computing Systems
DOE Office of Scientific and Technical Information (OSTI.GOV)
Lettsome, Annette K.; /Bethune-Cookman Coll. /SLAC
2005-12-15
Knowledge of the effectiveness and efficiency of computers is important when working with high performance systems. The monitoring of such systems is advantageous in order to foresee possible misfortunes or system failures. Ganglia is a software system designed for high performance computing systems to retrieve specific monitoring information. An alternative storage facility for Ganglia's collected data is needed since its default storage system, the round-robin database (RRD), struggles with data integrity. The creation of a script-driven MySQL database solves this dilemma. This paper describes the process took in the creation and implementation of the MySQL database for use by Ganglia.more » Comparisons between data storage by both databases are made using gnuplot and Ganglia's real-time graphical user interface.« less
Operationalizing Semantic Medline for meeting the information needs at point of care.
Rastegar-Mojarad, Majid; Li, Dingcheng; Liu, Hongfang
2015-01-01
Scientific literature is one of the popular resources for providing decision support at point of care. It is highly desirable to bring the most relevant literature to support the evidence-based clinical decision making process. Motivated by the recent advance in semantically enhanced information retrieval, we have developed a system, which aims to bring semantically enriched literature, Semantic Medline, to meet the information needs at point of care. This study reports our work towards operationalizing the system for real time use. We demonstrate that the migration of a relational database implementation to a NoSQL (Not only SQL) implementation significantly improves the performance and makes the use of Semantic Medline at point of care decision support possible.
MaGnET: Malaria Genome Exploration Tool.
Sharman, Joanna L; Gerloff, Dietlind L
2013-09-15
The Malaria Genome Exploration Tool (MaGnET) is a software tool enabling intuitive 'exploration-style' visualization of functional genomics data relating to the malaria parasite, Plasmodium falciparum. MaGnET provides innovative integrated graphic displays for different datasets, including genomic location of genes, mRNA expression data, protein-protein interactions and more. Any selection of genes to explore made by the user is easily carried over between the different viewers for different datasets, and can be changed interactively at any point (without returning to a search). Free online use (Java Web Start) or download (Java application archive and MySQL database; requires local MySQL installation) at http://malariagenomeexplorer.org joanna.sharman@ed.ac.uk or dgerloff@ffame.org Supplementary data are available at Bioinformatics online.
Operationalizing Semantic Medline for meeting the information needs at point of care
Rastegar-Mojarad, Majid; Li, Dingcheng; Liu, Hongfang
2015-01-01
Scientific literature is one of the popular resources for providing decision support at point of care. It is highly desirable to bring the most relevant literature to support the evidence-based clinical decision making process. Motivated by the recent advance in semantically enhanced information retrieval, we have developed a system, which aims to bring semantically enriched literature, Semantic Medline, to meet the information needs at point of care. This study reports our work towards operationalizing the system for real time use. We demonstrate that the migration of a relational database implementation to a NoSQL (Not only SQL) implementation significantly improves the performance and makes the use of Semantic Medline at point of care decision support possible. PMID:26306259
MySQL/PHP web database applications for IPAC proposal submission
NASA Astrophysics Data System (ADS)
Crane, Megan K.; Storrie-Lombardi, Lisa J.; Silbermann, Nancy A.; Rebull, Luisa M.
2008-07-01
The Infrared Processing and Analysis Center (IPAC) is NASA's multi-mission center of expertise for long-wavelength astrophysics. Proposals for various IPAC missions and programs are ingested via MySQL/PHP web database applications. Proposers use web forms to enter coversheet information and upload PDF files related to the proposal. Upon proposal submission, a unique directory is created on the webserver into which all of the uploaded files are placed. The coversheet information is converted into a PDF file using a PHP extension called FPDF. The files are concatenated into one PDF file using the command-line tool pdftk and then forwarded to the review committee. This work was performed at the California Institute of Technology under contract to the National Aeronautics and Space Administration.
Injector Design Tool Improvements: User's manual for FDNS V.4.5
NASA Technical Reports Server (NTRS)
Chen, Yen-Sen; Shang, Huan-Min; Wei, Hong; Liu, Jiwen
1998-01-01
The major emphasis of the current effort is in the development and validation of an efficient parallel machine computational model, based on the FDNS code, to analyze the fluid dynamics of a wide variety of liquid jet configurations for general liquid rocket engine injection system applications. This model includes physical models for droplet atomization, breakup/coalescence, evaporation, turbulence mixing and gas-phase combustion. Benchmark validation cases for liquid rocket engine chamber combustion conditions will be performed for model validation purpose. Test cases may include shear coaxial, swirl coaxial and impinging injection systems with combinations LOXIH2 or LOXISP-1 propellant injector elements used in rocket engine designs. As a final goal of this project, a well tested parallel CFD performance methodology together with a user's operation description in a final technical report will be reported at the end of the proposed research effort.
Implementation of an Enterprise Information Portal (EIP) in the Loyola University Health System
Price, Ronald N.; Hernandez, Kim
2001-01-01
Loyola University Chicago Stritch School of Medicine and Loyola University Medical Center have long histories in the development of applications to support the institutions' missions of education, research and clinical care. In late 1998, the institutions' application development group undertook an ambitious program to re-architecture more than 10 years of legacy application development (30+ core applications) into a unified World Wide Web (WWW) environment. The primary project objectives were to construct an environment that would support the rapid development of n-tier, web-based applications while providing standard methods for user authentication/validation, security/access control and definition of a user's organizational context. The project's efforts resulted in Loyola's Enterprise Information Portal (EIP), which meets the aforementioned objectives. This environment: 1) allows access to other vertical Intranet portals (e.g., electronic medical record, patient satisfaction information and faculty effort); 2) supports end-user desktop customization; and 3) provides a means for standardized application “look and feel.” The portal was constructed utilizing readily available hardware and software. Server hardware consists of multiprocessor (Intel Pentium 500Mhz) Compaq 6500 servers with one gigabyte of random access memory and 75 gigabytes of hard disk storage. Microsoft SQL Server was selected to house the portal's internal or security data structures. Netscape Enterprise Server was selected for the web server component of the environment and Allaire's ColdFusion was chosen for access and application tiers. Total costs for the portal environment was less than $40,000. User data storage is accomplished through two Microsoft SQL Servers and an existing SUN Microsystems enterprise server with eight processors, 750 gigabytes of disk storage operating Sybase relational database manager. Total storage capacity for all system exceeds one terabyte. In the past 12 months, the EIP has supported development of more than 88 applications and is utilized by more than 2,200 users.
DRUMS: Disk Repository with Update Management and Select option for high throughput sequencing data.
Nettling, Martin; Thieme, Nils; Both, Andreas; Grosse, Ivo
2014-02-04
New technologies for analyzing biological samples, like next generation sequencing, are producing a growing amount of data together with quality scores. Moreover, software tools (e.g., for mapping sequence reads), calculating transcription factor binding probabilities, estimating epigenetic modification enriched regions or determining single nucleotide polymorphism increase this amount of position-specific DNA-related data even further. Hence, requesting data becomes challenging and expensive and is often implemented using specialised hardware. In addition, picking specific data as fast as possible becomes increasingly important in many fields of science. The general problem of handling big data sets was addressed by developing specialized databases like HBase, HyperTable or Cassandra. However, these database solutions require also specialized or distributed hardware leading to expensive investments. To the best of our knowledge, there is no database capable of (i) storing billions of position-specific DNA-related records, (ii) performing fast and resource saving requests, and (iii) running on a single standard computer hardware. Here, we present DRUMS (Disk Repository with Update Management and Select option), satisfying demands (i)-(iii). It tackles the weaknesses of traditional databases while handling position-specific DNA-related data in an efficient manner. DRUMS is capable of storing up to billions of records. Moreover, it focuses on optimizing relating single lookups as range request, which are needed permanently for computations in bioinformatics. To validate the power of DRUMS, we compare it to the widely used MySQL database. The test setting considers two biological data sets. We use standard desktop hardware as test environment. DRUMS outperforms MySQL in writing and reading records by a factor of two up to a factor of 10000. Furthermore, it can work with significantly larger data sets. Our work focuses on mid-sized data sets up to several billion records without requiring cluster technology. Storing position-specific data is a general problem and the concept we present here is a generalized approach. Hence, it can be easily applied to other fields of bioinformatics.
MO-F-CAMPUS-T-05: SQL Database Queries to Determine Treatment Planning Resource Usage
DOE Office of Scientific and Technical Information (OSTI.GOV)
Fox, C; Gladstone, D
2015-06-15
Purpose: A radiation oncology clinic’s treatment capacity is traditionally thought to be limited by the number of machines in the clinic. As the number of fractions per course decrease and the number of adaptive plans increase, the question of how many treatment plans a clinic can plan becomes increasingly important. This work seeks to lay the ground work for assessing treatment planning resource usage. Methods: Care path templates were created using the Aria 11 care path interface. Care path tasks included key steps in the treatment planning process from the completion of CT simulation through the first radiation treatment. SQLmore » Server Management Studio was used to run SQL queries to extract task completion time stamps along with care path template information and diagnosis codes from the Aria database. 6 months of planning cycles were evaluated. Elapsed time was evaluated in terms of work hours within Monday – Friday, 7am to 5pm. Results: For the 195 validated treatment planning cycles, the average time for planning and MD review was 22.8 hours. Of those cases 33 were categorized as urgent. The average planning time for urgent plans was 5 hours. A strong correlation between diagnosis code and range of elapsed planning time was as well as between elapsed time and select diagnosis codes was observed. It was also observed that tasks were more likely to be completed on the date due than the time that they were due. Follow-up confirmed that most users did not look at the due time. Conclusion: Evaluation of elapsed planning time and other tasks suggest that care paths should be adjusted to allow for different contouring and planning times for certain diagnosis codes and urgent cases. Additional clinic training around task due times vs dates or a structuring of care paths around due dates is also needed.« less
ERIC Educational Resources Information Center
Homem, Vera; Alves, Arminda; Santos, Lu´cia
2014-01-01
A laboratory application with a strong component in analytical chemistry was designed for undergraduate students, in order to introduce a current problem in the environmental science field, the water contamination by antibiotics. Therefore, a simple and rapid method based on direct injection and high performance liquid chromatography-tandem mass…
A model for 3-D sonic/supersonic transverse fuel injection into a supersonic air stream
NASA Technical Reports Server (NTRS)
Bussing, Thomas R. A.; Lidstone, Gary L.
1989-01-01
A model for sonic/supersonic transverse fuel injection into a supersonic airstream is proposed. The model replaces the hydrogen jet up to the Mach disk plane and the elliptic parts of the air flow field around the jet by an equivalent body. The main features of the model were validated on the basis of experimental data.
In-injection port thermal desorption for explosives trace evidence analysis.
Sigman, M E; Ma, C Y
1999-10-01
A gas chromatographic method utilizing thermal desorption of a dry surface wipe for the analysis of explosives trace chemical evidence has been developed and validated using electron capture and negative ion chemical ionization mass spectrometric detection. Thermal desorption was performed within a split/splitless injection port with minimal instrument modification. Surface-abraded Teflon tubing provided the solid support for sample collection and desorption. Performance was characterized by desorption efficiency, reproducibility, linearity of the calibration, and method detection and quantitation limits. Method validation was performed with a series of dinitrotoluenes, trinitrotoluene, two nitroester explosives, and one nitramine explosive. The method was applied to the sampling of a single piece of debris from an explosion containing trinitrotoluene.
Automatic helmet-wearing detection for law enforcement using CCTV cameras
NASA Astrophysics Data System (ADS)
Wonghabut, P.; Kumphong, J.; Satiennam, T.; Ung-arunyawee, R.; Leelapatra, W.
2018-04-01
The objective of this research is to develop an application for enforcing helmet wearing using CCTV cameras. The developed application aims to help law enforcement by police, and eventually resulting in changing risk behaviours and consequently reducing the number of accidents and its severity. Conceptually, the application software implemented using C++ language and OpenCV library uses two different angle of view CCTV cameras. Video frames recorded by the wide-angle CCTV camera are used to detect motorcyclists. If any motorcyclist without helmet is found, then the zoomed (narrow-angle) CCTV is activated to capture image of the violating motorcyclist and the motorcycle license plate in real time. Captured images are managed by database implemented using MySQL for ticket issuing. The results show that the developed program is able to detect 81% of motorcyclists on various motorcycle types during daytime and night-time. The validation results reveal that the program achieves 74% accuracy in detecting the motorcyclist without helmet.
STILTS -- Starlink Tables Infrastructure Library Tool Set
NASA Astrophysics Data System (ADS)
Taylor, Mark
STILTS is a set of command-line tools for processing tabular data. It has been designed for, but is not restricted to, use on astronomical data such as source catalogues. It contains both generic (format-independent) table processing tools and tools for processing VOTable documents. Facilities offered include crossmatching, format conversion, format validation, column calculation and rearrangement, row selection, sorting, plotting, statistical calculations and metadata display. Calculations on cell data can be performed using a powerful and extensible expression language. The package is written in pure Java and based on STIL, the Starlink Tables Infrastructure Library. This gives it high portability, support for many data formats (including FITS, VOTable, text-based formats and SQL databases), extensibility and scalability. Where possible the tools are written to accept streamed data so the size of tables which can be processed is not limited by available memory. As well as the tutorial and reference information in this document, detailed on-line help is available from the tools themselves. STILTS is available under the GNU General Public Licence.
BIRS – Bioterrorism Information Retrieval System
Tewari, Ashish Kumar; Rashi; Wadhwa, Gulshan; Sharma, Sanjeev Kumar; Jain, Chakresh Kumar
2013-01-01
Bioterrorism is the intended use of pathogenic strains of microbes to widen terror in a population. There is a definite need to promote research for development of vaccines, therapeutics and diagnostic methods as a part of preparedness to any bioterror attack in the future. BIRS is an open-access database of collective information on the organisms related to bioterrorism. The architecture of database utilizes the current open-source technology viz PHP ver 5.3.19, MySQL and IIS server under windows platform for database designing. Database stores information on literature, generic- information and unique pathways of about 10 microorganisms involved in bioterrorism. This may serve as a collective repository to accelerate the drug discovery and vaccines designing process against such bioterrorist agents (microbes). The available data has been validated from various online resources and literature mining in order to provide the user with a comprehensive information system. Availability The database is freely available at http://www.bioterrorism.biowaves.org PMID:23390356
NASA Astrophysics Data System (ADS)
Li, Lang-quan; Huang, Wei; Yan, Li; Li, Shi-bin
2017-10-01
The dual transverse injection system with a front hydrogen porthole and a rear air porthole arranged in tandem is proposed, and this is a realistic approach for mixing enhancement and penetration improvement of transverse injection in a scramjet combustor. The influence of this dual transverse injection system on mixing characteristics has been evaluated numerically based on grid independency analysis and code validation. The numerical approach employed in the current study has been validated against the available experimental data in the open literature, and the predicted wall static pressure distributions show reasonable agreement with the experimental data for the cases with different jet-to-crossflow pressure ratios. The obtained results predicted by the three-dimensional Reynolds-average Navier - Stokes (RANS) equations coupled with the two equation k-ω shear stress transport (SST) turbulence model show that the air pothole has an great impact on penetration depth and mixing efficiency, and the effect of air jet on flow field varies with different values of the aspect ratio. The air porthole with larger aspect ratio can increase the fuel penetration depth. However, when the aspect ratio is relatively small, the fuel penetration depth decreases, and even smaller than that of the single injection system. At the same time, the air pothole has a highly remarkable improvement on mixing efficiency, especially in the near field. The smaller the aspect ratio of the air porthole is, the higher the mixing efficiency in the near field is. This is due to its larger circulation in the near field. The dual injection system owns more losses of stagnation pressure than the single injection system.
Flux-Level Transit Injection Experiments with NASA Pleiades Supercomputer
NASA Astrophysics Data System (ADS)
Li, Jie; Burke, Christopher J.; Catanzarite, Joseph; Seader, Shawn; Haas, Michael R.; Batalha, Natalie; Henze, Christopher; Christiansen, Jessie; Kepler Project, NASA Advanced Supercomputing Division
2016-06-01
Flux-Level Transit Injection (FLTI) experiments are executed with NASA's Pleiades supercomputer for the Kepler Mission. The latest release (9.3, January 2016) of the Kepler Science Operations Center Pipeline is used in the FLTI experiments. Their purpose is to validate the Analytic Completeness Model (ACM), which can be computed for all Kepler target stars, thereby enabling exoplanet occurrence rate studies. Pleiades, a facility of NASA's Advanced Supercomputing Division, is one of the world's most powerful supercomputers and represents NASA's state-of-the-art technology. We discuss the details of implementing the FLTI experiments on the Pleiades supercomputer. For example, taking into account that ~16 injections are generated by one core of the Pleiades processors in an hour, the “shallow” FLTI experiment, in which ~2000 injections are required per target star, can be done for 16% of all Kepler target stars in about 200 hours. Stripping down the transit search to bare bones, i.e. only searching adjacent high/low periods at high/low pulse durations, makes the computationally intensive FLTI experiments affordable. The design of the FLTI experiments and the analysis of the resulting data are presented in “Validating an Analytic Completeness Model for Kepler Target Stars Based on Flux-level Transit Injection Experiments” by Catanzarite et al. (#2494058).Kepler was selected as the 10th mission of the Discovery Program. Funding for the Kepler Mission has been provided by the NASA Science Mission Directorate.
Yang, Jie; Sun, Zhi; Li, Duolu; Duan, Fei; Li, Zhuolun; Lu, Jingli; Shi, Yingying; Xu, Tanye; Zhang, Xiaojian
2018-06-07
Shenkang injection is a traditional Chinese formula with good curative effect on chronic renal failure. In this paper, a novel, rapid and sensitive ultra high performance liquid chromatography coupled with Q Exactive hybrid quadrupole orbitrap high resolution accurate mass spectrometry was developed and validated for simultaneous determination of seven bioactive constituents of Shenkang injection in rat plasma and tissues after intravenous administration. Acetonitrile was used as protein precipitation agent in biological samples dispose with carbamazepine as internal standard. The chromatographic separation was carried out on a C 18 column with a gradient mobile phase consisting of acetonitrile and water (containing 0.1% formic acid). The MS analysis was performed in the full scan positive and negative ion mode. The lower limits of quantification for the seven analytes in rat plasma and tissues were 0.1-10 ng/mL. The validated method was successfully applied to tissue distribution and pharmacokinetic studies of Shenkang injection after intravenous administration. The results of the tissue distribution study showed that the high concentration of seven constituents were primarily in the kidney tract. This is the first time to report the application of Q-Orbitrap with full scan mass spectrometry in tissue distribution and pharmacokinetic studies of Shenkang injection. This article is protected by copyright. All rights reserved.
Energy optimization aspects by injection process technology
NASA Astrophysics Data System (ADS)
Tulbure, A.; Ciortea, M.; Hutanu, C.; Farcas, V.
2016-08-01
In the proposed paper, the authors examine the energy aspects related to the injection moulding process technology in the automotive industry. Theoretical considerations have been validated by experimental measurements on the manufacturing process, for two types of injections moulding machines, hydraulic and electric. Practical measurements have been taken with professional equipment separately on each technological operation: lamination, compression, injection and expansion. For results traceability, the following parameters were, whenever possible, maintained: cycle time, product weight and the relative time. The aim of the investigations was to carry out a professional energy audit with accurate losses identification. Base on technological diagram for each production cycle, at the end of this contribution, some measure to reduce the energy consumption were proposed.
FPGA-based multiprocessor system for injection molding control.
Muñoz-Barron, Benigno; Morales-Velazquez, Luis; Romero-Troncoso, Rene J; Rodriguez-Donate, Carlos; Trejo-Hernandez, Miguel; Benitez-Rangel, Juan P; Osornio-Rios, Roque A
2012-10-18
The plastic industry is a very important manufacturing sector and injection molding is a widely used forming method in that industry. The contribution of this work is the development of a strategy to retrofit control of an injection molding machine based on an embedded system microprocessors sensor network on a field programmable gate array (FPGA) device. Six types of embedded processors are included in the system: a smart-sensor processor, a micro fuzzy logic controller, a programmable logic controller, a system manager, an IO processor and a communication processor. Temperature, pressure and position are controlled by the proposed system and experimentation results show its feasibility and robustness. As validation of the present work, a particular sample was successfully injected.
Database Constraints Applied to Metabolic Pathway Reconstruction Tools
Vilaplana, Jordi; Solsona, Francesc; Teixido, Ivan; Usié, Anabel; Karathia, Hiren; Alves, Rui; Mateo, Jordi
2014-01-01
Our group developed two biological applications, Biblio-MetReS and Homol-MetReS, accessing the same database of organisms with annotated genes. Biblio-MetReS is a data-mining application that facilitates the reconstruction of molecular networks based on automated text-mining analysis of published scientific literature. Homol-MetReS allows functional (re)annotation of proteomes, to properly identify both the individual proteins involved in the process(es) of interest and their function. It also enables the sets of proteins involved in the process(es) in different organisms to be compared directly. The efficiency of these biological applications is directly related to the design of the shared database. We classified and analyzed the different kinds of access to the database. Based on this study, we tried to adjust and tune the configurable parameters of the database server to reach the best performance of the communication data link to/from the database system. Different database technologies were analyzed. We started the study with a public relational SQL database, MySQL. Then, the same database was implemented by a MapReduce-based database named HBase. The results indicated that the standard configuration of MySQL gives an acceptable performance for low or medium size databases. Nevertheless, tuning database parameters can greatly improve the performance and lead to very competitive runtimes. PMID:25202745
Database Reports Over the Internet
NASA Technical Reports Server (NTRS)
Smith, Dean Lance
2002-01-01
Most of the summer was spent developing software that would permit existing test report forms to be printed over the web on a printer that is supported by Adobe Acrobat Reader. The data is stored in a DBMS (Data Base Management System). The client asks for the information from the database using an HTML (Hyper Text Markup Language) form in a web browser. JavaScript is used with the forms to assist the user and verify the integrity of the entered data. Queries to a database are made in SQL (Sequential Query Language), a widely supported standard for making queries to databases. Java servlets, programs written in the Java programming language running under the control of network server software, interrogate the database and complete a PDF form template kept in a file. The completed report is sent to the browser requesting the report. Some errors are sent to the browser in an HTML web page, others are reported to the server. Access to the databases was restricted since the data are being transported to new DBMS software that will run on new hardware. However, the SQL queries were made to Microsoft Access, a DBMS that is available on most PCs (Personal Computers). Access does support the SQL commands that were used, and a database was created with Access that contained typical data for the report forms. Some of the problems and features are discussed below.
Design validation and labeling comprehension study for a new epinephrine autoinjector.
Edwards, Eric S; Edwards, Evan T; Gunn, Ronald; Patterson, Patricia; North, Robert
2013-03-01
To facilitate the correct use of epinephrine autoinjectors (EAIs) by patients and caregivers, a novel EAI (Auvi-Q) was designed to help minimize use-related hazards. To support validation of Auvi-Q final design and assess whether the instructions for use in the patient information leaflet (PIL) are effective in training participants on proper use of Auvi-Q. Healthy participants, 20 adult and 20 pediatric, were assessed for their ability to complete a simulated injection by following the Auvi-Q instructions for use. Participants relied only on the contents of the PIL and other labeling features (device labeling and its instructions for use, electronic voice instructions and visual prompts). The mean ± SD age of the adult and pediatric participants was 39.4 ± 11.6 and 10.9 ± 2.3 years, respectively. In total, 80% of adult and 35% of pediatric participants had prior experience with EAIs. All adults and 95% of pediatric participants completed a simulated injection on the first attempt; 1 pediatric participant required parental training and a second attempt. Three adult and 4 pediatric participants exhibited a noncritical issue while successfully completing the simulated injection. Most participants agreed that the injection steps were easy to follow and the PIL facilitated understanding on using Auvi-Q safely and effectively. The PIL and other labeling features were effective in communicating instructions for successful use of Auvi-Q. This study provided validation support for the final design and anticipated instructions for use of Auvi-Q. Copyright © 2013 American College of Allergy, Asthma & Immunology. Published by Elsevier Inc. All rights reserved.
Microfluidic Injector Models Based on Artificial Neural Networks
2005-06-15
medicine, and chemistry [1], [2]. They generally perform chemical analysis involving sample preparation, mixing , reaction, injection, separation analysis...algorithms have been validated against many ex- periments found in the literature demonstrating microfluidic mixing , joule heating, injection, and...385 [7] K. Seiler, Z. H. Fan, K. Fluri, and D. J. Harrison, “ Electroosmotic pump- ing and valveless control of fluid flow within a manifold of
New Directions in the NOAO Observing Proposal System
NASA Astrophysics Data System (ADS)
Gasson, David; Bell, Dave
For the past eight years NOAO has been refining its on-line observing proposal system. Virtually all related processes are now handled electronically. Members of the astronomical community can submit proposals through email, web form, or via the Gemini Phase I Tool. NOAO staff can use the system to do administrative tasks, scheduling, and compilation of various statistics. In addition, all information relevant to the TAC process is made available on-line, including the proposals themselves (in HTML, PDF and PostScript) and technical comments. Grades and TAC comments are entered and edited through web forms, and can be sorted and filtered according to specified criteria. Current developments include a move away from proprietary solutions, toward open standards such as SQL (in the form of the MySQL relational database system), Perl, PHP and XML.
PiCO QL: A software library for runtime interactive queries on program data
NASA Astrophysics Data System (ADS)
Fragkoulis, Marios; Spinellis, Diomidis; Louridas, Panos
PiCO QL is an open source C/C++ software whose scientific scope is real-time interactive analysis of in-memory data through SQL queries. It exposes a relational view of a system's or application's data structures, which is queryable through SQL. While the application or system is executing, users can input queries through a web-based interface or issue web service requests. Queries execute on the live data structures through the respective relational views. PiCO QL makes a good candidate for ad-hoc data analysis in applications and for diagnostics in systems settings. Applications of PiCO QL include the Linux kernel, the Valgrind instrumentation framework, a GIS application, a virtual real-time observatory of stellar objects, and a source code analyser.
The unified database for the fixed target experiment BM@N
NASA Astrophysics Data System (ADS)
Gertsenberger, K. V.
2016-09-01
The article describes the developed database designed as comprehensive data storage of the fixed target experiment BM@N [1] at Joint Institute for Nuclear Research (JINR) in Dubna. The structure and purposes of the BM@N facility will be briefly presented. The scheme of the unified database and its parameters will be described in detail. The use of the BM@N database implemented on the PostgreSQL database management system (DBMS) allows one to provide user access to the actual information of the experiment. Also the interfaces developed for the access to the database will be presented. One was implemented as the set of C++ classes to access the data without SQL statements, the other-Web-interface being available on the Web page of the BM@N experiment.
ExplorEnz: the primary source of the IUBMB enzyme list
McDonald, Andrew G.; Boyce, Sinéad; Tipton, Keith F.
2009-01-01
ExplorEnz is the MySQL database that is used for the curation and dissemination of the International Union of Biochemistry and Molecular Biology (IUBMB) Enzyme Nomenclature. A simple web-based query interface is provided, along with an advanced search engine for more complex Boolean queries. The WWW front-end is accessible at http://www.enzyme-database.org, from where downloads of the database as SQL and XML are also available. An associated form-based curatorial application has been developed to facilitate the curation of enzyme data as well as the internal and public review processes that occur before an enzyme entry is made official. Suggestions for new enzyme entries, or modifications to existing ones, can be made using the forms provided at http://www.enzyme-database.org/forms.php. PMID:18776214
NASA Astrophysics Data System (ADS)
Gaspar Aparicio, R.; Gomez, D.; Coterillo Coz, I.; Wojcik, D.
2012-12-01
At CERN a number of key database applications are running on user-managed MySQL database services. The database on demand project was born out of an idea to provide the CERN user community with an environment to develop and run database services outside of the actual centralised Oracle based database services. The Database on Demand (DBoD) empowers the user to perform certain actions that had been traditionally done by database administrators, DBA's, providing an enterprise platform for database applications. It also allows the CERN user community to run different database engines, e.g. presently open community version of MySQL and single instance Oracle database server. This article describes a technology approach to face this challenge, a service level agreement, the SLA that the project provides, and an evolution of possible scenarios.
Sukhinin, Dmitrii I.; Engel, Andreas K.; Manger, Paul; Hilgetag, Claus C.
2016-01-01
Databases of structural connections of the mammalian brain, such as CoCoMac (cocomac.g-node.org) or BAMS (https://bams1.org), are valuable resources for the analysis of brain connectivity and the modeling of brain dynamics in species such as the non-human primate or the rodent, and have also contributed to the computational modeling of the human brain. Another animal model that is widely used in electrophysiological or developmental studies is the ferret; however, no systematic compilation of brain connectivity is currently available for this species. Thus, we have started developing a database of anatomical connections and architectonic features of the ferret brain, the Ferret(connect)ome, www.Ferretome.org. The Ferretome database has adapted essential features of the CoCoMac methodology and legacy, such as the CoCoMac data model. This data model was simplified and extended in order to accommodate new data modalities that were not represented previously, such as the cytoarchitecture of brain areas. The Ferretome uses a semantic parcellation of brain regions as well as a logical brain map transformation algorithm (objective relational transformation, ORT). The ORT algorithm was also adopted for the transformation of architecture data. The database is being developed in MySQL and has been populated with literature reports on tract-tracing observations in the ferret brain using a custom-designed web interface that allows efficient and validated simultaneous input and proofreading by multiple curators. The database is equipped with a non-specialist web interface. This interface can be extended to produce connectivity matrices in several formats, including a graphical representation superimposed on established ferret brain maps. An important feature of the Ferretome database is the possibility to trace back entries in connectivity matrices to the original studies archived in the system. Currently, the Ferretome contains 50 reports on connections comprising 20 injection reports with more than 150 labeled source and target areas, the majority reflecting connectivity of subcortical nuclei and 15 descriptions of regional brain architecture. We hope that the Ferretome database will become a useful resource for neuroinformatics and neural modeling, and will support studies of the ferret brain as well as facilitate advances in comparative studies of mesoscopic brain connectivity. PMID:27242503
Sukhinin, Dmitrii I; Engel, Andreas K; Manger, Paul; Hilgetag, Claus C
2016-01-01
Databases of structural connections of the mammalian brain, such as CoCoMac (cocomac.g-node.org) or BAMS (https://bams1.org), are valuable resources for the analysis of brain connectivity and the modeling of brain dynamics in species such as the non-human primate or the rodent, and have also contributed to the computational modeling of the human brain. Another animal model that is widely used in electrophysiological or developmental studies is the ferret; however, no systematic compilation of brain connectivity is currently available for this species. Thus, we have started developing a database of anatomical connections and architectonic features of the ferret brain, the Ferret(connect)ome, www.Ferretome.org. The Ferretome database has adapted essential features of the CoCoMac methodology and legacy, such as the CoCoMac data model. This data model was simplified and extended in order to accommodate new data modalities that were not represented previously, such as the cytoarchitecture of brain areas. The Ferretome uses a semantic parcellation of brain regions as well as a logical brain map transformation algorithm (objective relational transformation, ORT). The ORT algorithm was also adopted for the transformation of architecture data. The database is being developed in MySQL and has been populated with literature reports on tract-tracing observations in the ferret brain using a custom-designed web interface that allows efficient and validated simultaneous input and proofreading by multiple curators. The database is equipped with a non-specialist web interface. This interface can be extended to produce connectivity matrices in several formats, including a graphical representation superimposed on established ferret brain maps. An important feature of the Ferretome database is the possibility to trace back entries in connectivity matrices to the original studies archived in the system. Currently, the Ferretome contains 50 reports on connections comprising 20 injection reports with more than 150 labeled source and target areas, the majority reflecting connectivity of subcortical nuclei and 15 descriptions of regional brain architecture. We hope that the Ferretome database will become a useful resource for neuroinformatics and neural modeling, and will support studies of the ferret brain as well as facilitate advances in comparative studies of mesoscopic brain connectivity.
Lazzari, Barbara; Caprera, Andrea; Cestaro, Alessandro; Merelli, Ivan; Del Corvo, Marcello; Fontana, Paolo; Milanesi, Luciano; Velasco, Riccardo; Stella, Alessandra
2009-06-29
Two complete genome sequences are available for Vitis vinifera Pinot noir. Based on the sequence and gene predictions produced by the IASMA, we performed an in silico detection of putative microRNA genes and of their targets, and collected the most reliable microRNA predictions in a web database. The application is available at http://www.itb.cnr.it/ptp/grapemirna/. The program FindMiRNA was used to detect putative microRNA genes in the grape genome. A very high number of predictions was retrieved, calling for validation. Nine parameters were calculated and, based on the grape microRNAs dataset available at miRBase, thresholds were defined and applied to FindMiRNA predictions having targets in gene exons. In the resulting subset, predictions were ranked according to precursor positions and sequence similarity, and to target identity. To further validate FindMiRNA predictions, comparisons to the Arabidopsis genome, to the grape Genoscope genome, and to the grape EST collection were performed. Results were stored in a MySQL database and a web interface was prepared to query the database and retrieve predictions of interest. The GrapeMiRNA database encompasses 5,778 microRNA predictions spanning the whole grape genome. Predictions are integrated with information that can be of use in selection procedures. Tools added in the web interface also allow to inspect predictions according to gene ontology classes and metabolic pathways of targets. The GrapeMiRNA database can be of help in selecting candidate microRNA genes to be validated.
Horii, Takuro; Arai, Yuji; Yamazaki, Miho; Morita, Sumiyo; Kimura, Mika; Itoh, Masahiro; Abe, Yumiko; Hatada, Izuho
2014-03-28
The CRISPR/Cas system, in which the Cas9 endonuclease and a guide RNA complementary to the target are sufficient for RNA-guided cleavage of the target DNA, is a powerful new approach recently developed for targeted gene disruption in various animal models. However, there is little verification of microinjection methods for generating knockout mice using this approach. Here, we report the verification of microinjection methods of the CRISPR/Cas system. We compared three methods for injection: (1) injection of DNA into the pronucleus, (2) injection of RNA into the pronucleus, and (3) injection of RNA into the cytoplasm. We found that injection of RNA into the cytoplasm was the most efficient method in terms of the numbers of viable blastocyst stage embryos and full-term pups generated. This method also showed the best overall knockout efficiency.
NASA Astrophysics Data System (ADS)
Desantes, J. M.; Salvador, F. J.; López, J. J.; de La Morena, J.
2011-02-01
In this paper, a research aimed at quantifying mass and momentum transfer in the near-nozzle field of diesel sprays injected into stagnant ambient air is reported. The study combines X-ray measurements for two different nozzles and axial positions, which provide mass distributions in the spray, with a theoretical model based on momentum flux conservation, which was previously validated. This investigation has allowed the validation of Gaussian profiles for local fuel concentration and velocity near the nozzle exit, as well as the determination of Schmidt number at realistic diesel spray conditions. This information could be very useful for those who are interested in spray modeling, especially at high-pressure injection conditions.
NASA Astrophysics Data System (ADS)
Therssen, E.; Delfosse, L.
1995-08-01
The design and setting up of a pulverized solid injection system for use in laboratory burners is presented. The original dual system consists of a screw feeder coupled to an acoustic sower. This laboratory device allows a good regularity and stability of the particle-gas mixture transported to the burner in a large scale of mass powder and gas vector rate flow. The thermal history of the particles has been followed by optical measurements. The quality of the particle cloud injected in the burner has been validated by the good agreement between experimental and modeling particle temperature.
NASA Astrophysics Data System (ADS)
Lamarre, Simon G.
Le moulage par injection a basse pression de poudre metallique est une technique de mise en forme de pieces de formes complexes. La poudre metallique est melangee avec des polymeres basse viscosite (ex. : cire) pour former un melange homogene a une temperature superieure a la temperature de fusion des polymeres. Pour faciliter l'injection dans la cavite du moule, la composition des melanges est ajustee pour diminuer la viscosite. D'une part, les melanges peu visqueux possedent une bonne moulabilite. D'autre part, le phenomene de la segregation se manifeste rapidement avec les melanges peu visqueux. Les machines commerciales sont munies d'un canal d'injection et d'une valve qui relient le reservoir de melange et la cavite du moule. Le melange reste stationnaire dans ces composantes entre deux sequences d'injection, ce qui le rend propice a la segregation. Plusieurs brevets tentent de resoudre ce probleme en utilisant des pompes et des canaux de recirculation. Ces composantes sont difficiles a nettoyer en raison de leur complexite. Une machine a injection basse pression a ete concue et fabriquee pour l'etude de l'aptitude au moulage des melanges de tres faible viscosite (ex. : 0.1 Pa˙s), qui tient compte du phenomene de segregation et des contraintes de nettoyage. Un piston d'injection puise le volume desire d'un reservoir. Ensuite, un mouvement lateral cisaille le melange a l'intersection entre le reservoir et le cylindre et bouche l'orifice de sortie du reservoir. Le cylindre est degage et peut recevoir le moule. A la suite de l'injection, le piston retourne a la position du reservoir et entre dans son orifice de sortie. Le melange residuel est retourne dans le reservoir, melange et desaere a nouveau. L'appareil a ete valide par des essais d'injectabilite avec un melange de poudre d'acier inoxydable et de liants a basse viscosite. Des essais d'injection ont montre que le melange contenant l'acide stearique a parcouru la plus grande distance dans le moule de forme spirale, suivi du melange contenant l'acide stearique et l'ethylene vinyle acetate et finalement du melange contenant seulement la cire de paraffine. Des essais de rheologie ont correle ces resultats pour une viscosite associee a un faible taux de cisaillement. Un essai de segregation a montre qu'un temps d'attente de moins d'une minute devait etre respecte avant de mesurer une segregation importante. De plus, le fonctionnement du systeme de vacuum a ete valide par des analyses radiographiques. Un temps de melange sous vacuum de 30 min est suffisant pour eliminer completement la presence de bulles d'air dans la piece injectee.
FPGA-Based Multiprocessor System for Injection Molding Control
Muñoz-Barron, Benigno; Morales-Velazquez, Luis; Romero-Troncoso, Rene J.; Rodriguez-Donate, Carlos; Trejo-Hernandez, Miguel; Benitez-Rangel, Juan P.; Osornio-Rios, Roque A.
2012-01-01
The plastic industry is a very important manufacturing sector and injection molding is a widely used forming method in that industry. The contribution of this work is the development of a strategy to retrofit control of an injection molding machine based on an embedded system microprocessors sensor network on a field programmable gate array (FPGA) device. Six types of embedded processors are included in the system: a smart-sensor processor, a micro fuzzy logic controller, a programmable logic controller, a system manager, an IO processor and a communication processor. Temperature, pressure and position are controlled by the proposed system and experimentation results show its feasibility and robustness. As validation of the present work, a particular sample was successfully injected. PMID:23202036
NASA Astrophysics Data System (ADS)
Sterpone, L.; Violante, M.
2007-08-01
Modern SRAM-based field programmable gate array (FPGA) devices offer high capability in implementing complex system. Unfortunately, SRAM-based FPGAs are extremely sensitive to single event upsets (SEUs) induced by radiation particles. In order to successfully deploy safety- or mission-critical applications, designer need to validate the correctness of the obtained designs. In this paper we describe a system based on partial-reconfiguration for running fault-injection experiments within the configuration memory of SRAM-based FPGAs. The proposed fault-injection system uses the internal configuration capabilities that modern FPGAs offer in order to inject SEU within the configuration memory. Detailed experimental results show that the technique is orders of magnitude faster than previously proposed ones.
Gauglitz, Gerd; Steckmeier, Stephanie; Pötschke, Julian; Schwaiger, Hannah
2017-01-01
Cohesive Polydensified Matrix ® hyaluronic acid (CPM-HA) volumizer has been used successfully for several years to reverse biometric volume loss during facial aging. This observational study explored the additive effect on nasolabial folds when CPM-HA volumizer is injected into the neighboring cheek area. In this open-label, prospective, postmarketing noninterventional study, 18 adult patients seeking esthetic enhancement of the lateral cheek hollows and cheekbone area were injected with CPM-HA volumizer integrated with lidocaine (CPM-HA-VL) in the upper or lower cheek area. Safety and performance of CPM-HA-VL up to 12 months after injection with follow-up visits at week 4 and month 3, 6, and 12 were assessed. The primary endpoint was improvement of cheek fullness on the validated Merz Aesthetics Scales. Additionally, changes in nasolabial folds were quantified using a phaseshift rapid in vivo measurement of skin optical three-dimensional (3D) in vivo measurement device. Patients (94.4% female, median age 52 years, age range 39-69 years) were injected with a mean volume of 2.5±1.1 mL CPM-HA-VL per side. Immediately after injection, mean severity for upper and lower cheek fullness assessed on the validated MAS improved from 2.5±0.6 and 2.8±0.5, respectively, to 1.0±0.0, and remained unchanged through month 12. Improvement in relation to baseline was attested on the Global Aesthetics Improvement Scale for all assessments. Compared with baseline, the following assessments offered a statistical significance in the reduction of wrinkle depth of nasolabial folds (maximum depth reduction by 30.4% at 3 months) according to optical 3D in vivo measurements. Pain during injection was minimal and abated within 30 minutes. Treatment was well tolerated and led to great patient satisfaction. CPM-HA-VL injected into the upper and lower cheeks led to long-lasting satisfactory cosmetic results in cheek augmentation as well as in reducing depth of nasolabial folds adjacent to the injection site.
Chandra monitoring, trends, and response
NASA Astrophysics Data System (ADS)
Spitzbart, Brad D.; Wolk, Scott J.; Isobe, Takashi
2002-12-01
The Chandra X-ray Observatory was launched in July, 1999 and has yielded extraordinary scientific results. Behind the scenes, our Monitoring and Trends Analysis (MTA) system has proven to be a valuable resource. With three years worth of on-orbit data, we have available a vast array of both telescope diagnostic information and analysis of scientific data to access Observatory performance. As part of Chandra's Science Operations Team (SOT), the primary goal of MTA is to provide tools for effective decision making leading to the most efficient production of quality science output from the Observatory. We occupy a middle ground between flight operations, chiefly concerned with the health and safety of the spacecraft, and validation and verification, concerned with the scientific validity of the data taken and whether or not they fulfill the observer's requirements. In that role we provide and receive support from systems engineers, instrument experts, operations managers, and scientific users. MTA tools, products, and services include real-time monitoring and alert generation for the most mission critical components, long term trending of all spacecraft systems, detailed analysis of various subsystems for life expectancy or anomaly resolution, and creating and maintaining a large SQL database of relevant information. This is accomplished through the use of a wide variety of input data sources and flexible, accessible programming and analysis techniques. This paper will discuss the overall design of the system, its evolution and the resources available.
Ben Hamida, Amen; Rafful, Claudia; Jain, Sonia; Sun, Shelly; Gonzalez-Zuniga, Patricia; Rangel, Gudelia; Strathdee, Steffanie A; Werb, Dan
2018-02-01
Although most people who inject drugs (PWID) report receiving assistance during injection initiation events, little research has focused on risk factors among PWID for providing injection initiation assistance. We therefore sought to determine the influence of non-injection drug use among PWID on their risk to initiate others. We used generalized estimating equation (GEE) models on longitudinal data among a prospective cohort of PWID in Tijuana, Mexico (Proyecto El Cuete IV), while controlling for potential confounders. At baseline, 534 participants provided data on injection initiation assistance. Overall, 14% reported ever initiating others, with 4% reporting this behavior recently (i.e., in the past 6 months). In a multivariable GEE model, recent non-injection drug use was independently associated with providing injection initiation assistance (adjusted odds ratio [AOR] = 2.42, 95% confidence interval [CI] = 1.39-4.20). Further, in subanalyses examining specific drug types, recent non-injection use of cocaine (AOR = 9.31, 95% CI = 3.98-21.78), heroin (AOR = 4.00, 95% CI = 1.88-8.54), and methamphetamine (AOR = 2.03, 95% CI = 1.16-3.55) were all significantly associated with reporting providing injection initiation assistance. Our findings may have important implications for the development of interventional approaches to reduce injection initiation and related harms. Further research is needed to validate findings and inform future approaches to preventing entry into drug injecting.
McCarthy, Ian D; Nicholls, Ruth; Malham, Shelagh K; Whiteley, Nia M
2016-01-01
For the first time, use of the flooding dose technique using (3)H-Phenylalanine is validated for measuring whole-animal and tissue-specific rates of protein synthesis in the blue mussel Mytilus edulis (61mm shell length; 4.0g fresh body mass). Following injection, the phenylalanine-specific radioactivities in the gill, mantle and whole-animal free pools were elevated within one hour and remained elevated and stable for up to 6h following injection of (3)H-phenylalanine into the posterior adductor muscle. Incorporation of (3)H-phenylalanine into body protein was linear over time following injection and the non-significant intercepts for the regressions suggested incorporation into body protein occurred rapidly after injection. These results validate the technique for measuring rates of protein synthesis in mussels. There were no differences in the calculated rates following 1-6h incubation in gill, mantle or whole-animal and fractional rates of protein synthesis from the combined time course data were 9.5±0.8%d(-1) for the gill, 2.5±0.3%d(-1) for the mantle and 2.6±0.3%d(-1) for the whole-animal, respectively (mean values±SEM). The whole-animal absolute rate of protein synthesis was calculated as 18.9±0.6mg protein day(-1). The use of this technique in measuring one of the major components of maintenance metabolism and growth will provide a valuable and convenient tool in furthering our understanding of the protein metabolism and energetics of this keystone marine invertebrate and its ability to adjust and respond to fluctuations, such as that expected as a result of climate change. Copyright © 2015 Elsevier Inc. All rights reserved.
3-D simulation of gases transport under condition of inert gas injection into goaf
NASA Astrophysics Data System (ADS)
Liu, Mao-Xi; Shi, Guo-Qing; Guo, Zhixiong; Wang, Yan-Ming; Ma, Li-Yang
2016-12-01
To prevent coal spontaneous combustion in mines, it is paramount to understand O2 gas distribution under condition of inert gas injection into goaf. In this study, the goaf was modeled as a 3-D porous medium based on stress distribution. The variation of O2 distribution influenced by CO2 or N2 injection was simulated based on the multi-component gases transport and the Navier-Stokes equations using Fluent. The numerical results without inert gas injection were compared with field measurements to validate the simulation model. Simulations with inert gas injection show that CO2 gas mainly accumulates at the goaf floor level; however, a notable portion of N2 gas moves upward. The evolution of the spontaneous combustion risky zone with continuous inert gas injection can be classified into three phases: slow inerting phase, rapid accelerating inerting phase, and stable inerting phase. The asphyxia zone with CO2 injection is about 1.25-2.4 times larger than that with N2 injection. The efficacy of preventing and putting out mine fires is strongly related with the inert gas injecting position. Ideal injections are located in the oxidation zone or the transitional zone between oxidation zone and heat dissipation zone.
Cluster dynamics modeling and experimental investigation of the effect of injected interstitials
NASA Astrophysics Data System (ADS)
Michaut, B.; Jourdan, T.; Malaplate, J.; Renault-Laborne, A.; Sefta, F.; Décamps, B.
2017-12-01
The effect of injected interstitials on loop and cavity microstructures is investigated experimentally and numerically for 304L austenitic stainless steel irradiated at 450 °C with 10 MeV Fe5+ ions up to about 100 dpa. A cluster dynamics model is parametrized on experimental results obtained by transmission electron microscopy (TEM) in a region where injected interstitials can be safely neglected. It is then used to model the damage profile and study the impact of self-ion injection. Results are compared to TEM observations on cross-sections of specimens. It is shown that injected interstitials have a significant effect on cavity density and mean size, even in the sink-dominated regime. To quantitatively match the experimental data in the self-ions injected area, a variation of some parameters is necessary. We propose that the fraction of freely migrating species may vary as a function of depth. Finally, we show that simple rate theory considerations do not seem to be valid for these experimental conditions.
PACSY, a relational database management system for protein structure and chemical shift analysis.
Lee, Woonghee; Yu, Wookyung; Kim, Suhkmann; Chang, Iksoo; Lee, Weontae; Markley, John L
2012-10-01
PACSY (Protein structure And Chemical Shift NMR spectroscopY) is a relational database management system that integrates information from the Protein Data Bank, the Biological Magnetic Resonance Data Bank, and the Structural Classification of Proteins database. PACSY provides three-dimensional coordinates and chemical shifts of atoms along with derived information such as torsion angles, solvent accessible surface areas, and hydrophobicity scales. PACSY consists of six relational table types linked to one another for coherence by key identification numbers. Database queries are enabled by advanced search functions supported by an RDBMS server such as MySQL or PostgreSQL. PACSY enables users to search for combinations of information from different database sources in support of their research. Two software packages, PACSY Maker for database creation and PACSY Analyzer for database analysis, are available from http://pacsy.nmrfam.wisc.edu.
Wiley, Laura K.; Sivley, R. Michael; Bush, William S.
2013-01-01
Efficient storage and retrieval of genomic annotations based on range intervals is necessary, given the amount of data produced by next-generation sequencing studies. The indexing strategies of relational database systems (such as MySQL) greatly inhibit their use in genomic annotation tasks. This has led to the development of stand-alone applications that are dependent on flat-file libraries. In this work, we introduce MyNCList, an implementation of the NCList data structure within a MySQL database. MyNCList enables the storage, update and rapid retrieval of genomic annotations from the convenience of a relational database system. Range-based annotations of 1 million variants are retrieved in under a minute, making this approach feasible for whole-genome annotation tasks. Database URL: https://github.com/bushlab/mynclist PMID:23894185
Wiley, Laura K; Sivley, R Michael; Bush, William S
2013-01-01
Efficient storage and retrieval of genomic annotations based on range intervals is necessary, given the amount of data produced by next-generation sequencing studies. The indexing strategies of relational database systems (such as MySQL) greatly inhibit their use in genomic annotation tasks. This has led to the development of stand-alone applications that are dependent on flat-file libraries. In this work, we introduce MyNCList, an implementation of the NCList data structure within a MySQL database. MyNCList enables the storage, update and rapid retrieval of genomic annotations from the convenience of a relational database system. Range-based annotations of 1 million variants are retrieved in under a minute, making this approach feasible for whole-genome annotation tasks. Database URL: https://github.com/bushlab/mynclist.
MaGnET: Malaria Genome Exploration Tool
Sharman, Joanna L.; Gerloff, Dietlind L.
2013-01-01
Summary: The Malaria Genome Exploration Tool (MaGnET) is a software tool enabling intuitive ‘exploration-style’ visualization of functional genomics data relating to the malaria parasite, Plasmodium falciparum. MaGnET provides innovative integrated graphic displays for different datasets, including genomic location of genes, mRNA expression data, protein–protein interactions and more. Any selection of genes to explore made by the user is easily carried over between the different viewers for different datasets, and can be changed interactively at any point (without returning to a search). Availability and Implementation: Free online use (Java Web Start) or download (Java application archive and MySQL database; requires local MySQL installation) at http://malariagenomeexplorer.org Contact: joanna.sharman@ed.ac.uk or dgerloff@ffame.org Supplementary information: Supplementary data are available at Bioinformatics online. PMID:23894142
A SQL-Database Based Meta-CASE System and its Query Subsystem
NASA Astrophysics Data System (ADS)
Eessaar, Erki; Sgirka, Rünno
Meta-CASE systems simplify the creation of CASE (Computer Aided System Engineering) systems. In this paper, we present a meta-CASE system that provides a web-based user interface and uses an object-relational database system (ORDBMS) as its basis. The use of ORDBMSs allows us to integrate different parts of the system and simplify the creation of meta-CASE and CASE systems. ORDBMSs provide powerful query mechanism. The proposed system allows developers to use queries to evaluate and gradually improve artifacts and calculate values of software measures. We illustrate the use of the systems by using SimpleM modeling language and discuss the use of SQL in the context of queries about artifacts. We have created a prototype of the meta-CASE system by using PostgreSQL™ ORDBMS and PHP scripting language.
Database on Demand: insight how to build your own DBaaS
NASA Astrophysics Data System (ADS)
Gaspar Aparicio, Ruben; Coterillo Coz, Ignacio
2015-12-01
At CERN, a number of key database applications are running on user-managed MySQL, PostgreSQL and Oracle database services. The Database on Demand (DBoD) project was born out of an idea to provide CERN user community with an environment to develop and run database services as a complement to the central Oracle based database service. The Database on Demand empowers the user to perform certain actions that had been traditionally done by database administrators, providing an enterprise platform for database applications. It also allows the CERN user community to run different database engines, e.g. presently three major RDBMS (relational database management system) vendors are offered. In this article we show the actual status of the service after almost three years of operations, some insight of our new redesign software engineering and near future evolution.
Lu, Ying-Hao; Kuo, Chen-Chun; Huang, Yaw-Bin
2011-08-01
We selected HTML, PHP and JavaScript as the programming languages to build "WebBio", a web-based system for patient data of biological products and used MySQL as database. WebBio is based on the PHP-MySQL suite and is run by Apache server on Linux machine. WebBio provides the functions of data management, searching function and data analysis for 20 kinds of biological products (plasma expanders, human immunoglobulin and hematological products). There are two particular features in WebBio: (1) pharmacists can rapidly find out whose patients used contaminated products for medication safety, and (2) the statistics charts for a specific product can be automatically generated to reduce pharmacist's work loading. WebBio has successfully turned traditional paper work into web-based data management.
A Comparison of Different Database Technologies for the CMS AsyncStageOut Transfer Database
DOE Office of Scientific and Technical Information (OSTI.GOV)
Ciangottini, D.; Balcas, J.; Mascheroni, M.
AsyncStageOut (ASO) is the component of the CMS distributed data analysis system (CRAB) that manages users transfers in a centrally controlled way using the File Transfer System (FTS3) at CERN. It addresses a major weakness of the previous, decentralized model, namely that the transfer of the user’s output data to a single remote site was part of the job execution, resulting in inefficient use of job slots and an unacceptable failure rate. Currently ASO manages up to 600k files of various sizes per day from more than 500 users per month, spread over more than 100 sites. ASO uses amore » NoSQL database (CouchDB) as internal bookkeeping and as way to communicate with other CRAB components. Since ASO/CRAB were put in production in 2014, the number of transfers constantly increased up to a point where the pressure to the central CouchDB instance became critical, creating new challenges for the system scalability, performance, and monitoring. This forced a re-engineering of the ASO application to increase its scalability and lowering its operational effort. In this contribution we present a comparison of the performance of the current NoSQL implementation and a new SQL implementation, and how their different strengths and features influenced the design choices and operational experience. We also discuss other architectural changes introduced in the system to handle the increasing load and latency in delivering output to the user.« less
A comparison of different database technologies for the CMS AsyncStageOut transfer database
NASA Astrophysics Data System (ADS)
Ciangottini, D.; Balcas, J.; Mascheroni, M.; Rupeika, E. A.; Vaandering, E.; Riahi, H.; Silva, J. M. D.; Hernandez, J. M.; Belforte, S.; Ivanov, T. T.
2017-10-01
AsyncStageOut (ASO) is the component of the CMS distributed data analysis system (CRAB) that manages users transfers in a centrally controlled way using the File Transfer System (FTS3) at CERN. It addresses a major weakness of the previous, decentralized model, namely that the transfer of the user’s output data to a single remote site was part of the job execution, resulting in inefficient use of job slots and an unacceptable failure rate. Currently ASO manages up to 600k files of various sizes per day from more than 500 users per month, spread over more than 100 sites. ASO uses a NoSQL database (CouchDB) as internal bookkeeping and as way to communicate with other CRAB components. Since ASO/CRAB were put in production in 2014, the number of transfers constantly increased up to a point where the pressure to the central CouchDB instance became critical, creating new challenges for the system scalability, performance, and monitoring. This forced a re-engineering of the ASO application to increase its scalability and lowering its operational effort. In this contribution we present a comparison of the performance of the current NoSQL implementation and a new SQL implementation, and how their different strengths and features influenced the design choices and operational experience. We also discuss other architectural changes introduced in the system to handle the increasing load and latency in delivering output to the user.
Temsch, W; Luger, A; Riedl, M
2008-01-01
This article presents a mathematical model to calculate HbA1c values based on self-measured blood glucose and past HbA1c levels, thereby enabling patients to monitor diabetes therapy between scheduled checkups. This method could help physicians to make treatment decisions if implemented in a system where glucose data are transferred to a remote server. The method, however, cannot replace HbA1c measurements; past HbA1c values are needed to gauge the method. The mathematical model of HbA1c formation was developed based on biochemical principles. Unlike an existing HbA1c formula, the new model respects the decreasing contribution of older glucose levels to current HbA1c values. About 12 standard SQL statements embedded in a php program were used to perform Fourier transform. Regression analysis was used to gauge results with previous HbA1c values. The method can be readily implemented in any SQL database. The predicted HbA1c values thus obtained were in accordance with measured values. They also matched the results of the HbA1c formula in the elevated range. By contrast, the formula was too "optimistic" in the range of better glycemic control. Individual analysis of two subjects improved the accuracy of values and reflected the bias introduced by different glucometers and individual measurement habits.
Improved oilfield GHG accounting using a global oilfield database
NASA Astrophysics Data System (ADS)
Roberts, S.; Brandt, A. R.; Masnadi, M.
2016-12-01
The definition of oil is shifting in considerable ways. Conventional oil resources are declining as oil sands, heavy oils, and others emerge. Technological advances mean that these unconventional hydrocarbons are now viable resources. Meanwhile, scientific evidence is mounting that climate change is occurring. The oil sector is responsible for 35% of global greenhouse gas (GHG) emissions, but the climate impacts of these new unconventional oils are not well understood. As such, the Oil Climate Index (OCI) project has been an international effort to evaluate the total life-cycle environmental GHG emissions of different oil fields globally. Over the course of the first and second phases of the project, 30 and 75 global oil fields have been investigated, respectively. The 75 fields account for about 25% of global oil production. For the third phase of the project, it is aimed to expand the OCI to contain closing to 100% of global oil production; leading to the analysis of 8000 fields. To accomplish this, a robust database system is required to handle and manipulate the data. Therefore, the integration of the data into the computer science language SQL (Structured Query Language) was performed. The implementation of SQL allows users to process the data more efficiently than would be possible by using the previously established program (Microsoft Excel). Next, a graphic user interface (gui) was implemented, in the computer science language of C#, in order to make the data interactive; enabling people to update the database without prior knowledge of SQL being necessary.
Correcting ligands, metabolites, and pathways
Ott, Martin A; Vriend, Gert
2006-01-01
Background A wide range of research areas in bioinformatics, molecular biology and medicinal chemistry require precise chemical structure information about molecules and reactions, e.g. drug design, ligand docking, metabolic network reconstruction, and systems biology. Most available databases, however, treat chemical structures more as illustrations than as a datafield in its own right. Lack of chemical accuracy impedes progress in the areas mentioned above. We present a database of metabolites called BioMeta that augments the existing pathway databases by explicitly assessing the validity, correctness, and completeness of chemical structure and reaction information. Description The main bulk of the data in BioMeta were obtained from the KEGG Ligand database. We developed a tool for chemical structure validation which assesses the chemical validity and stereochemical completeness of a molecule description. The validation tool was used to examine the compounds in BioMeta, showing that a relatively small number of compounds had an incorrect constitution (connectivity only, not considering stereochemistry) and that a considerable number (about one third) had incomplete or even incorrect stereochemistry. We made a large effort to correct the errors and to complete the structural descriptions. A total of 1468 structures were corrected and/or completed. We also established the reaction balance of the reactions in BioMeta and corrected 55% of the unbalanced (stoichiometrically incorrect) reactions in an automatic procedure. The BioMeta database was implemented in PostgreSQL and provided with a web-based interface. Conclusion We demonstrate that the validation of metabolite structures and reactions is a feasible and worthwhile undertaking, and that the validation results can be used to trigger corrections and improvements to BioMeta, our metabolite database. BioMeta provides some tools for rational drug design, reaction searches, and visualization. It is freely available at provided that the copyright notice of all original data is cited. The database will be useful for querying and browsing biochemical pathways, and to obtain reference information for identifying compounds. However, these applications require that the underlying data be correct, and that is the focus of BioMeta. PMID:17132165
DOE Office of Scientific and Technical Information (OSTI.GOV)
Hansen, C., E-mail: hansec@uw.edu; Columbia University, New York, New York 10027; Victor, B.
We present application of three scalar metrics derived from the Biorthogonal Decomposition (BD) technique to evaluate the level of agreement between macroscopic plasma dynamics in different data sets. BD decomposes large data sets, as produced by distributed diagnostic arrays, into principal mode structures without assumptions on spatial or temporal structure. These metrics have been applied to validation of the Hall-MHD model using experimental data from the Helicity Injected Torus with Steady Inductive helicity injection experiment. Each metric provides a measure of correlation between mode structures extracted from experimental data and simulations for an array of 192 surface-mounted magnetic probes. Numericalmore » validation studies have been performed using the NIMROD code, where the injectors are modeled as boundary conditions on the flux conserver, and the PSI-TET code, where the entire plasma volume is treated. Initial results from a comprehensive validation study of high performance operation with different injector frequencies are presented, illustrating application of the BD method. Using a simplified (constant, uniform density and temperature) Hall-MHD model, simulation results agree with experimental observation for two of the three defined metrics when the injectors are driven with a frequency of 14.5 kHz.« less
Thrust augmentation nozzle (TAN) concept for rocket engine booster applications
NASA Astrophysics Data System (ADS)
Forde, Scott; Bulman, Mel; Neill, Todd
2006-07-01
Aerojet used the patented thrust augmented nozzle (TAN) concept to validate a unique means of increasing sea-level thrust in a liquid rocket booster engine. We have used knowledge gained from hypersonic Scramjet research to inject propellants into the supersonic region of the rocket engine nozzle to significantly increase sea-level thrust without significantly impacting specific impulse. The TAN concept overcomes conventional engine limitations by injecting propellants and combusting in an annular region in the divergent section of the nozzle. This injection of propellants at moderate pressures allows for obtaining high thrust at takeoff without overexpansion thrust losses. The main chamber is operated at a constant pressure while maintaining a constant head rise and flow rate of the main propellant pumps. Recent hot-fire tests have validated the design approach and thrust augmentation ratios. Calculations of nozzle performance and wall pressures were made using computational fluid dynamics analyses with and without thrust augmentation flow, resulting in good agreement between calculated and measured quantities including augmentation thrust. This paper describes the TAN concept, the test setup, test results, and calculation results.
CRAB3: Establishing a new generation of services for distributed analysis at CMS
NASA Astrophysics Data System (ADS)
Cinquilli, M.; Spiga, D.; Grandi, C.; Hernàndez, J. M.; Konstantinov, P.; Mascheroni, M.; Riahi, H.; Vaandering, E.
2012-12-01
In CMS Computing the highest priorities for analysis tools are the improvement of the end users’ ability to produce and publish reliable samples and analysis results as well as a transition to a sustainable development and operations model. To achieve these goals CMS decided to incorporate analysis processing into the same framework as data and simulation processing. This strategy foresees that all workload tools (TierO, Tier1, production, analysis) share a common core with long term maintainability as well as the standardization of the operator interfaces. The re-engineered analysis workload manager, called CRAB3, makes use of newer technologies, such as RESTFul based web services and NoSQL Databases, aiming to increase the scalability and reliability of the system. As opposed to CRAB2, in CRAB3 all work is centrally injected and managed in a global queue. A pool of agents, which can be geographically distributed, consumes work from the central services serving the user tasks. The new architecture of CRAB substantially changes the deployment model and operations activities. In this paper we present the implementation of CRAB3, emphasizing how the new architecture improves the workflow automation and simplifies maintainability. In particular, we will highlight the impact of the new design on daily operations.
Simulators for training in ultrasound guided procedures.
Farjad Sultan, Syed; Shorten, George; Iohom, Gabrielle
2013-06-01
The four major categories of skill sets associated with proficiency in ultrasound guided regional anaesthesia are 1) understanding device operations, 2) image optimization, 3) image interpretation and 4) visualization of needle insertion and injection of the local anesthetic solution. Of these, visualization of needle insertion and injection of local anaesthetic solution can be practiced using simulators and phantoms. This survey of existing simulators summarizes advantages and disadvantages of each. Current deficits pertain to the validation process.
Anisotropic mechanical behavior of an injection molded short fiber reinforced thermoplastic
NASA Astrophysics Data System (ADS)
Lopez, Delphine; Thuillier, Sandrine; Bessières, Nicolas; Grohens, Yves
2016-10-01
A short fiber reinforced thermoplastic was injected into a rectangular mold, in order to prepare samples to characterize the mechanical behavior of the material. The injection process was simulated with Moldflow and a cutting pattern was deduced from the predicted fiber orientation, leading to samples with several well-defined orientations with respect to the injection direction. Monotonic tensile tests up to rupture, as well as complex cycles made of loading steps followed by relaxation steps at different strain levels were performed, in order to check the reproducibility for a given orientation. Moreover, the fiber orientation in the central part of the tensile samples was also analyzed with X-ray tomography. The results show that the mechanical behavior for each orientation (among 6) was rather reproducible, thus validating the cutting pattern.
Briggs, J. D.; Boyle, J. A.
1965-01-01
The level of plasma radioactivity following a single intravenous injection of 131I-labelled sodium orthoiodohippurate (radiohuppuran) falls with time in a tri-exponential fashion. The rate of fall of plasma radioactivity after an intravenous injection of radiohippuran was measured over the period 25 to 40 minutes from the time of injection and was expressed as the half-life of radiohippuran. The results suggest that this procedure may provide a valid measure of renal function which is more sensitive than the blood urea estimation but less sensitive than the creatinine clearance. PMID:14304256
Simulated fault injection - A methodology to evaluate fault tolerant microprocessor architectures
NASA Technical Reports Server (NTRS)
Choi, Gwan S.; Iyer, Ravishankar K.; Carreno, Victor A.
1990-01-01
A simulation-based fault-injection method for validating fault-tolerant microprocessor architectures is described. The approach uses mixed-mode simulation (electrical/logic analysis), and injects transient errors in run-time to assess the resulting fault impact. As an example, a fault-tolerant architecture which models the digital aspects of a dual-channel real-time jet-engine controller is used. The level of effectiveness of the dual configuration with respect to single and multiple transients is measured. The results indicate 100 percent coverage of single transients. Approximately 12 percent of the multiple transients affect both channels; none result in controller failure since two additional levels of redundancy exist.
Astronomical databases of Nikolaev Observatory
NASA Astrophysics Data System (ADS)
Protsyuk, Y.; Mazhaev, A.
2008-07-01
Several astronomical databases were created at Nikolaev Observatory during the last years. The databases are built by using MySQL search engine and PHP scripts. They are available on NAO web-site http://www.mao.nikolaev.ua.
NASA Astrophysics Data System (ADS)
Kulchitsky, A.; Maurits, S.; Watkins, B.
2006-12-01
With the widespread availability of the Internet today, many people can monitor various scientific research activities. It is important to accommodate this interest providing on-line access to dynamic and illustrative Web-resources, which could demonstrate different aspects of ongoing research. It is especially important to explain and these research activities for high school and undergraduate students, thereby providing more information for making decisions concerning their future studies. Such Web resources are also important to clarify scientific research for the general public, in order to achieve better awareness of research progress in various fields. Particularly rewarding is dissemination of information about ongoing projects within Universities and research centers to their local communities. The benefits of this type of scientific outreach are mutual, since development of Web-based automatic systems is prerequisite for many research projects targeting real-time monitoring and/or modeling of natural conditions. Continuous operation of such systems provide ongoing research opportunities for the statistically massive validation of the models, as well. We have developed a Web-based system to run the University of Alaska Fairbanks Polar Ionospheric Model in real-time. This model makes use of networking and computational resources at the Arctic Region Supercomputing Center. This system was designed to be portable among various operating systems and computational resources. Its components can be installed across different computers, separating Web servers and computational engines. The core of the system is a Real-Time Management module (RMM) written Python, which facilitates interactions of remote input data transfers, the ionospheric model runs, MySQL database filling, and PHP scripts for the Web-page preparations. The RMM downloads current geophysical inputs as soon as they become available at different on-line depositories. This information is processed to provide inputs for the next ionospheic model time step and then stored in a MySQL database as the first part of the time-specific record. The RMM then performs synchronization of the input times with the current model time, prepares a decision on initialization for the next model time step, and monitors its execution. Then, as soon as the model completes computations for the next time step, RMM visualizes the current model output into various short-term (about 1-2 hours) forecasting products and compares prior results with available ionospheric measurements. The RMM places prepared images into the MySQL database, which can be located on a different computer node, and then proceeds to the next time interval continuing the time-loop. The upper-level interface of this real-time system is the a PHP-based Web site (http://www.arsc.edu/SpaceWeather/new). This site provides general information about the Earth polar and adjacent mid-latitude ionosphere, allows for monitoring of the current developments and short-term forecasts, and facilitates access to the comparisons archive stored in the database.
Hewala, Ismail; El-Fatatre, Hamed; Emam, Ehab; Mubrouk, Mokhtar
2010-06-30
A simple, rapid and sensitive reversed phase high performance liquid chromatographic method using photodiode array detection was developed and validated for the simultaneous determination of granisetron hydrochloride, benzyl alcohol, 1-methyl-1H-indazole-3-carboxylic acid (the main degradation product of granisetron) and benzaldehyde (the main degradation product of benzyl alcohol) in granisetron injections. The separation was achieved on Hypersil BDS C8 (250 mm x 4.6 mm i.d., 5 microm particle diameter) column using a mobile phase consisted of acetonitrile:0.05 M KH(2)PO(4):triethylamine (22:100:0.15) adjusted to pH 4.8. The column was maintained at 25 degrees C and 20 microL of solutions was injected. Photodiode array detector was used to test the peak purity and the chromatograms were extracted at 210 nm. Naphazoline hydrochloride was used as internal standard. The method was validated with respect to specificity, linearity, accuracy, precision, limit of quantitation and limit of detection. The validation acceptance criteria were met in all cases. Identification of the pure peaks was carried out using library match programmer and wavelengths of derivative optima of the spectrograms of the peaks. The method was successfully applied to the determination of the investigated drugs and their degradation products in different batches of granisetron injections. The method was proved to be sensitive for the determination down to 0.03 and 0.01% of granisetron degradation product and benzaldehyde, respectively, which are far below the compendia limits for testing these degradation products in their corresponding intact drugs. Copyright 2010 Elsevier B.V. All rights reserved.
Phase and flow behavior of mixed gas hydrate systems during gas injection
NASA Astrophysics Data System (ADS)
Darnell, K.; Flemings, P. B.; DiCarlo, D. A.
2017-12-01
We present one-dimensional, multi-phase flow model results for injections of carbon dioxide and nitrogen mixtures, or flue gas, into methane hydrate bearing reservoirs. Our flow model is coupled to a thermodynamic simulator that predicts phase stabilities as a function of composition, so multiple phases can appear, disappear, or change composition as the injection invades the reservoir. We show that the coupling of multi-phase fluid flow with phase behavior causes preferential phase fractionation in which each component flows through the system at different speeds and in different phases. We further demonstrate that phase and flow behavior within the reservoir are driven by hydrate stability of each individual component in addition to the hydrate stability of the injection composition. For example, if carbon dioxide and nitrogen are both individually hydrate stable at the reservoir P-T conditions, then any injection composition will convert all available water into hydrate and plug the reservoir. In contrast, if only carbon dioxide is hydrate stable at the reservoir P-T conditions, then nitrogen preferentially stays in the gaseous phase, while the carbon dioxide partitions into the hydrate and liquid water phases. For all injections of this type, methane originally held in hydrate is released by dissociation into the nitrogen-rich gaseous phase. The net consequence is that a gas phase composed of nitrogen and methane propagates through the reservoir in a fast-moving front. A slower-moving front lags behind where carbon dioxide and nitrogen form a mixed hydrate, but methane is absent due to dissociation-induced methane stripping from the first, fast-moving front. The entire composition path traces through the phase space as the flow develops with each front moving at different, constant velocities. This behavior is qualitatively similar to the dynamics present in enhanced oil recovery or enhanced coalbed methane recovery. These results explain why the inclusion of nitrogen in mixed gas injection into methane hydrate reservoirs has been far more successful at producing methane than pure carbon dioxide injections. These results also provide a test for the validity of equilibrium thermodynamics in transport-dominated mixed hydrate systems that can be validated by laboratory-scale flow-through experiments.
WEB-GIS Decision Support System for CO2 storage
NASA Astrophysics Data System (ADS)
Gaitanaru, Dragos; Leonard, Anghel; Radu Gogu, Constantin; Le Guen, Yvi; Scradeanu, Daniel; Pagnejer, Mihaela
2013-04-01
Environmental decision support systems (DSS) paradigm evolves and changes as more knowledge and technology become available to the environmental community. Geographic Information Systems (GIS) can be used to extract, assess and disseminate some types of information, which are otherwise difficult to access by traditional methods. In the same time, with the help of the Internet and accompanying tools, creating and publishing online interactive maps has become easier and rich with options. The Decision Support System (MDSS) developed for the MUSTANG (A MUltiple Space and Time scale Approach for the quaNtification of deep saline formations for CO2 storaGe) project is a user friendly web based application that uses the GIS capabilities. MDSS can be exploited by the experts for CO2 injection and storage in deep saline aquifers. The main objective of the MDSS is to help the experts to take decisions based large structured types of data and information. In order to achieve this objective the MDSS has a geospatial objected-orientated database structure for a wide variety of data and information. The entire application is based on several principles leading to a series of capabilities and specific characteristics: (i) Open-Source - the entire platform (MDSS) is based on open-source technologies - (1) database engine, (2) application server, (3) geospatial server, (4) user interfaces, (5) add-ons, etc. (ii) Multiple database connections - MDSS is capable to connect to different databases that are located on different server machines. (iii)Desktop user experience - MDSS architecture and design follows the structure of a desktop software. (iv)Communication - the server side and the desktop are bound together by series functions that allows the user to upload, use, modify and download data within the application. The architecture of the system involves one database and a modular application composed by: (1) a visualization module, (2) an analysis module, (3) a guidelines module, and (4) a risk assessment module. The Database component is build by using the PostgreSQL and PostGIS open source technology. The visualization module allows the user to view data of CO2 injection sites in different ways: (1) geospatial visualization, (2) table view, (3) 3D visualization. The analysis module will allow the user to perform certain analysis like Injectivity, Containment and Capacity analysis. The Risk Assessment module focus on the site risk matrix approach. The Guidelines module contains the methodologies of CO2 injection and storage into deep saline aquifers guidelines.
Toxicity ForeCaster (ToxCast™) Data
Data is organized into different data sets and includes descriptions of ToxCast chemicals and assays and files summarizing the screening results, a MySQL database, chemicals screened through Tox21, and available data generated from animal toxicity studies.
NASA Astrophysics Data System (ADS)
Barouchou, Alexandra; Dendrinos, Markos
2015-02-01
An interesting issue in the domain of history of science and ideas is the concept of similarity of historical personalities. Similar objects of research of philosophers and scientists indicate prospective influences, caused either from one another's reading or meetings, communication or even cooperation. Key methodological role in the surfacing of the sought similarities play the keywords extracted from their works as well as their placement in a philosophical and scientific term taxonomy. The case study examined in the framework of this paper concerns scientists and philosophers, who lived in ancient Greece or Renaissance periods and dealt, in at least one work, with the subject God. All the available data (scientists, studies, recorded relations between scientists, keywords, and thematic hierarchy) have been organized in an RDBMS environment, aiming at the emergence of similarities and influences between scientists through properly created SQL queries based on date and thematic hierarchy criteria.
STINGRAY: system for integrated genomic resources and analysis.
Wagner, Glauber; Jardim, Rodrigo; Tschoeke, Diogo A; Loureiro, Daniel R; Ocaña, Kary A C S; Ribeiro, Antonio C B; Emmel, Vanessa E; Probst, Christian M; Pitaluga, André N; Grisard, Edmundo C; Cavalcanti, Maria C; Campos, Maria L M; Mattoso, Marta; Dávila, Alberto M R
2014-03-07
The STINGRAY system has been conceived to ease the tasks of integrating, analyzing, annotating and presenting genomic and expression data from Sanger and Next Generation Sequencing (NGS) platforms. STINGRAY includes: (a) a complete and integrated workflow (more than 20 bioinformatics tools) ranging from functional annotation to phylogeny; (b) a MySQL database schema, suitable for data integration and user access control; and (c) a user-friendly graphical web-based interface that makes the system intuitive, facilitating the tasks of data analysis and annotation. STINGRAY showed to be an easy to use and complete system for analyzing sequencing data. While both Sanger and NGS platforms are supported, the system could be faster using Sanger data, since the large NGS datasets could potentially slow down the MySQL database usage. STINGRAY is available at http://stingray.biowebdb.org and the open source code at http://sourceforge.net/projects/stingray-biowebdb/.
Kilintzis, Vassilis; Beredimas, Nikolaos; Chouvarda, Ioanna
2014-01-01
An integral part of a system that manages medical data is the persistent storage engine. For almost twenty five years Relational Database Management Systems(RDBMS) were considered the obvious decision, yet today new technologies have emerged that require our attention as possible alternatives. Triplestores store information in terms of RDF triples without necessarily binding to a specific predefined structural model. In this paper we present an attempt to compare the performance of Apache JENA-Fuseki and the Virtuoso Universal Server 6 triplestores with that of MySQL 5.6 RDBMS for storing and retrieving medical information that it is communicated as RDF/XML ontology instances over a RESTful web service. The results show that the performance, calculated as average time of storing and retrieving instances, is significantly better using Virtuoso Server while MySQL performed better than Fuseki.
Data Processing on Database Management Systems with Fuzzy Query
NASA Astrophysics Data System (ADS)
Şimşek, Irfan; Topuz, Vedat
In this study, a fuzzy query tool (SQLf) for non-fuzzy database management systems was developed. In addition, samples of fuzzy queries were made by using real data with the tool developed in this study. Performance of SQLf was tested with the data about the Marmara University students' food grant. The food grant data were collected in MySQL database by using a form which had been filled on the web. The students filled a form on the web to describe their social and economical conditions for the food grant request. This form consists of questions which have fuzzy and crisp answers. The main purpose of this fuzzy query is to determine the students who deserve the grant. The SQLf easily found the eligible students for the grant through predefined fuzzy values. The fuzzy query tool (SQLf) could be used easily with other database system like ORACLE and SQL server.
2017-01-01
Reusing the data from healthcare information systems can effectively facilitate clinical trials (CTs). How to select candidate patients eligible for CT recruitment criteria is a central task. Related work either depends on DBA (database administrator) to convert the recruitment criteria to native SQL queries or involves the data mapping between a standard ontology/information model and individual data source schema. This paper proposes an alternative computer-aided CT recruitment paradigm, based on syntax translation between different DSLs (domain-specific languages). In this paradigm, the CT recruitment criteria are first formally represented as production rules. The referenced rule variables are all from the underlying database schema. Then the production rule is translated to an intermediate query-oriented DSL (e.g., LINQ). Finally, the intermediate DSL is directly mapped to native database queries (e.g., SQL) automated by ORM (object-relational mapping). PMID:29065644
PACSY, a relational database management system for protein structure and chemical shift analysis
Lee, Woonghee; Yu, Wookyung; Kim, Suhkmann; Chang, Iksoo
2012-01-01
PACSY (Protein structure And Chemical Shift NMR spectroscopY) is a relational database management system that integrates information from the Protein Data Bank, the Biological Magnetic Resonance Data Bank, and the Structural Classification of Proteins database. PACSY provides three-dimensional coordinates and chemical shifts of atoms along with derived information such as torsion angles, solvent accessible surface areas, and hydrophobicity scales. PACSY consists of six relational table types linked to one another for coherence by key identification numbers. Database queries are enabled by advanced search functions supported by an RDBMS server such as MySQL or PostgreSQL. PACSY enables users to search for combinations of information from different database sources in support of their research. Two software packages, PACSY Maker for database creation and PACSY Analyzer for database analysis, are available from http://pacsy.nmrfam.wisc.edu. PMID:22903636
A UML Profile for Developing Databases that Conform to the Third Manifesto
NASA Astrophysics Data System (ADS)
Eessaar, Erki
The Third Manifesto (TTM) presents the principles of a relational database language that is free of deficiencies and ambiguities of SQL. There are database management systems that are created according to TTM. Developers need tools that support the development of databases by using these database management systems. UML is a widely used visual modeling language. It provides built-in extension mechanism that makes it possible to extend UML by creating profiles. In this paper, we introduce a UML profile for designing databases that correspond to the rules of TTM. We created the first version of the profile by translating existing profiles of SQL database design. After that, we extended and improved the profile. We implemented the profile by using UML CASE system StarUML™. We present an example of using the new profile. In addition, we describe problems that occurred during the profile development.
Development of noSQL data storage for the ATLAS PanDA Monitoring System
NASA Astrophysics Data System (ADS)
Potekhin, M.; ATLAS Collaboration
2012-06-01
For several years the PanDA Workload Management System has been the basis for distributed production and analysis for the ATLAS experiment at the LHC. Since the start of data taking PanDA usage has ramped up steadily, typically exceeding 500k completed jobs/day by June 2011. The associated monitoring data volume has been rising as well, to levels that present a new set of challenges in the areas of database scalability and monitoring system performance and efficiency. These challenges are being met with a R&D effort aimed at implementing a scalable and efficient monitoring data storage based on a noSQL solution (Cassandra). We present our motivations for using this technology, as well as data design and the techniques used for efficient indexing of the data. We also discuss the hardware requirements as they were determined by testing with actual data and realistic loads.
NASA Astrophysics Data System (ADS)
Dervos, D. A.; Skourlas, C.; Laiho, M.
2015-02-01
"DBTech VET Teachers" project is Leonardo da Vinci Multilateral Transfer of Innovation Project co-financed by the European Commission's Lifelong Learning Programme. The aim of the project is to renew the teaching of database technologies in VET (Vocational Education and Training) institutes, on the basis of the current and real needs of ICT industry in Europe. Training of the VET teachers is done with the systems used in working life and they are taught to guide students to learning by verifying. In this framework, a course module on SQL transactions is prepared and offered. In this paper we present and briefly discuss some qualitative/quantitative data collected from its first pilot offering to an international audience in Greece during May-June 2013. The questionnaire/evaluation results, and the types of participants who have attended the course offering, are presented. Conclusions are also presented.
GEOmetadb: powerful alternative search engine for the Gene Expression Omnibus
Zhu, Yuelin; Davis, Sean; Stephens, Robert; Meltzer, Paul S.; Chen, Yidong
2008-01-01
The NCBI Gene Expression Omnibus (GEO) represents the largest public repository of microarray data. However, finding data in GEO can be challenging. We have developed GEOmetadb in an attempt to make querying the GEO metadata both easier and more powerful. All GEO metadata records as well as the relationships between them are parsed and stored in a local MySQL database. A powerful, flexible web search interface with several convenient utilities provides query capabilities not available via NCBI tools. In addition, a Bioconductor package, GEOmetadb that utilizes a SQLite export of the entire GEOmetadb database is also available, rendering the entire GEO database accessible with full power of SQL-based queries from within R. Availability: The web interface and SQLite databases available at http://gbnci.abcc.ncifcrf.gov/geo/. The Bioconductor package is available via the Bioconductor project. The corresponding MATLAB implementation is also available at the same website. Contact: yidong@mail.nih.gov PMID:18842599
STINGRAY: system for integrated genomic resources and analysis
2014-01-01
Background The STINGRAY system has been conceived to ease the tasks of integrating, analyzing, annotating and presenting genomic and expression data from Sanger and Next Generation Sequencing (NGS) platforms. Findings STINGRAY includes: (a) a complete and integrated workflow (more than 20 bioinformatics tools) ranging from functional annotation to phylogeny; (b) a MySQL database schema, suitable for data integration and user access control; and (c) a user-friendly graphical web-based interface that makes the system intuitive, facilitating the tasks of data analysis and annotation. Conclusion STINGRAY showed to be an easy to use and complete system for analyzing sequencing data. While both Sanger and NGS platforms are supported, the system could be faster using Sanger data, since the large NGS datasets could potentially slow down the MySQL database usage. STINGRAY is available at http://stingray.biowebdb.org and the open source code at http://sourceforge.net/projects/stingray-biowebdb/. PMID:24606808
Levy, C.; Beauchamp, C.
1996-01-01
This poster describes the methods used and working prototype that was developed from an abstraction of the relational model from the VA's hierarchical DHCP database. Overlaying the relational model on DHCP permits multiple user views of the physical data structure, enhances access to the database by providing a link to commercial (SQL based) software, and supports a conceptual managed care data model based on primary and longitudinal patient care. The goal of this work was to create a relational abstraction of the existing hierarchical database; to construct, using SQL data definition language, user views of the database which reflect the clinical conceptual view of DHCP, and to allow the user to work directly with the logical view of the data using GUI based commercial software of their choosing. The workstation is intended to serve as a platform from which a managed care information model could be implemented and evaluated.
Hosting and pulishing astronomical data in SQL databases
NASA Astrophysics Data System (ADS)
Galkin, Anastasia; Klar, Jochen; Riebe, Kristin; Matokevic, Gal; Enke, Harry
2017-04-01
In astronomy, terabytes and petabytes of data are produced by ground instruments, satellite missions and simulations. At Leibniz-Institute for Astrophysics Potsdam (AIP) we host and publish terabytes of cosmological simulation and observational data. The public archive at AIP has now reached a size of 60TB and growing and helps to produce numerous scientific papers. The web framework Daiquiri offers a dedicated web interface for each of the hosted scientific databases. Scientists all around the world run SQL queries which include specific astrophysical functions and get their desired data in reasonable time. Daiquiri supports the scientific projects by offering a number of administration tools such as database and user management, contact messages to the staff and support for organization of meetings and workshops. The webpages can be customized and the Wordpress integration supports the participating scientists in maintaining the documentation and the projects' news sections.
Zhang, Yinsheng; Zhang, Guoming; Shang, Qian
2017-01-01
Reusing the data from healthcare information systems can effectively facilitate clinical trials (CTs). How to select candidate patients eligible for CT recruitment criteria is a central task. Related work either depends on DBA (database administrator) to convert the recruitment criteria to native SQL queries or involves the data mapping between a standard ontology/information model and individual data source schema. This paper proposes an alternative computer-aided CT recruitment paradigm, based on syntax translation between different DSLs (domain-specific languages). In this paradigm, the CT recruitment criteria are first formally represented as production rules. The referenced rule variables are all from the underlying database schema. Then the production rule is translated to an intermediate query-oriented DSL (e.g., LINQ). Finally, the intermediate DSL is directly mapped to native database queries (e.g., SQL) automated by ORM (object-relational mapping).
DOE Office of Scientific and Technical Information (OSTI.GOV)
Khalili, Farid; Danilishin, Stefan; Mueller-Ebhardt, Helge
We consider enhancing the sensitivity of future gravitational-wave detectors by using double optical spring. When the power, detuning and bandwidth of the two carriers are chosen appropriately, the effect of the double optical spring can be described as a 'negative inertia', which cancels the positive inertia of the test masses and thus increases their response to gravitational waves. This allows us to surpass the free-mass standard quantum limit (SQL) over a broad frequency band, through signal amplification, rather than noise cancellation, which has been the case for all broadband SQL-beating schemes so far considered for gravitational-wave detectors. The merit ofmore » such signal amplification schemes lies in the fact that they are less susceptible to optical losses than noise-cancellation schemes. We show that it is feasible to demonstrate such an effect with the Gingin High Optical Power Test Facility, and it can eventually be implemented in future advanced GW detectors.« less
A semi-automated vascular access system for preclinical models
NASA Astrophysics Data System (ADS)
Berry-Pusey, B. N.; Chang, Y. C.; Prince, S. W.; Chu, K.; David, J.; Taschereau, R.; Silverman, R. W.; Williams, D.; Ladno, W.; Stout, D.; Tsao, T. C.; Chatziioannou, A.
2013-08-01
Murine models are used extensively in biological and translational research. For many of these studies it is necessary to access the vasculature for the injection of biologically active agents. Among the possible methods for accessing the mouse vasculature, tail vein injections are a routine but critical step for many experimental protocols. To perform successful tail vein injections, a high skill set and experience is required, leaving most scientists ill-suited to perform this task. This can lead to a high variability between injections, which can impact experimental results. To allow more scientists to perform tail vein injections and to decrease the variability between injections, a vascular access system (VAS) that semi-automatically inserts a needle into the tail vein of a mouse was developed. The VAS uses near infrared light, image processing techniques, computer controlled motors, and a pressure feedback system to insert the needle and to validate its proper placement within the vein. The VAS was tested by injecting a commonly used radiolabeled probe (FDG) into the tail veins of five mice. These mice were then imaged using micro-positron emission tomography to measure the percentage of the injected probe remaining in the tail. These studies showed that, on average, the VAS leaves 3.4% of the injected probe in the tail. With these preliminary results, the VAS system demonstrates the potential for improving the accuracy of tail vein injections in mice.
NASA Astrophysics Data System (ADS)
Clamens, Olivier; Lecerf, Johann; Hudelot, Jean-Pascal; Duc, Bertrand; Cadiou, Thierry; Blaise, Patrick; Biard, Bruno
2018-01-01
CABRI is an experimental pulse reactor, funded by the French Nuclear Safety and Radioprotection Institute (IRSN) and operated by CEA at the Cadarache research center. It is designed to study fuel behavior under RIA conditions. In order to produce the power transients, reactivity is injected by depressurization of a neutron absorber (3He) situated in transient rods inside the reactor core. The shapes of power transients depend on the total amount of reactivity injected and on the injection speed. The injected reactivity can be calculated by conversion of the 3He gas density into units of reactivity. So, it is of upmost importance to properly master gas density evolution in transient rods during a power transient. The 3He depressurization was studied by CFD calculations and completed with measurements using pressure transducers. The CFD calculations show that the density evolution is slower than the pressure drop. Surrogate models were built based on CFD calculations and validated against preliminary tests in the CABRI transient system. Studies also show that it is harder to predict the depressurization during the power transients because of neutron/3He capture reactions that induce a gas heating. This phenomenon can be studied by a multiphysics approach based on reaction rate calculation thanks to Monte Carlo code and study the resulting heating effect with the validated CFD simulation.
VSDMIP: virtual screening data management on an integrated platform
NASA Astrophysics Data System (ADS)
Gil-Redondo, Rubén; Estrada, Jorge; Morreale, Antonio; Herranz, Fernando; Sancho, Javier; Ortiz, Ángel R.
2009-03-01
A novel software (VSDMIP) for the virtual screening (VS) of chemical libraries integrated within a MySQL relational database is presented. Two main features make VSDMIP clearly distinguishable from other existing computational tools: (i) its database, which stores not only ligand information but also the results from every step in the VS process, and (ii) its modular and pluggable architecture, which allows customization of the VS stages (such as the programs used for conformer generation or docking), through the definition of a detailed workflow employing user-configurable XML files. VSDMIP, therefore, facilitates the storage and retrieval of VS results, easily adapts to the specific requirements of each method and tool used in the experiments, and allows the comparison of different VS methodologies. To validate the usefulness of VSDMIP as an automated tool for carrying out VS several experiments were run on six protein targets (acetylcholinesterase, cyclin-dependent kinase 2, coagulation factor Xa, estrogen receptor alpha, p38 MAP kinase, and neuraminidase) using nine binary (actives/inactive) test sets. The performance of several VS configurations was evaluated by means of enrichment factors and receiver operating characteristic plots.
Provenance for Runtime Workflow Steering and Validation in Computational Seismology
NASA Astrophysics Data System (ADS)
Spinuso, A.; Krischer, L.; Krause, A.; Filgueira, R.; Magnoni, F.; Muraleedharan, V.; David, M.
2014-12-01
Provenance systems may be offered by modern workflow engines to collect metadata about the data transformations at runtime. If combined with effective visualisation and monitoring interfaces, these provenance recordings can speed up the validation process of an experiment, suggesting interactive or automated interventions with immediate effects on the lifecycle of a workflow run. For instance, in the field of computational seismology, if we consider research applications performing long lasting cross correlation analysis and high resolution simulations, the immediate notification of logical errors and the rapid access to intermediate results, can produce reactions which foster a more efficient progress of the research. These applications are often executed in secured and sophisticated HPC and HTC infrastructures, highlighting the need for a comprehensive framework that facilitates the extraction of fine grained provenance and the development of provenance aware components, leveraging the scalability characteristics of the adopted workflow engines, whose enactment can be mapped to different technologies (MPI, Storm clusters, etc). This work looks at the adoption of W3C-PROV concepts and data model within a user driven processing and validation framework for seismic data, supporting also computational and data management steering. Validation needs to balance automation with user intervention, considering the scientist as part of the archiving process. Therefore, the provenance data is enriched with community-specific metadata vocabularies and control messages, making an experiment reproducible and its description consistent with the community understandings. Moreover, it can contain user defined terms and annotations. The current implementation of the system is supported by the EU-Funded VERCE (http://verce.eu). It provides, as well as the provenance generation mechanisms, a prototypal browser-based user interface and a web API built on top of a NoSQL storage technology, experimenting ways to ensure a rapid and flexible access to the lineage traces. It supports the users with the visualisation of graphical products and offers combined operations to access and download the data which may be selectively stored at runtime, into dedicated data archives.
Visualization of Vgi Data Through the New NASA Web World Wind Virtual Globe
NASA Astrophysics Data System (ADS)
Brovelli, M. A.; Kilsedar, C. E.; Zamboni, G.
2016-06-01
GeoWeb 2.0, laying the foundations of Volunteered Geographic Information (VGI) systems, has led to platforms where users can contribute to the geographic knowledge that is open to access. Moreover, as a result of the advancements in 3D visualization, virtual globes able to visualize geographic data even on browsers emerged. However the integration of VGI systems and virtual globes has not been fully realized. The study presented aims to visualize volunteered data in 3D, considering also the ease of use aspects for general public, using Free and Open Source Software (FOSS). The new Application Programming Interface (API) of NASA, Web World Wind, written in JavaScript and based on Web Graphics Library (WebGL) is cross-platform and cross-browser, so that the virtual globe created using this API can be accessible through any WebGL supported browser on different operating systems and devices, as a result not requiring any installation or configuration on the client-side, making the collected data more usable to users, which is not the case with the World Wind for Java as installation and configuration of the Java Virtual Machine (JVM) is required. Furthermore, the data collected through various VGI platforms might be in different formats, stored in a traditional relational database or in a NoSQL database. The project developed aims to visualize and query data collected through Open Data Kit (ODK) platform and a cross-platform application, where data is stored in a relational PostgreSQL and NoSQL CouchDB databases respectively.
Social media based NPL system to find and retrieve ARM data: Concept paper
DOE Office of Scientific and Technical Information (OSTI.GOV)
Devarakonda, Ranjeet; Giansiracusa, Michael T.; Kumar, Jitendra
Information connectivity and retrieval has a role in our daily lives. The most pervasive source of online information is databases. The amount of data is growing at rapid rate and database technology is improving and having a profound effect. Almost all online applications are storing and retrieving information from databases. One challenge in supplying the public with wider access to informational databases is the need for knowledge of database languages like Structured Query Language (SQL). Although the SQL language has been published in many forms, not everybody is able to write SQL queries. Another challenge is that it may notmore » be practical to make the public aware of the structure of the database. There is a need for novice users to query relational databases using their natural language. To solve this problem, many natural language interfaces to structured databases have been developed. The goal is to provide more intuitive method for generating database queries and delivering responses. Social media makes it possible to interact with a wide section of the population. Through this medium, and with the help of Natural Language Processing (NLP) we can make the data of the Atmospheric Radiation Measurement Data Center (ADC) more accessible to the public. We propose an architecture for using Apache Lucene/Solr [1], OpenML [2,3], and Kafka [4] to generate an automated query/response system with inputs from Twitter5, our Cassandra DB, and our log database. Using the Twitter API and NLP we can give the public the ability to ask questions of our database and get automated responses.« less
DOE Office of Scientific and Technical Information (OSTI.GOV)
Devarakonda, Ranjeet; Giansiracusa, Michael T.; Kumar, Jitendra
Information connectivity and retrieval has a role in our daily lives. The most pervasive source of online information is databases. The amount of data is growing at rapid rate and database technology is improving and having a profound effect. Almost all online applications are storing and retrieving information from databases. One challenge in supplying the public with wider access to informational databases is the need for knowledge of database languages like Structured Query Language (SQL). Although the SQL language has been published in many forms, not everybody is able to write SQL queries. Another challenge is that it may notmore » be practical to make the public aware of the structure of the database. There is a need for novice users to query relational databases using their natural language. To solve this problem, many natural language interfaces to structured databases have been developed. The goal is to provide more intuitive method for generating database queries and delivering responses. Social media makes it possible to interact with a wide section of the population. Through this medium, and with the help of Natural Language Processing (NLP) we can make the data of the Atmospheric Radiation Measurement Data Center (ADC) more accessible to the public. We propose an architecture for using Apache Lucene/Solr [1], OpenML [2,3], and Kafka [4] to generate an automated query/response system with inputs from Twitter5, our Cassandra DB, and our log database. Using the Twitter API and NLP we can give the public the ability to ask questions of our database and get automated responses.« less
DOE Office of Scientific and Technical Information (OSTI.GOV)
Liu, Yong-Liang; Department of Chemistry and Chemical Engineering, Shaanxi Key Laboratory of Comprehensive Utilization of Tailings Resources, Shang Luo University, Shang Luo 726000; Wu, Ya-Pan
2015-03-15
Two new interpenetrating Cu{sup II}/Ni{sup II} coordination polymers, based on a unsymmetrical bifunctional N/O-tectonic 3-(pyrid-4′-yl)-5-(4″-carbonylphenyl)-1,2,4-triazolyl (H{sub 2}pycz), ([Cu-(Hpycz){sub 2}]·2H{sub 2}O){sub n} (1) and ([Ni(Hpycz){sub 2}]·H{sub 2}O){sub n} (2), have been solvothermally synthesized and structure characterization. Single crystal X-ray analysis indicates that compound 1 shows 2-fold parallel interpenetrated 4{sup 4}-sql layers with the same handedness. The overall structure of 1 is achiral—in each layer of doubly interpenetrating nets, the two individual nets have the opposite handedness to the corresponding nets in the adjoining layers—while 2 features a rare 8-fold interpenetrating 6{sup 6}-dia network that belongs to class IIIa interpenetration. In addition,more » compounds 1 and 2 both show similar paramagnetic characteristic properties. - Graphical abstract: Two new Cu(II)/Ni(II) coordination polymers present 2D parallel 2-fold interpenetrated 4{sup 4}-sql layers and a rare 3D 8-fold interpenetrating 6{sup 6}-dia network. In addition, magnetic susceptibility measurements show similar paramagnetic characteristic for two complexes. - Highlights: • A new unsymmetrical bifunctional N/O-tectonic as 4-connected spacer. • A 2-fold parallel interpenetrated sql layer with the same handedness. • A rare 8-fold interpenetrating dia network (class IIIa)« less
Fritz, Jan; Niemeyer, Thomas; Clasen, Stephan; Wiskirchen, Jakub; Tepe, Gunnar; Kastler, Bruno; Nägele, Thomas; König, Claudius W; Claussen, Claus D; Pereira, Philippe L
2007-01-01
If low back pain does not improve with conservative management, the cause of the pain must be determined before further therapy is initiated. Information obtained from the patient's medical history, physical examination, and imaging may suffice to rule out many common causes of chronic pain (eg, fracture, malignancy, visceral or metabolic abnormality, deformity, inflammation, and infection). However, in most cases, the initial clinical and imaging findings have a low predictive value for the identification of specific pain-producing spinal structures. Diagnostic spinal injections performed in conjunction with imaging may be necessary to test the hypothesis that a particular structure is the source of pain. To ensure a valid test result, diagnostic injection procedures should be monitored with fluoroscopy, computed tomography, or magnetic resonance imaging. The use of controlled and comparative injections helps maximize the reliability of the test results. After a symptomatic structure has been identified, therapeutic spinal injections may be administered as an adjunct to conservative management, especially in patients with inoperable conditions. Therapeutic injections also may help hasten the recovery of patients with persistent or recurrent pain after spinal surgery. RSNA, 2007
DOE Office of Scientific and Technical Information (OSTI.GOV)
Nguyen, Ba Nghiep; Kunc, Vlastimil; Jin, Xiaoshi
2013-12-18
This article illustrates the predictive capabilities for long-fiber thermoplastic (LFT) composites that first simulate the injection molding of LFT structures by Autodesk® Simulation Moldflow® Insight (ASMI) to accurately predict fiber orientation and length distributions in these structures. After validating fiber orientation and length predictions against the experimental data, the predicted results are used by ASMI to compute distributions of elastic properties in the molded structures. In addition, local stress-strain responses and damage accumulation under tensile loading are predicted by an elastic-plastic damage model of EMTA-NLA, a nonlinear analysis tool implemented in ABAQUS® via user-subroutines using an incremental Eshelby-Mori-Tanaka approach. Predictedmore » stress-strain responses up to failure and damage accumulations are compared to the experimental results to validate the model.« less
Magarelli, N; Palmieri, D; Ottaviano, L; Savastano, M; Barbato, M; Leone, A; Maggialetti, A; Ciampa, F P; Bonomo, L
2008-01-01
Hyaluronic Acid (HA) is an alternative method for the treatment of osteoarthritis (OA), which acts on pain through a double action: anti-inflammatory and synovial fluid (SF) visco-supplementation. Magnetic Resonance Imaging (MRI), utilizing specific sequences, is a valid method for studying the initial phase of chondral damage. The analysis of the data, obtained through the intensity of values taken by positioning Region of Interest (ROIs) within the lesion, determining the differences before and after treatment with HA injected into the knee. The results obtained after six months and one year from the injection were statistically different in respect to those taken before, immediately and after three months of treatment. MRI represents a valid tool to evaluate the grade of chondromalacia patellae and also to follow the cartilage modification induced by HA therapy.
Pribbenow, Susanne; Wachter, Bettina; Ludwig, Carsten; Weigold, Annika; Dehnhard, Martin
2016-03-01
In mammals, the sex hormone testosterone is the major endocrine variable to objectify testicular activity and thus reproductive function in males. Testosterone is involved in the development and function of male reproductive physiology and sex-related behaviour. The development of a reliable androgen enzyme-immunoassay (EIA) to monitor faecal testosterone metabolites (fTM) is a powerful tool to non-invasively assess the gonadal status of males. We validated an epiandrosterone EIA for male cheetahs by performing a testosterone radiometabolism study followed by high-performance liquid chromatography (HPLC) analyses and excluding possible cross-reactivities with androgenic metabolites not derived from testosterone metabolism. The physiological and biological relevance of the epiandrosterone EIA was validated by demonstrating (1) a significant increase in fTM concentrations within one day in response to a testosterone injection, (2) a significant increase in fTM concentrations within one day in response to a gonadotropin-releasing hormone (GnRH) injection, which failed following a placebo injection, and (3) significant differences in fTM concentrations between adult male and adult female cheetahs and between adult and juvenile male cheetahs of a free-ranging population. Finally, we demonstrated stability of fTM concentrations measured in faecal samples exposed to ambient temperatures up to 72h. Our results clearly demonstrate that the epiandrosterone EIA is a reliable non-invasive method to monitor testicular activity in male cheetahs. Copyright © 2016 Elsevier Inc. All rights reserved.
Sentinel lymph nodes in endometrial cancer: is hysteroscopic injection valid?
Clement, D; Bats, A S; Ghazzar-Pierquet, N; Le Frere Belda, M A; Larousserie, F; Nos, C; Lecuru, F
2008-01-01
We aimed to describe hysteroscopic peritumoral tracer injection for detecting sentinel lymph nodes (SLNs) in patients with endometrial cancer and to evaluate tolerance of the procedure, detection rate and location of SLNs. Five patients with early endometrial cancer underwent hysteroscopic radiotracer injection followed by lymphoscintigraphy, then by surgery with hysteroscopic peritumoral blue dye injection, and radioactivity measurement using an endoscopic handheld gamma probe. SLNs and other nodes were sent separately to the pathology laboratory. SLNs were evaluated by hematoxylin-eosin-saffron staining and, when negative, by immunohistochemistry. Tolerance of the injection by the patients was poor (mean visual analog scale score, 8/10). SLNs were detected in only two patients (external iliac and common iliac+paraaortic, respectively). Detection rates were 1/5 by radiotracer, 1/5 by dye, and 2/5 by the combined method. One SLN was involved in a patient whose other nodes were negative. In three patients no SLNs were found by radiotracer or blue dye. Of the 83 non sentinel nodes removed from these patients, none was involved. Hysteroscopic peritumoral injection may be more difficult than cervical injection and, in our experience, carries a lower SLN detection rate.
NASA Astrophysics Data System (ADS)
Zhang, Xiaowei; Xing, Peiqi; Geng, Xiujuan; Sun, Daofeng; Xiao, Zhenyu; Wang, Lei
2015-09-01
Eight new coordination polymers (CPs), namely, [Zn(1,2-mbix)(tbtpa)]n (1), [Co(1,2-mbix)(tbtpa)]n (2), [CdCl(1,2-mbix)(tbtpa)0.5]n (3), {[Cd(1,2-bix)(tbtpa)]·H2O}n (4), {[Cd0.5(1,2-bix)(tbtpa)0.5]·H2O}n (5), {[Co0.5(1,2-bix)(tbtpa)0.5]·2H2O}n (6), {[Co(1,2-bix)(tbtpa)]·H2O}n (7) and {[Co(1,2-bix)(tbtpa)]·Diox·2H2O}n (8), were synthesized under solvothermal conditions based on mix-ligand strategy (H2tbtpa=tetrabromoterephthalic acid and 1,2-mbix=1,2-bis((2-methyl-1H-imidazol-1-yl)methyl)benzene, 1,2-bix=1,2-bis(imidazol-1-ylmethyl)benzene). All of the CPs have been structurally characterized by single-crystal X-ray diffraction analyses and further characterized by elemental analyses, IR spectroscopy, powder X-ray diffraction (PXRD), and thermogravimetric analyses (TGA). X-ray diffraction analyses show that 1 and 2 are isotypics which have 2D highly undulated networks with (4,4)-sql topology with the existence of C-H ⋯Br interactions; for 3, it has a 2D planar network with (4,4)-sql topology with the occurrence of C-H ⋯Cl interactions other than C-H ⋯Br interactions; 4 shows a 3D 2-fold interpenetrated nets with rare 65·8-mok topology which has a self-catention property. As the same case as 1 and 2, 5 and 6 are also isostructural with planar layers with 44-sql topology which further assembled into 3D supramolecular structure through the interdigitated stacking fashion and the C-Br ⋯Cph interactions. As for 7, it has a 2D slightly undulated networks with (4,4)-sql topology which has one dimension channel. While 8 has a 2-fold interpenetrated networks with (3,4)-connect jeb topology with point symbol {63}{65·8}. And their structures can be tuned by conformations of bis(imidazol) ligands and solvent mixture. Besides, the TGA properties for all compounds and the luminescent properties for 1, 3, 4, 5 are discussed in detail.
Enhancing SAMOS Data Access in DOMS via a Neo4j Property Graph Database.
NASA Astrophysics Data System (ADS)
Stallard, A. P.; Smith, S. R.; Elya, J. L.
2016-12-01
The Shipboard Automated Meteorological and Oceanographic System (SAMOS) initiative provides routine access to high-quality marine meteorological and near-surface oceanographic observations from research vessels. The Distributed Oceanographic Match-Up Service (DOMS) under development is a centralized service that allows researchers to easily match in situ and satellite oceanographic data from distributed sources to facilitate satellite calibration, validation, and retrieval algorithm development. The service currently uses Apache Solr as a backend search engine on each node in the distributed network. While Solr is a high-performance solution that facilitates creation and maintenance of indexed data, it is limited in the sense that its schema is fixed. The property graph model escapes this limitation by creating relationships between data objects. The authors will present the development of the SAMOS Neo4j property graph database including new search possibilities that take advantage of the property graph model, performance comparisons with Apache Solr, and a vision for graph databases as a storage tool for oceanographic data. The integration of the SAMOS Neo4j graph into DOMS will also be described. Currently, Neo4j contains spatial and temporal records from SAMOS which are modeled into a time tree and r-tree using Graph Aware and Spatial plugin tools for Neo4j. These extensions provide callable Java procedures within CYPHER (Neo4j's query language) that generate in-graph structures. Once generated, these structures can be queried using procedures from these libraries, or directly via CYPHER statements. Neo4j excels at performing relationship and path-based queries, which challenge relational-SQL databases because they require memory intensive joins due to the limitation of their design. Consider a user who wants to find records over several years, but only for specific months. If a traditional database only stores timestamps, this type of query would be complex and likely prohibitively slow. Using the time tree model, one can specify a path from the root to the data which restricts resolutions to certain timeframes (e.g., months). This query can be executed without joins, unions, or other compute-intensive operations, putting Neo4j at a computational advantage to the SQL database alternative.
DRUMS: Disk Repository with Update Management and Select option for high throughput sequencing data
2014-01-01
Background New technologies for analyzing biological samples, like next generation sequencing, are producing a growing amount of data together with quality scores. Moreover, software tools (e.g., for mapping sequence reads), calculating transcription factor binding probabilities, estimating epigenetic modification enriched regions or determining single nucleotide polymorphism increase this amount of position-specific DNA-related data even further. Hence, requesting data becomes challenging and expensive and is often implemented using specialised hardware. In addition, picking specific data as fast as possible becomes increasingly important in many fields of science. The general problem of handling big data sets was addressed by developing specialized databases like HBase, HyperTable or Cassandra. However, these database solutions require also specialized or distributed hardware leading to expensive investments. To the best of our knowledge, there is no database capable of (i) storing billions of position-specific DNA-related records, (ii) performing fast and resource saving requests, and (iii) running on a single standard computer hardware. Results Here, we present DRUMS (Disk Repository with Update Management and Select option), satisfying demands (i)-(iii). It tackles the weaknesses of traditional databases while handling position-specific DNA-related data in an efficient manner. DRUMS is capable of storing up to billions of records. Moreover, it focuses on optimizing relating single lookups as range request, which are needed permanently for computations in bioinformatics. To validate the power of DRUMS, we compare it to the widely used MySQL database. The test setting considers two biological data sets. We use standard desktop hardware as test environment. Conclusions DRUMS outperforms MySQL in writing and reading records by a factor of two up to a factor of 10000. Furthermore, it can work with significantly larger data sets. Our work focuses on mid-sized data sets up to several billion records without requiring cluster technology. Storing position-specific data is a general problem and the concept we present here is a generalized approach. Hence, it can be easily applied to other fields of bioinformatics. PMID:24495746
MAZZOLA, R.F.; CANTARELLA, G.; TORRETTA, S.; SBARBATI, A.; LAZZARI, L.; PIGNATARO, L.
2011-01-01
SUMMARY Minimally-invasive autologous fat injection of the head and neck region can be considered a valid alternative to major invasive surgical procedures both for aesthetic and functional purposes. The favourable outcomes of autologous fat injection in otolaryngological practice are due to the filling of soft tissue and, mainly, to the potential regenerative effect of adipose-derived mesenchymal stem cells. Herewith, some important biological preliminary remarks are described underlying the potential of autologous fat injection in regenerative medicine, and personal experience in using it for both consolidated clinical applications, such as fat grafting to the face and vocal fold augmentation in the treatment of glottic incompetence, and more recent applications including the treatment of post-parotidectomy Frey syndrome and velopharyngeal insufficiency. PMID:22058586
NASA Astrophysics Data System (ADS)
Martin, Ffion A.; Warrior, Nicholas A.; Simacek, Pavel; Advani, Suresh; Hughes, Adrian; Darlington, Roger; Senan, Eissa
2018-03-01
Very short manufacture cycle times are required if continuous carbon fibre and epoxy composite components are to be economically viable solutions for high volume composite production for the automotive industry. Here, a manufacturing process variant of resin transfer moulding (RTM), targets a reduction of in-mould manufacture time by reducing the time to inject and cure components. The process involves two stages; resin injection followed by compression. A flow simulation methodology using an RTM solver for the process has been developed. This paper compares the simulation prediction to experiments performed using industrial equipment. The issues encountered during the manufacturing are included in the simulation and their sensitivity to the process is explored.
Bidirectional light-scattering image processing method for high-concentration jet sprays
NASA Astrophysics Data System (ADS)
Shimizu, I.; Emori, Y.; Yang, W.-J.; Shimoda, M.; Suzuki, T.
1985-01-01
In order to study the distributions of droplet size and volume density in high-concentration jet sprays, a new technique is developed, which combines the forward and backward light scattering method and an image processing method. A pulsed ruby laser is used as the light source. The Mie scattering theory is applied to the results obtained from image processing on the scattering photographs. The time history is obtained for the droplet size and volume density distributions, and the method is demonstrated by diesel fuel sprays under various injecting conditions. The validity of the technique is verified by a good agreement in the injected fuel volume distributions obtained by the present method and by injection rate measurements.
NASA Technical Reports Server (NTRS)
Fletcher, D. G.; Mcdaniel, J. C.
1987-01-01
A preliminary quantitative study of the compressible flowfield in a steady, nonreacting model SCRAMJET combustor using laser-induced iodine fluorescence (LIIF) is reported. Measurements of density, temperature, and velocity were conducted with the calibrated, nonintrusive, optical technique for two different combustor operating conditions. First, measurements were made in the supersonic flow over a rearward-facing step without transverse injection for comparison with calculated pressure profiles. The second configuration was staged injection behind the rearward-facing step at an injection dynamic pressure ratio of 1.06. These experimental results will be used to validate computational fluid dynamic (CFD) codes being developed to model supersonic combustor flowfields.
Asymmetric injection and distribution of space charges in propylene carbonate under impulse voltage
NASA Astrophysics Data System (ADS)
Sima, Wenxia; Chen, Qiulin; Sun, Potao; Yang, Ming; Guo, Hongda; Ye, Lian
2018-05-01
Space charge can distort the electric field in high voltage stressed liquid dielectrics and lead to breakdown. Observing the evolution of space charge in real time and determining the influencing factors are of considerable significance. The spatio-temporal evolution of space charge in propylene carbonate, which is very complex under impulse voltage, was measured in this study through the time-continuous Kerr electro-optic field mapping measurement. We found that the injection charge from a brass electrode displayed an asymmetric effect; that is, the negative charge injection near the cathode lags behind the positive charge injection near the anode. Physical mechanisms, including charge generation and drift, are analyzed, and a voltage-dependent saturated drift rectification model was established to explain the interesting phenomena. Mutual validation of models and our measurement data indicated that a barrier layer, which is similar to metal-semiconductor contact, was formed in the contact interface between the electrode and propylene carbonate and played an important role in the space charge injection.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Cong, Yongzheng; Katipamula, Shanta; Geng, Tao
2016-02-01
A microfluidic platform was developed to perform online electrokinetic sample preconcentration and rapid hydrodynamic sample injection for electrophoresis using a single microvalve. The PDMS microchip consists of a separation channel, a side channel for sample introduction, and a control channel which is used as a pneumatic microvalve aligned at the intersection of the two flow channels. The closed microvalve, created by multilayer soft lithography, can serve as a preconcentrator under an applied electric potential, enabling current to pass through while blocking bulk flow. Once analytes are concentrated, the valve is briefly opened and the stacked sample is pressure injected intomore » the separation channel for electrophoretic separation. Fluorescently labeled peptides were enriched by a factor of ~450 in 230 s. The performance of the platform was validated by the online preconcentration, injection and electrophoretic separation of fluorescently labeled peptides. This method enables both rapid analyte concentration and controlled injection volume for high sensitivity, high resolution capillary electrophoresis.« less
Liu, Yanhong; Zhang, Weihua; Yang, Yuhui
2008-10-19
A hydrophilic interaction high performance liquid chromatography-tandem mass spectrometric method has been developed and validated for simultaneous quantification of dacarbazine (DTIC) and its terminal metabolite, 5-amino-4-imidazole-carboxamide (AIC) in human plasma. The plasma samples are first extracted by a C8+SCX mixed-mode 96-well plate to extend the extraction stability of DTIC and AIC. The extracted residues are further cleaned by a primary and secondary amine (PSA) adsorbent for minimization of matrix effect. Analyses are done on an Amide-80 HPLC column coupled to a tandem mass spectrometer fitted with an atmospheric pressure turbo ion spray ionization interface in the positive-ion mode. Both DTIC and AIC have reproducible retention times on the Amide-80 HPLC column. This type of column not only has an excellent column life (over 4000 injections), but also has zero carryover effect. The injection volume should be limited at 10 microL or less to avoid the peak splitting. The validated concentration ranges are from 0.5 to 500 ng/mL for DTIC and from 2.0 to 500 ng/mL for AIC. The validated method has been successfully applied to determine the pharmacokinetic profiles for human patients receiving DTIC infusions.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Gonzalez Gonzalez, R.; Petruzzi, A.; D'Auria, F.
2012-07-01
Atucha-2 is a Siemens-designed PHWR reactor under construction in the Republic of Argentina. Its geometrical complexity and (e.g., oblique Control Rods, Positive Void coefficient) required a developed and validated complex three dimensional (3D) neutron kinetics (NK) coupled thermal hydraulic (TH) model. Reactor shut-down is obtained by oblique CRs and, during accidental conditions, by an emergency shut-down system (JDJ) injecting a highly concentrated boron solution (boron clouds) in the moderator tank, the boron clouds reconstruction is obtained using a CFD (CFX) code calculation. A complete LBLOCA calculation implies the application of the RELAP5-3D{sup C} system code. Within the framework of themore » third Agreement 'NA-SA - Univ. of Pisa' a new RELAP5-3D control system for the boron injection system was developed and implemented in the validated coupled RELAP5-3D/NESTLE model of the Atucha 2 NPP. The aim of this activity is to find out the limiting case (maximum break area size) for the Peak Cladding Temperature for LOCAs under fixed boundary conditions. (authors)« less
Validated flow-injection method for rapid aluminium determination in anti-perspirants.
López-Gonzálvez, A; Ruiz, M A; Barbas, C
2008-09-29
A flow-injection (FI) method for the rapid determination of aluminium in anti-perspirants has been developed. The method is based on the spectrophotometric detection at 535nm of the complex formed between Al ions and the chromogenic reagent eriochrome cyanine R. Both the batch and FI methods were validated by checking the parameters included in the ISO-3543-1 regulation. Variables involved in the FI method were optimized by using appropriate statistical tools. The method does not exhibit interference from other substances present in anti-perspirants and it shows a high precision with a R.S.D. value (n=6) of 0.9%. Moreover, the accuracy of the method was evaluated by comparison with a back complexometric titration method, which is currently used for routine analysis in pharmaceutical laboratories. The Student's t-test showed that the results obtained by both methods were not significantly different for a significance level of 95%. A response time of 12s and a sample analysis time, by performing triplicate injections, of 60s were achieved. The analytical figures of merit make the method highly appropriate to substitute the time-consuming complexometric method for this kind of analysis.
Overview of HIT-SI3 experiment: Simulations, Diagnostics, and Summary of Current Results
NASA Astrophysics Data System (ADS)
Penna, James; Jarboe, Thomas; Nelson, Brian; Hossack, Aaron; Sutherland, Derek; Morgan, Kyle; Hansen, Chris; Benedett, Thomas; Everson, Chris; Victor, Brian
2016-10-01
The Helicity Injected Torus - Steady Inductive 3(HIT-SI3)experiment forms and maintains spheromaks via Steady Inductive Helicity Injection (SIHI). Three injector units allow for continuous injection of helicity into a copper flux conserver in order to sustain a spheromak. Firing of the injectors with a phase difference allows finite rotation of the plasma to provide a stabilizing effect. Simulations in the MHD code NIMROD and the fluid-model code PSI-TET provide validation and a basis for interpretation of the observed experimental data. Thompson Scattering (TS) and Far Infrared (FIR) Interferometer systems allow temperature and line-averaged density measurements to be taken. An Ion Doppler Spectroscopy (IDS) system allows measurement of the plasma rotation and velocity. HIT-SI3 data has been used for validation of IDCD predictions, in particular the projected impedance of helicity injectors according to the theory. The experimental impedances have been calculated here for the first time for different HIT-SI3 regimes. Such experimental evidence will contribute to the design of future experiments employing IDCD as a current-drive mechanism. Work supported by the D.O.E., Office of Science, Office of Fusion Science.
Tompkins, D. Andrew; Bigelow, George E.; Harrison, Joseph A.; Johnson, Rolley E.; Fudala, Paul J.; Strain, Eric C.
2009-01-01
Introduction The Clinical Opiate Withdrawal Scale (COWS) is an 11-item clinician-administered scale assessing opioid withdrawal. Though commonly used in clinical practice, it has not been systematically validated. The present study validated the COWS in comparison to the validated Clinical Institute Narcotic Assessment (CINA) scale. Method Opioid-dependent volunteers were enrolled in a residential trial and stabilized on morphine 30 mg given subcutaneously four times daily. Subjects then underwent double-blind, randomized challenges of intramuscularly administered placebo and naloxone (0.4 mg) on separate days, during which the COWS, CINA, and visual analog scale (VAS) assessments were concurrently obtained. Subjects completing both challenges were included (N=46). Correlations between mean peak COWS and CINA scores as well as self-report VAS questions were calculated. Results Mean peak COWS and CINA scores of 7.6 and 24.4, respectively, occurred on average 30 minutes post-injection of naloxone. Mean COWS and CINA scores 30 minutes after placebo injection were 1.3 and 18.9, respectively. The Pearson correlation coefficient for peak COWS and CINA scores during the naloxone challenge session was 0.85 (p<0.001). Peak COWS scores also correlated well with peak VAS self-report scores of bad drug effect (r=0.57, p<0.001) and feeling sick (r=0.57, p<0.001), providing additional evidence of concurrent validity. Placebo was not associated with any significant elevation of COWS, CINA, or VAS scores, indicating discriminant validity. Cronbach’s alpha for the COWS was 0.78, indicating good internal consistency (reliability). Discussion COWS, CINA, and certain VAS items are all valid measurement tools for acute opiate withdrawal. PMID:19647958
VIRmiRNA: a comprehensive resource for experimentally validated viral miRNAs and their targets.
Qureshi, Abid; Thakur, Nishant; Monga, Isha; Thakur, Anamika; Kumar, Manoj
2014-01-01
Viral microRNAs (miRNAs) regulate gene expression of viral and/or host genes to benefit the virus. Hence, miRNAs play a key role in host-virus interactions and pathogenesis of viral diseases. Lately, miRNAs have also shown potential as important targets for the development of novel antiviral therapeutics. Although several miRNA and their target repositories are available for human and other organisms in literature, but a dedicated resource on viral miRNAs and their targets are lacking. Therefore, we have developed a comprehensive viral miRNA resource harboring information of 9133 entries in three subdatabases. This includes 1308 experimentally validated miRNA sequences with their isomiRs encoded by 44 viruses in viral miRNA ' VIRMIRNA: ' and 7283 of their target genes in ' VIRMIRTAR': . Additionally, there is information of 542 antiviral miRNAs encoded by the host against 24 viruses in antiviral miRNA ' AVIRMIR': . The web interface was developed using Linux-Apache-MySQL-PHP (LAMP) software bundle. User-friendly browse, search, advanced search and useful analysis tools are also provided on the web interface. VIRmiRNA is the first specialized resource of experimentally proven virus-encoded miRNAs and their associated targets. This database would enhance the understanding of viral/host gene regulation and may also prove beneficial in the development of antiviral therapeutics. Database URL: http://crdd.osdd.net/servers/virmirna. © The Author(s) 2014. Published by Oxford University Press.
Weng, Jun; Li, Yang; Cai, Lei; Li, Ting; Peng, Gongze; Fu, Chaoyi; Han, Xu; Li, Haiyan; Jiang, Zesheng; Zhang, Zhi; Du, Jiang; Peng, Qing; Gao, Yi
2017-01-01
Background/Aims: The use of antibiotics to eliminate Mycoplasma contamination has some serious limitations. Mycoplasma contamination can be eliminated by intraperitoneal injection of BALB/c mice with contaminated cells combined with screening monoclonal cells. However, in vivo passage in mice after injection with contaminated cells requires a long duration (20-54 days). Furthermore, it is important to monitor for cross-contamination of mouse and human cells, xenotropic murine leukemia virus-related virus (XMRV) infection, and altered cell function after the in vivo treatment. The present study aimed to validate a reliable and simplified method to eliminate mycoplasma contamination from human hepatocytes. BALB/c mice were injected with paraffin oil prior to injection with cells, in order to shorten duration of intraperitoneal passage. Cross-contamination of mouse and human cells, XMRV infection and cell function-related genes and proteins were also evaluated. Methods: PCR and DNA sequencing were used to confirm Mycoplasma hyorhinis ( M. hyorhinis ) contamination in human hepatocyte C3A cells. Five BALB/c mice were intraperitoneally injected with 0.5 ml paraffin oil 1 week before injection of the cells. The mice were then intraperitoneally injected with C3A hepatocytes (5.0 × 10 6 /ml) contaminated with M. hyorhinis (6.2 ± 2.2 × 10 8 CFU/ml). Ascites were collected for monoclonal cell screening on the 14th day after injection of contaminated cells. Elimination of mycoplasma from cells was determined by PCR and Transmission Electron Microscopy (TEM). Human-mouse cell and XMRV contamination were also detected by PCR. Quantitative reverse transcription PCR and western blotting were used to compare the expression of genes and proteins among treated cells, non-treated infected cells, and uninfected cells. Results: Fourteen days after injection with cells, 4 of the 5 mice had ascites. Hepatocyte colonies extracted from the ascites of four mice were all mycoplasma-free. There was no cell cross-contamination or XMRV infection in treated cell cultures. Elimination of Mycoplasma resulted in partial or complete recovery in the expression of ALB, TF, and CYP3A4 genes as well as proteins. Proliferation of the treated cells was not significantly affected by this management. Conclusion: The method of elimination of Mycoplasma contamination in this study was validated and reproducible. Success was achieved in four of five cases examined. Compared to the previous studies, the duration of intraperitoneal passage in this study was significantly shorter.
Botulinum toxin treatment of false vocal folds in adductor spasmodic dysphonia: Functional outcomes.
Simpson, C Blake; Lee, Christopher T; Hatcher, Jeanne L; Michalek, Joel
2016-01-01
Supraglottic injection of botulinum toxin (Botox) has been described as an effective treatment for adductor spasmodic dysphonia (ADSD). Anecdotal evidence suggests that the patients have little to no breathiness after injection, but no formal longitudinal studies have been carried out to date. The purpose of this study was to examine the voice outcomes in patients with ADSD after supraglottic Botox injection. Retrospective case series. Patients with ADSD who were treated with supraglottic Botox injections completed a qualitative self-evaluation of voice function after injection using the percentage of normal function (PNF) scale, a validated, quantitative scale from 0% (no function) to 100% (normal function). Posttreatment voice improvement after injection was determined using a Voice Handicap Index-10 (VHI-10) questionnaire. A total of 198 supraglottic injections were performed between July 2011 and October 2014. Twenty-five questionnaires were completed. Mean postinjection PNF was 95.0% ± 8.4% and was significantly increased from the preinjection mean PNF (62.5%) ± 22.6% (P < 0.001). The mean best VHI-10 for all injections was 7.23. In 19 of 25 patients (76%), there was no reduction in PNF in the early postinjection period. In the remaining six patients (24%), the decline in mean vocal function was 9.2%. Supraglottic Botox injection is an effective treatment for ADSD. Postinjection voice is significantly improved, and the majority of patients do not experience breathy voice/decline in vocal function after injection. © 2015 The American Laryngological, Rhinological and Otological Society, Inc.
Direct injection GC method for measuring light hydrocarbon emissions from cooling-tower water.
Lee, Max M; Logan, Tim D; Sun, Kefu; Hurley, N Spencer; Swatloski, Robert A; Gluck, Steve J
2003-12-15
A Direct Injection GC method for quantifying low levels of light hydrocarbons (C6 and below) in cooling water has been developed. It is intended to overcome the limitations of the currently available technology. The principle of this method is to use a stripper column in a GC to strip waterfrom the hydrocarbons prior to entering the separation column. No sample preparation is required since the water sample is introduced directly into the GC. Method validation indicates that the Direct Injection GC method offers approximately 15 min analysis time with excellent precision and recovery. The calibration studies with ethylene and propylene show that both liquid and gas standards are suitable for routine calibration and calibration verification. The sampling method using zero headspace traditional VOA (Volatile Organic Analysis) vials and a sample chiller has also been validated. It is apparent that the sampling method is sufficient to minimize the potential for losses of light hydrocarbons, and samples can be held at 4 degrees C for up to 7 days with more than 93% recovery. The Direct Injection GC method also offers <1 ppb (w/v) level method detection limits for ethylene, propylene, and benzene. It is superior to the existing El Paso stripper method. In addition to lower detection limits for ethylene and propylene, the Direct Injection GC method quantifies individual light hydrocarbons in cooling water, provides better recoveries, and requires less maintenance and setup costs. Since the instrumentation and supplies are readily available, this technique could easily be established as a standard or alternative method for routine emission monitoring and leak detection of light hydrocarbons in cooling-tower water.
Fry, Craig L; Lintzeris, Nick
2003-02-01
To develop a standard measure of blood-borne virus transmission risk behaviour, and examine the underlying psychometric properties. The Blood-borne Virus Transmission Risk Assessment Questionnaire (BBV-TRAQ) was developed over three consecutive phases of the original BBV-TRAQ study in adherence to classical scale development procedures, culminating in the recruitment of a development sample of current injecting drug users via convenience and snowball sampling. Needle and syringe programmes (NSPs), medical clinics, alcohol/drug agencies, peer-based and outreach organizations across inner and outer metropolitan Melbourne. Two hundred and nine current injecting drug users. The mean age was 27 years, 68% were male, 65% unemployed, 36% with prison history and 25% in methadone maintenance. BBV-TRAQ items cover specific injecting, sexual and skin penetration risk practices. BBV-TRAQ characteristics were assessed via measures of internal and test-retest reliability; collateral validation; and principal components analyses. The BBV-TRAQ has satisfactory psychometric properties. Internal (a=0.87), test-retest (r=0.84) and inter-observer reliability results were high, suggesting that the instrument provides a reliable measure of BBV risk behaviour and is reliable over time and across interviewers. A principal components analysis with varimax rotation produced a parsimonious factor solution despite modest communality, and indicated that three factors (injecting, sex and skin penetration/hygiene risks) are required to describe BBV risk behaviour. The BBV-TRAQ is reliable and represents the first risk assessment tool to incorporate sufficient coverage of injecting, sex and other skin penetration risk practices to be considered truly content valid. The questionnaire is indicated for use in addictions research, clinical, peer education and BBV risk behaviour surveillance settings.
Senra, Hugo; Balaskas, Konstantinos; Mahmoodi, Neda; Aslam, Tariq
2017-05-01
To investigate detailed patient experiences specific to receiving vascular endothelial growth factor inhibitors (anti-VEGF) for wet age-related macular degeneration (wAMD), and to acquire a snapshot of the frequency of clinically significant levels of depression, anxiety, and posttraumatic stress among patients and levels of burden in patients' carers. Observational cross-sectional mixed-methods study. Three hundred patients with wAMD receiving anti-VEGF treatment and 100 patient carers were recruited. Qualitative data on patients' experience of treatment were collected using a structured survey. Standardized validated questionnaires were used to quantify clinically significant levels of anxiety, depression, and posttraumatic stress, as well as cognitive function and carers' burden. Qualitative data showed that 56% of patients (n = 132) reported anxiety related to anti-VEGF treatment. The main sources of anxiety were fear of going blind owing to intravitreal injections and concerns about treatment effectiveness, rather than around pain. From validated questionnaires, 17% of patients (n = 52) showed clinical levels of anxiety and 12% (n = 36) showed clinical levels of depression. Depression levels, but not anxiety, were significantly higher in patients who received up to 3 injections compared with patients who received from 4 to 12 injections (analysis of variance [ANOVA] P = .027) and compared with patients who received more than 12 injections (ANOVA P = .001). Anti-VEGF treatment is often experienced with some anxiety related to treatment, regardless of the number of injections received. Clinical levels of depression seem to be more frequent in patients at early stages of anti-VEGF treatment. Strategies to improve patient experience of treatment and minimize morbidity are suggested. Crown Copyright © 2017. Published by Elsevier Inc. All rights reserved.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Thi, Thanh Binh Nguyen; Morioka, Mizuki; Yokoyama, Atsushi
Numerical prediction of the fiber orientation in the short-glass fiber (GF) reinforced polyamide 6 (PA6) composites with the fiber weight concentration of 30%, 50%, and 70% manufactured by the injection molding process is presented. And the fiber orientation was also directly observed and measured through X-ray computed tomography. During the injection molding process of the short-fiber/thermoplastic composite, the fiber orientation is produced by the flow states and the fiber-fiber interaction. Folgar and Tucker equation is the well known for modeling the fiber orientation in a concentrated suspension. They included into Jeffrey’s equation a diffusive type of term by introducing amore » phenomenological coefficient to account for the fiber-fiber interaction. Our developed model for the fiber-fiber interaction was proposed by modifying the rotary diffusion term of the Folgar-Tucker equation. This model was presented in a conference paper of the 29{sup th} International Conference of the Polymer Processing Society published by AIP conference proceeding. For modeling fiber interaction, the fiber dynamic simulation was introduced in order to obtain a global fiber interaction coefficient, which is sum function of the fiber concentration, aspect ratio, and angular velocity. The fiber orientation is predicted by using the proposed fiber interaction model incorporated into a computer aided engineering simulation package C-Mold. An experimental program has been carried out in which the fiber orientation distribution has been measured in 100 x 100 x 2 mm injection-molded plate and 100 x 80 x 2 mm injection-molded weld by analyzed with a high resolution 3D X-ray computed tomography system XVA-160α, and calculated by X-ray computed tomography imaging. The numerical prediction shows a good agreement with experimental validation. And the complex fiber orientation in the injection-molded weld was investigated.« less
DOE Office of Scientific and Technical Information (OSTI.GOV)
Lambson, K; Lafata, K; Miles, D
Purpose: To validate the use of a PRESAGE dosimeter as a method to quantitatively measure dose distributions of injectable brachytherapy based on elastin-like polypeptide (ELP) nanoparticles. PRESAGE is a solid, transparent polyurethane-based dosimeter whose dose is proportional to a change in optical density, making it useful for visualizing the dose from a radionuclide-tagged-ELP injection. Methods: A PRESAGE dosimeter was designed to simulate an ELP injection. To calibrate, cuvette samples from the batch of PRESAGE were exposed to varying levels of radiation from 0–35.9Gy applied via a linear accelerator, then placed into a spectrophotometer to obtain the optical density change asmore » a function of dose. A pre-optical-CT scan was acquired of the phantom to obtain a baseline tomographic optical density. A 1cc saline solution of I-125 tagged-ELP with and activity concentration of 1mCi/cc was injected into the phantom and left for five days. After five days, the ELP was removed and the cavity cleaned of all remaining radioactive material. Post tomographic optical images were acquired to obtain a differential optical density dataset. Results: Initial results after the 5-day exposure revealed an opaque white film that resembled the volume of the ELP solution injected into the phantom. We think this is possibly due to the saline solution diffusing into the PRESAGE and causing a change in the index of refraction at this shallow depth. Therefore, initially the optical scanner yielded inconclusive results. After several more days, the saline seemed to have evaporated out of the injection site and the ELP dose distribution was visible via color change in the dosimeter. Conclusion: We have created the first experimental design to measure the dose distribution of I-125-tagged-ELP. The PRESAGE formulation proves to be a feasible option for such measurements. Future experimental measurements need to be obtained to further characterize ELP dosimetry.« less
NASA Astrophysics Data System (ADS)
Thi, Thanh Binh Nguyen; Morioka, Mizuki; Yokoyama, Atsushi; Hamanaka, Senji; Yamashita, Katsuhisa; Nonomura, Chisato
2015-05-01
Numerical prediction of the fiber orientation in the short-glass fiber (GF) reinforced polyamide 6 (PA6) composites with the fiber weight concentration of 30%, 50%, and 70% manufactured by the injection molding process is presented. And the fiber orientation was also directly observed and measured through X-ray computed tomography. During the injection molding process of the short-fiber/thermoplastic composite, the fiber orientation is produced by the flow states and the fiber-fiber interaction. Folgar and Tucker equation is the well known for modeling the fiber orientation in a concentrated suspension. They included into Jeffrey's equation a diffusive type of term by introducing a phenomenological coefficient to account for the fiber-fiber interaction. Our developed model for the fiber-fiber interaction was proposed by modifying the rotary diffusion term of the Folgar-Tucker equation. This model was presented in a conference paper of the 29th International Conference of the Polymer Processing Society published by AIP conference proceeding. For modeling fiber interaction, the fiber dynamic simulation was introduced in order to obtain a global fiber interaction coefficient, which is sum function of the fiber concentration, aspect ratio, and angular velocity. The fiber orientation is predicted by using the proposed fiber interaction model incorporated into a computer aided engineering simulation package C-Mold. An experimental program has been carried out in which the fiber orientation distribution has been measured in 100 x 100 x 2 mm injection-molded plate and 100 x 80 x 2 mm injection-molded weld by analyzed with a high resolution 3D X-ray computed tomography system XVA-160α, and calculated by X-ray computed tomography imaging. The numerical prediction shows a good agreement with experimental validation. And the complex fiber orientation in the injection-molded weld was investigated.
NASA Astrophysics Data System (ADS)
Basirat, Farzad; Perroud, Hervé; Lofi, Johanna; Denchik, Nataliya; Lods, Gérard; Fagerlund, Fritjof; Sharma, Prabhakar; Pezard, Philippe; Niemi, Auli
2015-04-01
In this study, TOUGH2/EOS7CA model is used to simulate the shallow injection-monitoring experiment carried out at Maguelone, France, during 2012 and 2013. The possibility of CO2 leakage from storage reservoir to upper layers is one of the issues that need to be addressed in CCS projects. Developing reliable monitoring techniques to detect and characterize CO2 leakage is necessary for the safety of CO2 storage in reservoir formations. To test and cross-validate different monitoring techniques, a series of shallow gas injection-monitoring experiments (SIMEx) has been carried out at the Maguelone. The experimental site is documented in Lofi et al [2013]. At the site, a series of nitrogen and one CO2 injection experiment have been carried out during 2012-2013 and different monitoring techniques have been applied. The purpose of modelling is to acquire understanding of the system performance as well as to further develop and validate modelling approaches for gas transport in the shallow subsurface, against the well-controlled data sets. The preliminary simulation of the experiment including the simulation for the Nitrogen injection test in 2012 was presented in Basirat et al [2013]. In this work, the simulations represent the gaseous CO2 distribution and dissolved CO2 within range obtained by monitoring approaches. The Multiphase modelling in combination with geophysical monitoring can be used for process understanding of gas phase migration- and mass transfer processes resulting from gaseous CO2 injection. Basirat, F., A. Niemi, H. Perroud, J. Lofi, N. Denchik, G. Lods, P. Pezard, P. Sharma, and F. Fagerlund (2013), Modeling Gas Transport in the Shallow Subsurface in Maguelone Field Experiment, Energy Procedia, 40, 337-345. Lofi, J., P. Pezard, F. Bouchette, O. Raynal, P. Sabatier, N. Denchik, A. Levannier, L. Dezileau, and R. Certain (2013), Integrated Onshore-Offshore Investigation of a Mediterranean Layered Coastal Aquifer, Groundwater, 51(4), 550-561.
Rittenberg, Lauren; Morrissey, Darlene; El-Khawand, Dominique; Whitmore, Kristene
2017-01-01
Introduction This study aims to evaluate the effectiveness of kenalog injection into Hunner's lesions. Materials and Methods All patients had cystoscopy and bladder hydrodistention with corticosteroid injection into Hunner's lesions over a 2.5-year period. Data include patient characteristics and pre- and post-operative validated questionnaires. Spearman Correlation and Wilcoxon t-tests were used for analysis. Results One hundred patients were reviewed retrospectively. There was a 1.1 point decrease in pain at 12 weeks post-operation (p = 0.435). Urinary frequency decreased from a mean of 11.7 to 9.1 daily episodes (p = 0.05), and nocturia from a mean of 3 to 1.6 nightly episodes (p = 0.008). Conclusion The use of a corticosteroid may be beneficial to symptom control and improvement in the quality of life of interstitial cystitis/painful bladder syndrome patients. Patients had improved frequency and nocturia 12 weeks post injection. PMID:28878600
Rittenberg, Lauren; Morrissey, Darlene; El-Khawand, Dominique; Whitmore, Kristene
2017-08-01
This study aims to evaluate the effectiveness of kenalog injection into Hunner's lesions. All patients had cystoscopy and bladder hydrodistention with corticosteroid injection into Hunner's lesions over a 2.5-year period. Data include patient characteristics and pre- and post-operative validated questionnaires. Spearman Correlation and Wilcoxon t -tests were used for analysis. One hundred patients were reviewed retrospectively. There was a 1.1 point decrease in pain at 12 weeks post-operation (p = 0.435). Urinary frequency decreased from a mean of 11.7 to 9.1 daily episodes (p = 0.05), and nocturia from a mean of 3 to 1.6 nightly episodes (p = 0.008). The use of a corticosteroid may be beneficial to symptom control and improvement in the quality of life of interstitial cystitis/painful bladder syndrome patients. Patients had improved frequency and nocturia 12 weeks post injection.
Gas-injection-start and shutdown characteristics of a 2-kilowatt to 15-kilowatt Brayton power system
NASA Technical Reports Server (NTRS)
Cantoni, D. A.
1972-01-01
Two methods of starting the Brayton power system have been considered: (1) using the alternator as a motor to spin the Brayton rotating unit (BRU), and (2) spinning the BRU by forced gas injection. The first method requires the use of an auxiliary electrical power source. An alternating voltage is applied to the terminals of the alternator to drive it as an induction motor. Only gas-injection starts are discussed in this report. The gas-injection starting method requires high-pressure gas storage and valves to route the gas flow to provide correct BRU rotation. An analog computer simulation was used to size hardware and to determine safe start and shutdown procedures. The simulation was also used to define the range of conditions for successful startups. Experimental data were also obtained under various test conditions. These data verify the validity of the start and shutdown procedures.
Menth-Meier, S E; Imoberdorf, R; Regli, B; Kipfer, B; Turgay, M; Ballmer, P E
2001-05-01
Validation of plasma volume (PV) determination by indocyanine green (ICG) in comparison to the gold-standard method with radioiodinated albumin, and investigation of the effect of commonly used plasma expanders (albumin, hydroxyethyl starch, and polygelatine) on PV in the early postoperative phase in patients undergoing cardiac surgery. Prospective clinical study. Department of medicine and intensive care unit at a university hospital. Ten healthy volunteers and 21 patients after elective open-heart surgery. PV of subjects was measured by i.v. injecting 5 microCi [125I]albumin (I-ALB). One hour later, PV was determined by a peripheral i. v. injection of 0.25 mg/kg body weight ICG (ICG1). In five subjects PV was measured repeatedly by ICG (ICG2) 1 h after ICG1. Mean PV of I-ALB and ICG1 or ICG2 showed consistent results. Further, we investigated central vs peripheral intravenous injection of ICG in six patients after open-heart surgery compared to [125I]albumin. There was no difference between mean PV measured by [125I]albumin and peripheral ICG (P = 0.40). PV determined by central injection of ICG was significantly higher than by the other methods. In 15 patients PV was determined by [125I]albumin. Thereafter, patients were randomly divided into three groups. Group ALB was infused with 1.75 ml/kg body weight human albumin 20%, group HAES with 5.25 ml/kg body weight hydroxyethyl starch 6%, and group HAEM with 7.0 ml/kg body weight polygelatine 3.5%. PV was measured 1 h and 4 h after infusion by ICG. There were no significant changes in PV between the groups. PV determination by peripheral i. v. injection of ICG produced reliable and consistent results when a reactive hyperaemia was produced by a tourniquet prior to injection. Therefore, central venous injection of ICG may not be prerequisite for precise measurements of PV. The expected acute increase in PV after infusion of commonly used plasma expanders after cardiac surgery was not found.
Quality Control of Injection Molded Eyewear by Non-Contact Deflectometry
NASA Astrophysics Data System (ADS)
Speck, A.; Zelzer, B.; Langenbucher, A.; Eppig, T.
2014-07-01
Occupational eye wear such as safety spectacles are manufactured by injection molding techniques. Testing of the assembled safety spectacle lenses in transmission is state of the art, but there is a lack of surface measurement systems for occupational safety lenses. The purpose of this work was to validate a deflectometric setup for topography measurement, detection of defects and visualization of the polishing quality, e.g. casting indentations or impressions, for the production process of safety spectacles. The setup is based on a customized stereo phase measuring deflectometer (PMD), equipped with 3 cameras with f'1,2 = 16 mm and f'3 = 8.5 mm and a specified measurement uncertainty of ± 3 μm. Sixteen plastic lenses and 8 corresponding injection molds from 4 parallel cavities were used for validation of the deflectometer. For comparison an interferometric method and a reference standard (< λ/10 super polished) was used. The accuracy and bias with a spherical safety spectacle sample was below 1 μm, according to DIN ISO 5725-2.2002-12. The repeatability was 2.1 μm and 35.7 μm for a blind radius fit. In conclusion, the PMD technique is an appropriate tool for characterizing occupational safety spectacle and injections mold surfaces. With the presented setup we were able to quantify the surface quality. This can be useful and may optimize the quality of the end product, in addition to standardized measuring systems in transmission.
Code of Federal Regulations, 2011 CFR
2011-10-01
... Microsoft Open Database Connectivity (ODBC) standard. ODBC is a Windows technology that allows a database software package to import data from a database created using a different software package. We currently...-compatible format. All databases must be supported with adequate documentation on data attributes, SQL...
SAMSON Technology Demonstrator
2014-06-01
requested. The SAMSON TD was testing with two different policy engines: 1. A custom XACML-based element matching engine using a MySQL database for...performed during the course of the event. Full information protection across the sphere of access management, information protection and auditing was in...
Business Demands for Web-Related Skills as Compared to Other Computer Skills.
ERIC Educational Resources Information Center
Groneman, Nancy
2000-01-01
Analysis of 23,704 want ads for computer-related jobs revealed that the most frequently mentioned skills were UNIX, SQL programming, and computer security. Curriculum implications were derived from the most desired and less frequently mentioned skills. (SK)
Scale-Independent Relational Query Processing
2013-10-04
source options are also available, including Postgresql, MySQL , and SQLite. These mod- ern relational databases are generally very complex software systems...and Their Application to Data Stream Management. IGI Global, 2010. [68] George Reese. Database Programming with JDBC and Java , Second Edition. Ed. by
Methods, Knowledge Support, and Experimental Tools for Modeling
2006-10-01
open source software entities: the PostgreSQL relational database management system (http://www.postgres.org), the Apache web server (http...past. The revision control system allows the program to capture disagreements, and allows users to explore the history of such disagreements by
Sessions-Bresnahan, D R; Graham, J K; Carnevale, E M
2014-07-15
IVF in horses is rarely successful. One reason for this could be the failure of sperm to fully capacitate or exhibit hyperactive motility. We hypothesized that the zona pellucida (ZP) of equine oocytes prevents fertilization in vitro, and bypassing the ZP would increase fertilization rates. Limited availability of equine oocytes for research has necessitated the use of heterologous oocyte binding assays using bovine oocytes. We sought to validate an assay using bovine oocytes and equine sperm and then to demonstrate that bypassing the ZP using perivitelline sperm injections (PVIs) with equine sperm capacitated with dilauroyl phosphatidylcholine would result in higher fertilization rates than standard IVF in bovine and equine oocytes. In experiment 1, bovine oocytes were used for (1) IVF with bovine sperm, (2) IVF with equine sperm, and (3) intracytoplasmic sperm injections (ICSIs) with equine sperm. Presumptive zygotes were either stained with 4',6-diamidino-2-phenylindole from 18 to 26 hours at 2-hour intervals or evaluated for cleavage at 56 hours after addition of sperm. Equine sperm fertilized bovine oocytes; however, pronuclei formation was delayed compared with bovine sperm after IVF. The delayed pronuclear formation was not seen after ICSI. In experiment 2, bovine oocytes were assigned to the following five groups: (1) cumulus oocyte complexes (COCs) coincubated with bovine sperm; (2) COC exposed to sucrose then coincubated with bovine sperm; (3) COC coincubated with equine sperm; (4) COC exposed to sucrose, and coincubated with equine sperm; and (5) oocytes exposed to sucrose, and 10 to 15 equine sperm injected into the perivitelline (PV) space. Equine sperm tended (P = 0.08) to fertilize more bovine oocytes when injected into the PV space than after IVF. In experiment 3, oocytes were assigned to the following four groups: (1) IVF, equine, and bovine COC coincubated with equine sperm; (2) PVI of equine and bovine oocytes; (3) PVI with equine oocytes pretreated with sucrose; and (4) ICSI of equine oocytes. Oocytes were examined at 24 hours for cleavage. No equine oocytes cleaved after IVF or PVI. However, ICSI conducted with equine sperm treated with dilauroyl phosphatidylcholine resulted in 85% of the oocytes cleaving. Sperm injected into the PV space of equine oocytes did not appear to enter the ooplasm. This study validated the use of bovine oocytes for equine sperm studies and indicates that failure of equine IVF is more than an inability of equine sperm to penetrate the ZP. Copyright © 2014 Elsevier Inc. All rights reserved.
Airport Remote Tower Sensor Systems
NASA Technical Reports Server (NTRS)
Maluf, David A.; Gawdiak, Yuri; Leidichj, Christopher; Papasin, Richard; Tran, Peter B.; Bass, Kevin
2006-01-01
Networks of video cameras, meteorological sensors, and ancillary electronic equipment are under development in collaboration among NASA Ames Research Center, the Federal Aviation Administration (FAA), and the National Oceanic Atmospheric Administration (NOAA). These networks are to be established at and near airports to provide real-time information on local weather conditions that affect aircraft approaches and landings. The prototype network is an airport-approach-zone camera system (AAZCS), which has been deployed at San Francisco International Airport (SFO) and San Carlos Airport (SQL). The AAZCS includes remotely controlled color video cameras located on top of SFO and SQL air-traffic control towers. The cameras are controlled by the NOAA Center Weather Service Unit located at the Oakland Air Route Traffic Control Center and are accessible via a secure Web site. The AAZCS cameras can be zoomed and can be panned and tilted to cover a field of view 220 wide. The NOAA observer can see the sky condition as it is changing, thereby making possible a real-time evaluation of the conditions along the approach zones of SFO and SQL. The next-generation network, denoted a remote tower sensor system (RTSS), will soon be deployed at the Half Moon Bay Airport and a version of it will eventually be deployed at Los Angeles International Airport. In addition to remote control of video cameras via secure Web links, the RTSS offers realtime weather observations, remote sensing, portability, and a capability for deployment at remote and uninhabited sites. The RTSS can be used at airports that lack control towers, as well as at major airport hubs, to provide synthetic augmentation of vision for both local and remote operations under what would otherwise be conditions of low or even zero visibility.
ClimateSpark: An In-memory Distributed Computing Framework for Big Climate Data Analytics
NASA Astrophysics Data System (ADS)
Hu, F.; Yang, C. P.; Duffy, D.; Schnase, J. L.; Li, Z.
2016-12-01
Massive array-based climate data is being generated from global surveillance systems and model simulations. They are widely used to analyze the environment problems, such as climate changes, natural hazards, and public health. However, knowing the underlying information from these big climate datasets is challenging due to both data- and computing- intensive issues in data processing and analyzing. To tackle the challenges, this paper proposes ClimateSpark, an in-memory distributed computing framework to support big climate data processing. In ClimateSpark, the spatiotemporal index is developed to enable Apache Spark to treat the array-based climate data (e.g. netCDF4, HDF4) as native formats, which are stored in Hadoop Distributed File System (HDFS) without any preprocessing. Based on the index, the spatiotemporal query services are provided to retrieve dataset according to a defined geospatial and temporal bounding box. The data subsets will be read out, and a data partition strategy will be applied to equally split the queried data to each computing node, and store them in memory as climateRDDs for processing. By leveraging Spark SQL and User Defined Function (UDFs), the climate data analysis operations can be conducted by the intuitive SQL language. ClimateSpark is evaluated by two use cases using the NASA Modern-Era Retrospective Analysis for Research and Applications (MERRA) climate reanalysis dataset. One use case is to conduct the spatiotemporal query and visualize the subset results in animation; the other one is to compare different climate model outputs using Taylor-diagram service. Experimental results show that ClimateSpark can significantly accelerate data query and processing, and enable the complex analysis services served in the SQL-style fashion.
Motamed, Cyrus; Bourgain, Jean Louis
2016-06-01
Anaesthesia Information Management Systems (AIMS) generate large amounts of data, which might be useful for quality assurance programs. This study was designed to highlight the multiple contributions of our AIMS system in extracting quality indicators over a period of 10years. The study was conducted from 2002 to 2011. Two methods were used to extract anaesthesia indicators: the manual extraction of individual files for monitoring neuromuscular relaxation and structured query language (SQL) extraction for other indicators which were postoperative nausea and vomiting (PONV), pain, sedation scores, pain-related medications, scores and postoperative hypothermia. For each indicator, a program of information/meetings and adaptation/suggestions for operating room and PACU personnel was initiated to improve quality assurance, while data were extracted each year. The study included 77,573 patients. The mean overall completeness of data for the initial years ranged from 55 to 85% and was indicator-dependent, which then improved to 95% completeness for the last 5years. The incidence of neuromuscular monitoring was initially 67% and then increased to 95% (P<0.05). The rate of pharmacological reversal remained around 53% throughout the study. Regarding SQL data, an improvement of severe postoperative pain and PONV scores was observed throughout the study, while mild postoperative hypothermia remained a challenge, despite efforts for improvement. The AIMS system permitted the follow-up of certain indicators through manual sampling and many more via SQL extraction in a sustained and non-time-consuming way across years. However, it requires competent and especially dedicated resources to handle the database. Copyright © 2016 Société française d'anesthésie et de réanimation (Sfar). Published by Elsevier Masson SAS. All rights reserved.
WebDB Component Builder - Lessons Learned
DOE Office of Scientific and Technical Information (OSTI.GOV)
Macedo, C.
2000-02-15
Oracle WebDB is the easiest way to produce web enabled lightweight and enterprise-centric applications. This concept from Oracle has tantalized our taste for simplistic web development by using a purely web based tool that lives nowhere else but in the database. The use of online wizards, templates, and query builders, which produces PL/SQL behind the curtains, can be used straight ''out of the box'' by both novice and seasoned developers. The topic of this presentation will introduce lessons learned by developing and deploying applications built using the WebDB Component Builder in conjunction with custom PL/SQL code to empower a hybridmore » application. There are two kinds of WebDB components: those that display data to end users via reporting, and those that let end users update data in the database via entry forms. The presentation will also discuss various methods within the Component Builder to enhance the applications pushed to the desktop. The demonstrated example is an application entitled HOME (Helping Other's More Effectively) that was built to manage a yearly United Way Campaign effort. Our task was to build an end to end application which could manage approximately 900 non-profit agencies, an average of 4,100 individual contributions, and $1.2 million dollars. Using WebDB, the shell of the application was put together in a matter of a few weeks. However, we did encounter some hurdles that WebDB, in it's stage of infancy (v2.0), could not solve for us directly. Together with custom PL/SQL, WebDB's Component Builder became a powerful tool that enabled us to produce a very flexible hybrid application.« less
Shen, Boxuan; Linko, Veikko; Dietz, Hendrik; Toppari, J Jussi
2015-01-01
DNA origami is a widely used method for fabrication of custom-shaped nanostructures. However, to utilize such structures, one needs to controllably position them on nanoscale. Here we demonstrate how different types of 3D scaffolded multilayer origamis can be accurately anchored to lithographically fabricated nanoelectrodes on a silicon dioxide substrate by DEP. Straight brick-like origami structures, constructed both in square (SQL) and honeycomb lattices, as well as curved "C"-shaped and angular "L"-shaped origamis were trapped with nanoscale precision and single-structure accuracy. We show that the positioning and immobilization of all these structures can be realized with or without thiol-linkers. In general, structural deformations of the origami during the DEP trapping are highly dependent on the shape and the construction of the structure. The SQL brick turned out to be the most robust structure under the high DEP forces, and accordingly, its single-structure trapping yield was also highest. In addition, the electrical conductivity of single immobilized plain brick-like structures was characterized. The electrical measurements revealed that the conductivity is negligible (insulating behavior). However, we observed that the trapping process of the SQL brick equipped with thiol-linkers tended to induce an etched "nanocanyon" in the silicon dioxide substrate. The nanocanyon was formed exactly between the electrodes, that is, at the location of the DEP-trapped origami. The results show that the demonstrated DEP-trapping technique can be readily exploited in assembling and arranging complex multilayered origami geometries. In addition, DNA origamis could be utilized in DEP-assisted deformation of the substrates onto which they are attached. © 2014 WILEY-VCH Verlag GmbH & Co. KGaA, Weinheim.
An improved resource management model based on MDS
NASA Astrophysics Data System (ADS)
Yuan, Man; Sun, Changying; Li, Pengfei; Sun, Yongdong; He, Rui
2005-11-01
GRID technology provides a kind of convenient method for managing GRID resources. This service is so-called monitoring, discovering service. This method is proposed by Globus Alliance, in this GRID environment, all kinds of resources, such as computational resources, storage resources and other resources can be organized by MDS specifications. However, this MDS is a theory framework, particularly, in a small world intranet, in the case of limit of resources, the MDS has its own limitation. Based on MDS, an improved light method for managing corporation computational resources and storage resources is proposed in intranet(IMDS). Firstly, in MDS, all kinds of resource description information is stored in LDAP, it is well known although LDAP is a light directory access protocol, in practice, programmers rarely master how to access and store resource information into LDAP store, in such way, it limits MDS to be used. So, in intranet, these resources' description information can be stored in RDBMS, programmers and users can access this information by standard SQL. Secondly, in MDS, how to monitor all kinds of resources in GRID is not transparent for programmers and users. In such way, it limits its application scope, in general, resource monitoring method base on SNMP is widely employed in intranet, therefore, a kind of resource monitoring method based on SNMP is integrated into MDS. Finally, all kinds of resources in the intranet can be described by XML, and all kinds of resources' description information is stored in RDBMS, such as MySql, and retrieved by standard SQL, dynamic information for all kinds of resources can be sent to resource storage by SNMP, A prototype resource description, monitoring is designed and implemented in intranet.
NASA Astrophysics Data System (ADS)
Vacca, G.; Pili, D.; Fiorino, D. R.; Pintus, V.
2017-05-01
The presented work is part of the research project, titled "Tecniche murarie tradizionali: conoscenza per la conservazione ed il miglioramento prestazionale" (Traditional building techniques: from knowledge to conservation and performance improvement), with the purpose of studying the building techniques of the 13th-18th centuries in the Sardinia Region (Italy) for their knowledge, conservation, and promotion. The end purpose of the entire study is to improve the performance of the examined structures. In particular, the task of the authors within the research project was to build a WebGIS to manage the data collected during the examination and study phases. This infrastructure was entirely built using Open Source software. The work consisted of designing a database built in PostgreSQL and its spatial extension PostGIS, which allows to store and manage feature geometries and spatial data. The data input is performed via a form built in HTML and PHP. The HTML part is based on Bootstrap, an open tools library for websites and web applications. The implementation of this template used both PHP and Javascript code. The PHP code manages the reading and writing of data to the database, using embedded SQL queries. As of today, we surveyed and archived more than 300 buildings, belonging to three main macro categories: fortification architectures, religious architectures, residential architectures. The masonry samples investigated in relation to the construction techniques are more than 150. The database is published on the Internet as a WebGIS built using the Leaflet Javascript open libraries, which allows creating map sites with background maps and navigation, input and query tools. This too uses an interaction of HTML, Javascript, PHP and SQL code.
Changes in Exercise Data Management
NASA Technical Reports Server (NTRS)
Buxton, R. E.; Kalogera, K. L.; Hanson, A. M.
2018-01-01
The suite of exercise hardware aboard the International Space Station (ISS) generates an immense amount of data. The data collected from the treadmill, cycle ergometer, and resistance strength training hardware are basic exercise parameters (time, heart rate, speed, load, etc.). The raw data are post processed in the laboratory and more detailed parameters are calculated from each exercise data file. Updates have recently been made to how this valuable data are stored, adding an additional level of data security, increasing data accessibility, and resulting in overall increased efficiency of medical report delivery. Questions regarding exercise performance or how exercise may influence other variables of crew health frequently arise within the crew health care community. Inquiries over the health of the exercise hardware often need quick analysis and response to ensure the exercise system is operable on a continuous basis. Consolidating all of the exercise system data in a single repository enables a quick response to both the medical and engineering communities. A SQL server database is currently in use, and provides a secure location for all of the exercise data starting at ISS Expedition 1 - current day. The database has been structured to update derived metrics automatically, making analysis and reporting available within minutes of dropping the inflight data it into the database. Commercial tools were evaluated to help aggregate and visualize data from the SQL database. The Tableau software provides manageable interface, which has improved the laboratory's output time of crew reports by 67%. Expansion of the SQL database to be inclusive of additional medical requirement metrics, addition of 'app-like' tools for mobile visualization, and collaborative use (e.g. operational support teams, research groups, and International Partners) of the data system is currently being explored.
The Influence of Injection Molding Parameter on Properties of Thermally Conductive Plastic
NASA Astrophysics Data System (ADS)
Hafizah Azis, N.; Zulafif Rahim, M.; Sa'ude, Nasuha; Rafai, N.; Yusof, M. S.; Tobi, ALM; Sharif, ZM; Rasidi Ibrahim, M.; Ismail, A. E.
2017-05-01
Thermally conductive plastic is the composite between metal-plastic material that is becoming popular because if it special characteristic. Injection moulding was regarded as the best process for mass manufacturing of the plastic composite due to its low production cost. The objective of this research is to find the best combination of the injection parameter setting and to find the most significant factor that effect the strength and thermal conductivity of the composite. Several parameter such as the volume percentage of copper powder, nozzle temperature and injection pressure of injection moulding machine were investigated. The analysis was done using Design Expert Software by implementing design of experiment method. From the analysis, the significant effects were determined and mathematical models of only significant effect were established. In order to ensure the validity of the model, confirmation run was done and percentage errors were calculated. It was found that the best combination parameter setting to maximize the value of tensile strength is volume percentage of copper powder of 3.00%, the nozzle temperature of 195°C and the injection pressure of 65%, and the best combination parameter settings to maximize the value of thermal conductivity is volume percentage of copper powder of 7.00%, the nozzle temperature of 195°C and the injection pressure of 65% as recommended..
DOE Office of Scientific and Technical Information (OSTI.GOV)
Daley, Thomas M.; Vasco, Don; Ajo-Franklin, Jonathan
After learning that the TDS value in the target injection formation at the Kevin Dome site is too low to qualify for an EPA Class VI CO2 injection permit, the BSCSP project was re-scoped such that injection of CO2 is no longer planned. With no injection planned, the Geomechanics project was closed. In this final report, we describe the objective and approach of the project as proposed, and the limited results obtained before stopping work. The objective of the proposed research was the development & validation of an integrated monitoring approach for quantifying the interactions between large-scale geological carbon storagemore » (GCS) and subsurface geomechanical state, particularly perturbations relevant to reservoir integrity such as fault reactivation and induced fracturing. In the short period of work before knowing the fate of the Kevin Dome project, we (1) researched designs for both the proposed InSAR corner reflectors as well as the near-surface 3C seismic stations; (2) developed preliminary elastic geomechanical models; (3) developed a second generation deformation prediction for the BSCSP Kevin Dome injection site; and (4) completed a preliminary map of InSAR monuments and shallow MEQ wells in the vicinity of the BSCSP injection pad.« less
Dubreil, Estelle; Gautier, Sophie; Fourmond, Marie-Pierre; Bessiral, Mélaine; Gaugain, Murielle; Verdon, Eric; Pessel, Dominique
2017-04-01
An approach is described to validate a fast and simple targeted screening method for antibiotic analysis in meat and aquaculture products by LC-MS/MS. The strategy of validation was applied for a panel of 75 antibiotics belonging to different families, i.e., penicillins, cephalosporins, sulfonamides, macrolides, quinolones and phenicols. The samples were extracted once with acetonitrile, concentrated by evaporation and injected into the LC-MS/MS system. The approach chosen for the validation was based on the Community Reference Laboratory (CRL) guidelines for the validation of screening qualitative methods. The aim of the validation was to prove sufficient sensitivity of the method to detect all the targeted antibiotics at the level of interest, generally the maximum residue limit (MRL). A robustness study was also performed to test the influence of different factors. The validation showed that the method is valid to detect and identify 73 antibiotics of the 75 antibiotics studied in meat and aquaculture products at the validation levels.
Zhang, Xiao-Yu; Wang, Yan-Ping; Lin, Li-Kai; Shang, Hong-Cai; Wang, Yong-Yan
2017-08-01
As an important representative of modern Chinese medicine, traditional Chinese medicine (TCM) injzection has become an indispensable part of the Chinese medicine industry. However, its development is now restricted by the bottleneck of insufficient core competitiveness, low-level research and production, even injection quality and the safe use are not guaranteed. Thus, it is urgent to reevaluate post-marketing TCM injection generally and to make secondary development. Under current circumstances, taking major brands which have good clinical and market foundation, as well as research value, as the main subject of cultivation and evaluation is an important approach to innovative development of TCM injection industry. Unlike oral proprietary Chinese medicine, the cultivatation of major brands of TCM injection needs higher technical support, quality standards and more timely feedback. Therefore, a post-market integral evaluation system adaptive to TCM injection is required. This article discussed some key points on the construction of a post-market integral evaluation system of TCM injection in three levels: optimizing evaluation methods, building synergistic innovation platforms which combine the medical research institutions and pharmaceutical enterprises, and finally constructing the integral evaluation system. A "five to one" structure has been proposed to enhance TCM injection effectiveness, safety and adaptability on the whole, which are from the following aspects: mechanism research, clinical evidence validation, literature information mining, sustainable development of resources and industrialization operation. Copyright© by the Chinese Pharmaceutical Association.
Martín-Díaz, M Laura; Gagné, François; Blaise, Christian
2009-09-01
A biomarker approach was undertaken using the mussel Elliptio complanata to assess the ecotoxicological effects after injection of a range concentration (0-10mM) of three different PPCPs: carbamazepine, caffeine, methotrexate; and an effluent extract (C8) from St. Lawrence wastewaters treatment plant (Montreal, Canada). A battery of biomarkers, involving oxidative stress and genotoxicity responses: glutation-S-transferase (GST), ethoxyresorufin O-deethylase (EROD), dibenzylflourescein dealkylase (DBF), xanthine oxidoreductase (XOR) activities, lipid peroxidation (LPO) and DNA damage were determined in gonad and digestive gland tissues after 48 h of injection. Results showed an induction of the oxidative metabolism with increasing pharmaceutical concentration in those mussels injected with the PPCPs and the effluent extract. Phase I detoxification enzymes were significantly induced (p<0.05), concretely DBF activity was significantly induced after caffeine, carbamazipine and C8 injection; and EROD activity after C8 and methotrexate injection. Oxidative stress induction only lead to lipid peroxidation (p<0.05) in organisms injected with carbamazepine and caffeine and DNA damage in organisms injected with methotrexate (p<0.05). EROD and DBF enzymatic activities have been found to be suitable biomarkers to determine bioavailability of pharmaceuticals. LPO and DNA damage to determine possible associated adverse effects. Nevertheless, their validation in realistic exposure scenarios and under exposure conditions should be performed in future research.
Silva-Rodríguez, Jesús; Aguiar, Pablo; Sánchez, Manuel; Mosquera, Javier; Luna-Vega, Víctor; Cortés, Julia; Garrido, Miguel; Pombar, Miguel; Ruibal, Alvaro
2014-05-01
Current procedure guidelines for whole body [18F]fluoro-2-deoxy-D-glucose (FDG)-positron emission tomography (PET) state that studies with visible dose extravasations should be rejected for quantification protocols. Our work is focused on the development and validation of methods for estimating extravasated doses in order to correct standard uptake value (SUV) values for this effect in clinical routine. One thousand three hundred sixty-seven consecutive whole body FDG-PET studies were visually inspected looking for extravasation cases. Two methods for estimating the extravasated dose were proposed and validated in different scenarios using Monte Carlo simulations. All visible extravasations were retrospectively evaluated using a manual ROI based method. In addition, the 50 patients with higher extravasated doses were also evaluated using a threshold-based method. Simulation studies showed that the proposed methods for estimating extravasated doses allow us to compensate the impact of extravasations on SUV values with an error below 5%. The quantitative evaluation of patient studies revealed that paravenous injection is a relatively frequent effect (18%) with a small fraction of patients presenting considerable extravasations ranging from 1% to a maximum of 22% of the injected dose. A criterion based on the extravasated volume and maximum concentration was established in order to identify this fraction of patients that might be corrected for paravenous injection effect. The authors propose the use of a manual ROI based method for estimating the effectively administered FDG dose and then correct SUV quantification in those patients fulfilling the proposed criterion.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Silva-Rodríguez, Jesús, E-mail: jesus.silva.rodriguez@sergas.es; Aguiar, Pablo, E-mail: pablo.aguiar.fernandez@sergas.es; Servicio de Medicina Nuclear, Complexo Hospitalario Universidade de Santiago de Compostela
Purpose: Current procedure guidelines for whole body [18F]fluoro-2-deoxy-D-glucose (FDG)-positron emission tomography (PET) state that studies with visible dose extravasations should be rejected for quantification protocols. Our work is focused on the development and validation of methods for estimating extravasated doses in order to correct standard uptake value (SUV) values for this effect in clinical routine. Methods: One thousand three hundred sixty-seven consecutive whole body FDG-PET studies were visually inspected looking for extravasation cases. Two methods for estimating the extravasated dose were proposed and validated in different scenarios using Monte Carlo simulations. All visible extravasations were retrospectively evaluated using a manualmore » ROI based method. In addition, the 50 patients with higher extravasated doses were also evaluated using a threshold-based method. Results: Simulation studies showed that the proposed methods for estimating extravasated doses allow us to compensate the impact of extravasations on SUV values with an error below 5%. The quantitative evaluation of patient studies revealed that paravenous injection is a relatively frequent effect (18%) with a small fraction of patients presenting considerable extravasations ranging from 1% to a maximum of 22% of the injected dose. A criterion based on the extravasated volume and maximum concentration was established in order to identify this fraction of patients that might be corrected for paravenous injection effect. Conclusions: The authors propose the use of a manual ROI based method for estimating the effectively administered FDG dose and then correct SUV quantification in those patients fulfilling the proposed criterion.« less
Galan, Alba; Barcelona, Pablo F; Nedev, Hinyu; Sarunic, Marinko V; Jian, Yifan; Saragovi, H Uri
2017-06-01
The p75NTR is a novel therapeutic target validated in a streptozotocin mouse model of diabetic retinopathy. Intravitreal (IVT) injection of small molecule p75NTR antagonist THX-B was therapeutic and resolved the inflammatory, vascular, and neurodegenerative phases of the retinal pathology. To simplify clinical translation, we sought a superior drug delivery method that circumvents risks associated with IVT injections. We compared the pharmacokinetics of a single 40 μg subconjunctival (SCJ) depot to the reported effective 5 μg IVT injections of THX-B. We quantified therapeutic efficacy, with endpoints of inflammation, edema, and neuronal death. The subconjunctival depot affords retinal exposure equal to IVT injection, without resulting in detectable drug in circulation. At week 2 of diabetic retinopathy, the SCJ depot provided therapeutic efficacy similar to IVT injections, with reduced inflammation, reduced edema, reduced neuronal death, and a long-lasting protection of the retinal structure. Subconjunctival injections are a safe and effective route for retinal delivery of p75NTR antagonists. The subconjunctival route offers an advantageous, less-invasive, more compliant, and nonsystemic method to deliver p75NTR antagonists for the treatment of retinal diseases.
Evaluation of a Murine Single-Blood-Injection SAH Model
Sommer, Clemens; Steiger, Hans-Jakob; Schneider, Toni; Hänggi, Daniel
2014-01-01
The molecular pathways underlying the pathogenesis after subarachnoid haemorrhage (SAH) are poorly understood and continue to be a matter of debate. A valid murine SAH injection model is not yet available but would be the prerequisite for further transgenic studies assessing the mechanisms following SAH. Using the murine single injection model, we examined the effects of SAH on regional cerebral blood flow (rCBF) in the somatosensory (S1) and cerebellar cortex, neuro-behavioural and morphological integrity and changes in quantitative electrocorticographic and electrocardiographic parameters. Micro CT imaging verified successful blood delivery into the cisterna magna. An acute impairment of rCBF was observed immediately after injection in the SAH and after 6, 12 and 24 hours in the S1 and 6 and 12 hours after SAH in the cerebellum. Injection of blood into the foramen magnum reduced telemetric recorded total ECoG power by an average of 65%. Spectral analysis of ECoGs revealed significantly increased absolute delta power, i.e., slowing, cortical depolarisations and changes in ripples and fast ripple oscillations 12 hours and 24 hours after SAH. Therefore, murine single-blood-injection SAH model is suitable for pathophysiological and further molecular analysis following SAH. PMID:25545775
Relational Algebra and SQL: Better Together
ERIC Educational Resources Information Center
McMaster, Kirby; Sambasivam, Samuel; Hadfield, Steven; Wolthuis, Stuart
2013-01-01
In this paper, we describe how database instructors can teach Relational Algebra and Structured Query Language together through programming. Students write query programs consisting of sequences of Relational Algebra operations vs. Structured Query Language SELECT statements. The query programs can then be run interactively, allowing students to…
Ontological Approach to Military Knowledge Modeling and Management
2004-03-01
federated search mechanism has to reformulate user queries (expressed using the ontology) in the query languages of the different sources (e.g. SQL...ontologies as a common terminology – Unified query to perform federated search • Query processing – Ontology mapping to sources reformulate queries
tcpl: The ToxCast Pipeline for High-Throughput Screening Data
Motivation: The large and diverse high-throughput chemical screening efforts carried out by the US EPAToxCast program requires an efficient, transparent, and reproducible data pipeline.Summary: The tcpl R package and its associated MySQL database provide a generalized platform fo...
Celesti, Antonio; Fazio, Maria; Romano, Agata; Bramanti, Alessia; Bramanti, Placido; Villari, Massimo
2018-05-01
The Open Archive Information System (OAIS) is a reference model for organizing people and resources in a system, and it is already adopted in care centers and medical systems to efficiently manage clinical data, medical personnel, and patients. Archival storage systems are typically implemented using traditional relational database systems, but the relation-oriented technology strongly limits the efficiency in the management of huge amount of patients' clinical data, especially in emerging cloud-based, that are distributed. In this paper, we present an OAIS healthcare architecture useful to manage a huge amount of HL7 clinical documents in a scalable way. Specifically, it is based on a NoSQL column-oriented Data Base Management System deployed in the cloud, thus to benefit from a big tables and wide rows available over a virtual distributed infrastructure. We developed a prototype of the proposed architecture at the IRCCS, and we evaluated its efficiency in a real case of study.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Christopher Slominski
2009-10-01
Archiving a large fraction of the EPICS signals within the Jefferson Lab (JLAB) Accelerator control system is vital for postmortem and real-time analysis of the accelerator performance. This analysis is performed on a daily basis by scientists, operators, engineers, technicians, and software developers. Archiving poses unique challenges due to the magnitude of the control system. A MySQL Archiving system (Mya) was developed to scale to the needs of the control system; currently archiving 58,000 EPICS variables, updating at a rate of 11,000 events per second. In addition to the large collection rate, retrieval of the archived data must also bemore » fast and robust. Archived data retrieval clients obtain data at a rate over 100,000 data points per second. Managing the data in a relational database provides a number of benefits. This paper describes an archiving solution that uses an open source database and standard off the shelf hardware to reach high performance archiving needs. Mya has been in production at Jefferson Lab since February of 2007.« less
Computerized database management system for breast cancer patients.
Sim, Kok Swee; Chong, Sze Siang; Tso, Chih Ping; Nia, Mohsen Esmaeili; Chong, Aun Kee; Abbas, Siti Fathimah
2014-01-01
Data analysis based on breast cancer risk factors such as age, race, breastfeeding, hormone replacement therapy, family history, and obesity was conducted on breast cancer patients using a new enhanced computerized database management system. My Structural Query Language (MySQL) is selected as the application for database management system to store the patient data collected from hospitals in Malaysia. An automatic calculation tool is embedded in this system to assist the data analysis. The results are plotted automatically and a user-friendly graphical user interface is developed that can control the MySQL database. Case studies show breast cancer incidence rate is highest among Malay women, followed by Chinese and Indian. The peak age for breast cancer incidence is from 50 to 59 years old. Results suggest that the chance of developing breast cancer is increased in older women, and reduced with breastfeeding practice. The weight status might affect the breast cancer risk differently. Additional studies are needed to confirm these findings.
Techniques of Photometry and Astrometry with APASS, Gaia, and Pan-STARRs Results (Abstract)
NASA Astrophysics Data System (ADS)
Green, W.
2017-12-01
(Abstract only) The databases with the APASS DR9, Gaia DR1, and the Pan-STARRs 3pi DR1 data releases are publicly available for use. There is a bit of data-mining involved to download and manage these reference stars. This paper discusses the use of these databases to acquire accurate photometric references as well as techniques for improving results. Images are prepared in the usual way: zero, dark, flat-fields, and WCS solutions with Astrometry.net. Images are then processed with Sextractor to produce an ASCII table of identifying photometric features. The database manages photometics catalogs and images converted to ASCII tables. Scripts convert the files into SQL and assimilate them into database tables. Using SQL techniques, each image star is merged with reference data to produce publishable results. The VYSOS has over 13,000 images of the ONC5 field to process with roughly 100 total fields in the campaign. This paper provides the overview for this daunting task.
Fill factor in organic solar cells can exceed the Shockley-Queisser limit
NASA Astrophysics Data System (ADS)
Trukhanov, Vasily A.; Bruevich, Vladimir V.; Paraschuk, Dmitry Yu.
2015-06-01
The ultimate efficiency of organic solar cells (OSC) is under active debate. The solar cell efficiency is calculated from the current-voltage characteristic as a product of the open-circuit voltage (VOC), short-circuit current (JSC), and the fill factor (FF). While the factors limiting VOC and JSC for OSC were extensively studied, the ultimate FF for OSC is scarcely explored. Using numerical drift-diffusion modeling, we have found that the FF in OSC can exceed the Shockley-Queisser limit (SQL) established for inorganic p-n junction solar cells. Comparing charge generation and recombination in organic donor-acceptor bilayer heterojunction and inorganic p-n junction, we show that such distinctive properties of OSC as interface charge generation and heterojunction facilitate high FF, but the necessary condition for FF exceeding the SQL in OSC is field-dependence of charge recombination at the donor-acceptor interface. These findings can serve as a guideline for further improvement of OSC.
A JEE RESTful service to access Conditions Data in ATLAS
NASA Astrophysics Data System (ADS)
Formica, Andrea; Gallas, E. J.
2015-12-01
Usage of condition data in ATLAS is extensive for offline reconstruction and analysis (e.g. alignment, calibration, data quality). The system is based on the LCG Conditions Database infrastructure, with read and write access via an ad hoc C++ API (COOL), a system which was developed before Run 1 data taking began. The infrastructure dictates that the data is organized into separate schemas (assigned to subsystems/groups storing distinct and independent sets of conditions), making it difficult to access information from several schemas at the same time. We have thus created PL/SQL functions containing queries to provide content extraction at multi-schema level. The PL/SQL API has been exposed to external clients by means of a Java application providing DB access via REST services, deployed inside an application server (JBoss WildFly). The services allow navigation over multiple schemas via simple URLs. The data can be retrieved either in XML or JSON formats, via simple clients (like curl or Web browsers).
Constructing a Graph Database for Semantic Literature-Based Discovery.
Hristovski, Dimitar; Kastrin, Andrej; Dinevski, Dejan; Rindflesch, Thomas C
2015-01-01
Literature-based discovery (LBD) generates discoveries, or hypotheses, by combining what is already known in the literature. Potential discoveries have the form of relations between biomedical concepts; for example, a drug may be determined to treat a disease other than the one for which it was intended. LBD views the knowledge in a domain as a network; a set of concepts along with the relations between them. As a starting point, we used SemMedDB, a database of semantic relations between biomedical concepts extracted with SemRep from Medline. SemMedDB is distributed as a MySQL relational database, which has some problems when dealing with network data. We transformed and uploaded SemMedDB into the Neo4j graph database, and implemented the basic LBD discovery algorithms with the Cypher query language. We conclude that storing the data needed for semantic LBD is more natural in a graph database. Also, implementing LBD discovery algorithms is conceptually simpler with a graph query language when compared with standard SQL.
NoSQL Based 3D City Model Management System
NASA Astrophysics Data System (ADS)
Mao, B.; Harrie, L.; Cao, J.; Wu, Z.; Shen, J.
2014-04-01
To manage increasingly complicated 3D city models, a framework based on NoSQL database is proposed in this paper. The framework supports import and export of 3D city model according to international standards such as CityGML, KML/COLLADA and X3D. We also suggest and implement 3D model analysis and visualization in the framework. For city model analysis, 3D geometry data and semantic information (such as name, height, area, price and so on) are stored and processed separately. We use a Map-Reduce method to deal with the 3D geometry data since it is more complex, while the semantic analysis is mainly based on database query operation. For visualization, a multiple 3D city representation structure CityTree is implemented within the framework to support dynamic LODs based on user viewpoint. Also, the proposed framework is easily extensible and supports geoindexes to speed up the querying. Our experimental results show that the proposed 3D city management system can efficiently fulfil the analysis and visualization requirements.
Incremental Query Rewriting with Resolution
NASA Astrophysics Data System (ADS)
Riazanov, Alexandre; Aragão, Marcelo A. T.
We address the problem of semantic querying of relational databases (RDB) modulo knowledge bases using very expressive knowledge representation formalisms, such as full first-order logic or its various fragments. We propose to use a resolution-based first-order logic (FOL) reasoner for computing schematic answers to deductive queries, with the subsequent translation of these schematic answers to SQL queries which are evaluated using a conventional relational DBMS. We call our method incremental query rewriting, because an original semantic query is rewritten into a (potentially infinite) series of SQL queries. In this chapter, we outline the main idea of our technique - using abstractions of databases and constrained clauses for deriving schematic answers, and provide completeness and soundness proofs to justify the applicability of this technique to the case of resolution for FOL without equality. The proposed method can be directly used with regular RDBs, including legacy databases. Moreover, we propose it as a potential basis for an efficient Web-scale semantic search technology.
NASA Astrophysics Data System (ADS)
Barnsley, R. M.; Steele, Iain A.; Smith, R. J.; Mawson, Neil R.
2014-07-01
The Small Telescopes Installed at the Liverpool Telescope (STILT) project has been in operation since March 2009, collecting data with three wide field unfiltered cameras: SkycamA, SkycamT and SkycamZ. To process the data, a pipeline was developed to automate source extraction, catalogue cross-matching, photometric calibration and database storage. In this paper, modifications and further developments to this pipeline will be discussed, including a complete refactor of the pipeline's codebase into Python, migration of the back-end database technology from MySQL to PostgreSQL, and changing the catalogue used for source cross-matching from USNO-B1 to APASS. In addition to this, details will be given relating to the development of a preliminary front-end to the source extracted database which will allow a user to perform common queries such as cone searches and light curve comparisons of catalogue and non-catalogue matched objects. Some next steps and future ideas for the project will also be presented.
StreptomycesInforSys: A web-enabled information repository
Jain, Chakresh Kumar; Gupta, Vidhi; Gupta, Ashvarya; Gupta, Sanjay; Wadhwa, Gulshan; Sharma, Sanjeev Kumar; Sarethy, Indira P
2012-01-01
Members of Streptomyces produce 70% of natural bioactive products. There is considerable amount of information available based on polyphasic approach for classification of Streptomyces. However, this information based on phenotypic, genotypic and bioactive component production profiles is crucial for pharmacological screening programmes. This is scattered across various journals, books and other resources, many of which are not freely accessible. The designed database incorporates polyphasic typing information using combinations of search options to aid in efficient screening of new isolates. This will help in the preliminary categorization of appropriate groups. It is a free relational database compatible with existing operating systems. A cross platform technology with XAMPP Web server has been used to develop, manage, and facilitate the user query effectively with database support. Employment of PHP, a platform-independent scripting language, embedded in HTML and the database management software MySQL will facilitate dynamic information storage and retrieval. The user-friendly, open and flexible freeware (PHP, MySQL and Apache) is foreseen to reduce running and maintenance cost. Availability www.sis.biowaves.org PMID:23275736
A Web-based Tool for SDSS and 2MASS Database Searches
NASA Astrophysics Data System (ADS)
Hendrickson, M. A.; Uomoto, A.; Golimowski, D. A.
We have developed a web site using HTML, Php, Python, and MySQL that extracts, processes, and displays data from the Sloan Digital Sky Survey (SDSS) and the Two-Micron All-Sky Survey (2MASS). The goal is to locate brown dwarf candidates in the SDSS database by looking at color cuts; however, this site could also be useful for targeted searches of other databases as well. MySQL databases are created from broad searches of SDSS and 2MASS data. Broad queries on the SDSS and 2MASS database servers are run weekly so that observers have the most up-to-date information from which to select candidates for observation. Observers can look at detailed information about specific objects including finding charts, images, and available spectra. In addition, updates from previous observations can be added by any collaborators; this format makes observational collaboration simple. Observers can also restrict the database search, just before or during an observing run, to select objects of special interest.
Image query and indexing for digital x rays
NASA Astrophysics Data System (ADS)
Long, L. Rodney; Thoma, George R.
1998-12-01
The web-based medical information retrieval system (WebMIRS) allows interned access to databases containing 17,000 digitized x-ray spine images and associated text data from National Health and Nutrition Examination Surveys (NHANES). WebMIRS allows SQL query of the text, and viewing of the returned text records and images using a standard browser. We are now working (1) to determine utility of data directly derived from the images in our databases, and (2) to investigate the feasibility of computer-assisted or automated indexing of the images to support image retrieval of images of interest to biomedical researchers in the field of osteoarthritis. To build an initial database based on image data, we are manually segmenting a subset of the vertebrae, using techniques from vertebral morphometry. From this, we will derive and add to the database vertebral features. This image-derived data will enhance the user's data access capability by enabling the creation of combined SQL/image-content queries.
StreptomycesInforSys: A web-enabled information repository.
Jain, Chakresh Kumar; Gupta, Vidhi; Gupta, Ashvarya; Gupta, Sanjay; Wadhwa, Gulshan; Sharma, Sanjeev Kumar; Sarethy, Indira P
2012-01-01
Members of Streptomyces produce 70% of natural bioactive products. There is considerable amount of information available based on polyphasic approach for classification of Streptomyces. However, this information based on phenotypic, genotypic and bioactive component production profiles is crucial for pharmacological screening programmes. This is scattered across various journals, books and other resources, many of which are not freely accessible. The designed database incorporates polyphasic typing information using combinations of search options to aid in efficient screening of new isolates. This will help in the preliminary categorization of appropriate groups. It is a free relational database compatible with existing operating systems. A cross platform technology with XAMPP Web server has been used to develop, manage, and facilitate the user query effectively with database support. Employment of PHP, a platform-independent scripting language, embedded in HTML and the database management software MySQL will facilitate dynamic information storage and retrieval. The user-friendly, open and flexible freeware (PHP, MySQL and Apache) is foreseen to reduce running and maintenance cost. www.sis.biowaves.org.
An integrated data-analysis and database system for AMS 14C
NASA Astrophysics Data System (ADS)
Kjeldsen, Henrik; Olsen, Jesper; Heinemeier, Jan
2010-04-01
AMSdata is the name of a combined database and data-analysis system for AMS 14C and stable-isotope work that has been developed at Aarhus University. The system (1) contains routines for data analysis of AMS and MS data, (2) allows a flexible and accurate description of sample extraction and pretreatment, also when samples are split into several fractions, and (3) keeps track of all measured, calculated and attributed data. The structure of the database is flexible and allows an unlimited number of measurement and pretreatment procedures. The AMS 14C data analysis routine is fairly advanced and flexible, and it can be easily optimized for different kinds of measuring processes. Technically, the system is based on a Microsoft SQL server and includes stored SQL procedures for the data analysis. Microsoft Office Access is used for the (graphical) user interface, and in addition Excel, Word and Origin are exploited for input and output of data, e.g. for plotting data during data analysis.
Aromatherapy: composition of the gaseous phase at equilibrium with liquid bergamot essential oil.
Leggio, Antonella; Leotta, Vanessa; Belsito, Emilia Lucia; Di Gioia, Maria Luisa; Romio, Emanuela; Santoro, Ilaria; Taverna, Domenico; Sindona, Giovanni; Liguori, Angelo
2017-11-02
This work compares the composition at different temperatures of gaseous phase of bergamot essential oil at equilibrium with the liquid phase. A new GC-MS methodology to determine quantitatively the volatile aroma compounds was developed. The adopted methodology involved the direct injection of headspace gas into injection port of GC-MS system and of known amounts of the corresponding authentic volatile compounds. The methodology was validated. This study showed that gaseous phase composition is different from that of the liquid phase at equilibrium with it.
Clark, Toshimasa J; Wilson, Gregory J; Maki, Jeffrey H
2017-07-01
Contrast-enhanced (CE)-MRA optimization involves interactions of sequence duration, bolus timing, contrast recirculation, and both R 1 relaxivity and R2*-related reduction of signal. Prior data suggest superior image quality with slower gadolinium injection rates than typically used. A computer-based model of CE-MRA was developed, with contrast injection, physiologic, and image acquisition parameters varied over a wide gamut. Gadolinium concentration was derived using Verhoeven's model with recirculation, R 1 and R2* calculated at each time point, and modulation transfer curves used to determine injection rates, resulting in optimal resolution and image contrast for renal and carotid artery CE-MRA. Validation was via a vessel stenosis phantom and example patients who underwent carotid CE-MRA with low effective injection rates. Optimal resolution for renal and carotid CE-MRA is achieved with injection rates between 0.5 to 0.9 mL/s and 0.2 to 0.3 mL/s, respectively, dependent on contrast volume. Optimal image contrast requires slightly faster injection rates. Expected signal-to-noise ratio varies with both contrast volume and cardiac output. Simulated vessel phantom and clinical carotid CE-MRA exams at an effective contrast injection rate of 0.4 to 0.5 mL/s demonstrate increased resolution. Optimal image resolution is achieved at intuitively low, effective injection rates (0.2-0.9 mL/s, dependent on imaging parameters and contrast injection volume). Magn Reson Med 78:357-369, 2017. © 2016 International Society for Magnetic Resonance in Medicine. © 2016 International Society for Magnetic Resonance in Medicine.
NASA Technical Reports Server (NTRS)
Kalinowski, Kevin F.; Tucker, George E.; Moralez, Ernesto, III
2006-01-01
Engineering development and qualification of a Research Flight Control System (RFCS) for the Rotorcraft Aircrew Systems Concepts Airborne Laboratory (RASCAL) JUH-60A has motivated the development of a pilot rating scale for evaluating failure transients in fly-by-wire flight control systems. The RASCAL RFCS includes a highly-reliable, dual-channel Servo Control Unit (SCU) to command and monitor the performance of the fly-by-wire actuators and protect against the effects of erroneous commands from the flexible, but single-thread Flight Control Computer. During the design phase of the RFCS, two piloted simulations were conducted on the Ames Research Center Vertical Motion Simulator (VMS) to help define the required performance characteristics of the safety monitoring algorithms in the SCU. Simulated failures, including hard-over and slow-over commands, were injected into the command path, and the aircraft response and safety monitor performance were evaluated. A subjective Failure/Recovery Rating (F/RR) scale was developed as a means of quantifying the effects of the injected failures on the aircraft state and the degree of pilot effort required to safely recover the aircraft. A brief evaluation of the rating scale was also conducted on the Army/NASA CH-47B variable stability helicopter to confirm that the rating scale was likely to be equally applicable to in-flight evaluations. Following the initial research flight qualification of the RFCS in 2002, a flight test effort was begun to validate the performance of the safety monitors and to validate their design for the safe conduct of research flight testing. Simulated failures were injected into the SCU, and the F/RR scale was applied to assess the results. The results validate the performance of the monitors, and indicate that the Failure/Recovery Rating scale is a very useful tool for evaluating failure transients in fly-by-wire flight control systems.
Development and validation of a periarticular injection technique of the sacroiliac joint in horses.
Engeli, E; Haussler, K K; Erb, H N
2004-05-01
Sacroiliac joint osteoarthritis has been recognised as a significant cause of poor performance in competition and racehorses. Reliable diagnostic tools are currently lacking. The diagnosis has been based typically on exclusion of other possible causes of poor performance, back pain and hindlimb lameness. To develop a safe, reliable and minimally invasive periarticular or intra-articular injection technique of potential use for diagnosis and therapy of sacroiliac joint disease in horses. Twenty-six horses were used to develop and assess a medial approach to the sacroiliac joint with a 15 gauge, 25 cm long spinal needle. In Part I, the cadaveric study, the spinal needle was introduced cranial to the contralateral tuber sacrale and advanced along the medial aspect of the ipsilateral iliac wing until the dorsal surface of the sacrum was encountered. One ml methylene blue (MB) was injected in both sacroiliac joint regions of the sacropelvic specimens. The location of MB-stained tissues relative to the sacroiliac joints was recorded after dissection and disarticulation of the sacroiliac joint. In Part II, the in vivo study, 18 horses were used to validate the in vivo application of the sacroiliac joint injection technique. Horses were restrained in stocks and sedated in preparation for needle placement. One ml MB was injected bilaterally prior to euthanasia. Stained tissues were identified and recorded at necropsy. Successful joint injections were characterised as having MB located intra-articularly or < or = 2 cm periarticularly from the sacroiliac joint margin and localised to the middle or caudal third of the sacroiliac joint. Intra-articular MB was not observed in any specimen. However, MB-stained tissue was identified periarticularly in all injection sites (n = 48). Based on the predetermined success criteria, 96% of the methylene blue depots were located at the middle or caudal third of the sacroiliac joint. Dye-stained tissue was located < or = 2 cm from the sacroiliac joint margins in 88% of the specimens. Median distance of the MB from the sacroiliac joint margins was 1.0 cm (range 0.2-3.8 cm). The overall success rate considering both location and distance of the MB-stained tissue relative to sacroiliac joint margins was 83% (40 of 48 joints). The injection technique provides a reliable, easy to perform and consistent access to the medial periarticular aspect of the sacroiliac joint. The described injection technique has the potential for both diagnostic and therapeutic applications in the medical management of equine sacroiliac joint disease. Further investigation is necessary to evaluate clinical efficacy and potential adverse effects.
DAMAGE MODELING OF INJECTION-MOLDED SHORT- AND LONG-FIBER THERMOPLASTICS
DOE Office of Scientific and Technical Information (OSTI.GOV)
Nguyen, Ba Nghiep; Kunc, Vlastimil; Bapanapalli, Satish K.
2009-10-30
This article applies the recent anisotropic rotary diffusion – reduced strain closure (ARD-RSC) model for predicting fiber orientation and a new damage model for injection-molded long-fiber thermoplastics (LFTs) to analyze progressive damage leading to total failure of injection-molded long-glass-fiber/polypropylene (PP) specimens. The ARD-RSC model was implemented in a research version of the Autodesk Moldflow Plastics Insight (MPI) processing code, and it has been used to simulate injection-molding of a long-glass-fiber/PP plaque. The damage model combines micromechanical modeling with a continuum damage mechanics description to predict the nonlinear behavior due to plasticity coupled with damage in LFTs. This model has beenmore » implemented in the ABAQUS finite element code via user-subroutines and has been used in the damage analyses of tensile specimens removed from the injection-molded long-glass-fiber/PP plaques. Experimental characterization and mechanical testing were performed to provide input data to support and validate both process modeling and damage analyses. The predictions are in agreement with the experimental results.« less
Scaling of seismicity induced by nonlinear fluid-rock interaction after an injection stop
NASA Astrophysics Data System (ADS)
Johann, L.; Dinske, C.; Shapiro, S. A.
2016-11-01
Fluid injections into unconventional reservoirs, performed for fluid-mobility enhancement, are accompanied by microseismic activity also after the injection. Previous studies revealed that the triggering of seismic events can be effectively described by nonlinear diffusion of pore fluid pressure perturbations where the hydraulic diffusivity becomes pressure dependent. The spatiotemporal distribution of postinjection-induced microseismicity has two important features: the triggering front, corresponding to early and distant events, and the back front, representing the time-dependent spatial envelope of the growing seismic quiescence zone. Here for the first time, we describe analytically the temporal behavior of these two fronts after the injection stop in the case of nonlinear pore fluid pressure diffusion. We propose a scaling law for the fronts and show that they are sensitive to the degree of nonlinearity and to the Euclidean dimension of the dominant growth of seismicity clouds. To validate the theoretical finding, we numerically model nonlinear pore fluid pressure diffusion and generate synthetic catalogs of seismicity. Additionally, we apply the new scaling relation to several case studies of injection-induced seismicity. The derived scaling laws describe well synthetic and real data.
Junsomboon, Jaroon; Jakmunee, Jaroon
2008-07-15
A simple flow injection system using three 3-way solenoid valves as an electric control injection valve and with a simple home-made chloride ion selective electrode based on Ag/AgCl wire as a sensor for determination of water soluble chloride in admixtures and aggregates for cement has been developed. A liquid sample or an extract was injected into a water carrier stream which was then merged with 0.1M KNO(3) stream and flowed through a flow cell where the solution will be in contact with the sensor, producing a potential change recorded as a peak. A calibration graph in range of 10-100 mg L(-1) was obtained with a detection limit of 2 mg L(-1). Relative standard deviations for 7 replicates injecting of 20, 60 and 90 mg L(-1) chloride solutions were 1.0, 1.2 and 0.6%, respectively. Sample throughput of 60 h(-1) was achieved with the consumption of 1 mL each of electrolyte solution and water carrier. The developed method was validated by the British Standard methods.
Simultaneous determination of ezetimibe and simvastatin in pharmaceutical preparations by MEKC.
Yardimci, Ceren; Ozaltin, Nuran
2010-02-01
A micellar electrokinetic capillary chromatography method was developed and validated for the simultaneous determination of ezetimibe and simvastatin in pharmaceutical preparations. The influence of buffer concentration, buffer pH, sodium dodecyl sulphate (SDS) concentration, organic modifier, capillary temperature, applied voltage, and injection time was investigated, and the method validation studies were performed. The optimum separation for these analytes was achieved in less than 10 min at 30 degrees C with a fused-silica capillary column (56 cm x 50 microm i.d.) and a 25mM borate buffer at pH 9.0 containing 25mM SDS and 10% (v/v) acetonitrile. The samples were injected hydrodynamically for 3 s at 50 mbar, and the applied voltage was +30.0 kV. Detection wavelength was set at 238 nm. Diflunisal was used as internal standard. The method was suitably validated with respect to stability, specificity, linearity, limits of detection and quantification, accuracy, precision, and robustness. The limits of detection and quantification were 1.0 and 2.0 microg/mL for both ezetimibe and simvastatin, respectively. The method developed was successfully applied to the simultaneous determination of ezetimibe and simvastatin in pharmaceutical preparations.
Large-Eddy Simulation of Subsonic Jets
NASA Astrophysics Data System (ADS)
Vuorinen, Ville; Wehrfritz, Armin; Yu, Jingzhou; Kaario, Ossi; Larmi, Martti; Boersma, Bendiks Jan
2011-12-01
The present study deals with development and validation of a fully explicit, compressible Runge-Kutta-4 (RK4) Navier-Stokes solver in the opensource CFD programming environment OpenFOAM. The background motivation is to shift towards explicit density based solution strategy and thereby avoid using the pressure based algorithms which are currently proposed in the standard OpenFOAM release for Large-Eddy Simulation (LES). This shift is considered necessary in strongly compressible flows when Ma > 0.5. Our application of interest is related to the pre-mixing stage in direct injection gas engines where high injection pressures are typically utilized. First, the developed flow solver is discussed and validated. Then, the implementation of subsonic inflow conditions using a forcing region in combination with a simplified nozzle geometry is discussed and validated. After this, LES of mixing in compressible, round jets at Ma = 0.3, 0.5 and 0.65 are carried out. Respectively, the Reynolds numbers of the jets correspond to Re = 6000, 10000 and 13000. Results for two meshes are presented. The results imply that the present solver produces turbulent structures, resolves a range of turbulent eddy frequencies and gives also mesh independent results within satisfactory limits for mean flow and turbulence statistics.
Herbrink, M; de Vries, N; Rosing, H; Huitema, A D R; Nuijen, B; Schellens, J H M; Beijnen, J H
2018-04-01
To support therapeutic drug monitoring of patients with cancer, a fast and accurate method for simultaneous quantification of the registered anticancer drugs afatinib, axitinib, ceritinib, crizotinib, dabrafenib, enzalutamide, regorafenib and trametinib in human plasma using liquid chromatography tandem mass spectrometry was developed and validated. Human plasma samples were collected from treated patients and stored at -20°C. Analytes and internal standards (stable isotopically labeled analytes) were extracted with acetonitrile. An equal amount of 10 mm NH 4 CO 3 was added to the supernatant to yield the final extract. A 2 μL aliquot of this extract was injected onto a C 18 -column, gradient elution was applied and triple-quadrupole mass spectrometry in positive-ion mode was used for detection. All results were within the acceptance criteria of the latest US Food and Drug Administration guidance and European Medicines Agency guidelines on method validation, except for the carry-over of ceritinib and crizotinib. These were corrected for by the injection order of samples. Additional stability tests were carried out for axitinib and dabrafenib in relation to their reported photostability. In conclusion, the described method to simultaneously quantify the eight selected anticancer drugs in human plasma was successfully validated and applied for therapeutic drug monitoring in cancer patients treated with these drugs. Copyright © 2017 John Wiley & Sons, Ltd.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Grigg, Reid; McPherson, Brian; Lee, Rober
The Southwest Regional Partnership on Carbon Sequestration (SWP) one of seven regional partnerships sponsored by the U.S. Department of Energy (USDOE) carried out five field pilot tests in its Phase II Carbon Sequestration Demonstration effort, to validate the most promising sequestration technologies and infrastructure concepts, including three geologic pilot tests and two terrestrial pilot programs. This field testing demonstrated the efficacy of proposed sequestration technologies to reduce or offset greenhouse gas emissions in the region. Risk mitigation, optimization of monitoring, verification, and accounting (MVA) protocols, and effective outreach and communication were additional critical goals of these field validation tests. Themore » program included geologic pilot tests located in Utah, New Mexico, Texas, and a region-wide terrestrial analysis. Each geologic sequestration test site was intended to include injection of a minimum of ~75,000 tons/year CO{sub 2}, with minimum injection duration of one year. These pilots represent medium- scale validation tests in sinks that host capacity for possible larger-scale sequestration operations in the future. These validation tests also demonstrated a broad variety of carbon sink targets and multiple value-added benefits, including testing of enhanced oil recovery and sequestration, enhanced coalbed methane production and a geologic sequestration test combined with a local terrestrial sequestration pilot. A regional terrestrial sequestration demonstration was also carried out, with a focus on improved terrestrial MVA methods and reporting approaches specific for the Southwest region.« less
ERIC Educational Resources Information Center
Bosc, P.; Lietard, L.; Pivert, O.
2003-01-01
Considers flexible querying of relational databases. Highlights include SQL languages and basic aggregate operators; Sugeno's fuzzy integral; evaluation examples; and how and under what conditions other aggregate functions could be applied to fuzzy sets in a flexible query. (Author/LRW)
Unobtrusive Social Network Data From Email
2008-12-01
PERSON Standard Form 298 (Rev. 8-98) Prescribed by ANSI Std Z39-18 outlook archived files and stores that data into an SQL - database. Communication...Applications ( VBA ) program was installed on the personal computers (PC) of all participants, in the session window of their Microsoft Outlook. Details of
Dr. John H. Hopps Jr. Research Scholars Program
2014-10-20
Program staff, alumni and existing participants. Over the course of the last five months, SageFox has successfully obtained IRB approval for all...and awards. Progress made in development of the HoppsNet system included design and implementation of a relational database in MySQL , development of
Acoustic Metadata Management and Transparent Access to Networked Oceanographic Data Sets
2013-09-30
connectivity (ODBC) compliant data source for which drivers are available (e.g. MySQL , Oracle database, Postgres) can now be imported. Implementation...the possibility of speeding data transmission through compression (implemented) or the potential to use alternative data formats such as Java script
Social Impacts Module (SIM) Transition
2012-09-28
User String The authorized user’s name to access the PAVE database. Applies only to Microsoft SQL Server; leave blank, otherwise. passwd String The...otherwise. passwd String The password if an authorized user’s name is required; otherwise, leave blank driver String The class name for the driver to
Microsoft Repository Version 2 and the Open Information Model.
ERIC Educational Resources Information Center
Bernstein, Philip A.; Bergstraesser, Thomas; Carlson, Jason; Pal, Shankar; Sanders, Paul; Shutt, David
1999-01-01
Describes the programming interface and implementation of the repository engine and the Open Information Model for Microsoft Repository, an object-oriented meta-data management facility that ships in Microsoft Visual Studio and Microsoft SQL Server. Discusses Microsoft's component object model, object manipulation, queries, and information…
NASA Astrophysics Data System (ADS)
Luk, Alex T.; Lin, Yuting; Grimmond, Brian; Sood, Anup; Uzgiris, Egidijus E.; Nalcioglu, Orhan; Gulsen, Gultekin
2013-03-01
Since diffuse optical tomography (DOT) is a low spatial resolution modality, it is desirable to validate its quantitative accuracy with another well-established imaging modality, such as magnetic resonance imaging (MRI). In this work, we have used a polymer based bi-functional MRI-optical contrast agent (Gd-DTPA-polylysine-IR800) in collaboration with GE Global Research. This multi-modality contrast agent provided not only co-localization but also the same kinetics, to cross-validate two imaging modalities. Bi-functional agents are injected to the rats and pharmacokinetics at the bladder are recovered using both optical and MR imaging. DOT results are validated using MRI results as "gold standard"
Kornowski, R; Fuchs, S; Tio, F O; Pierre, A; Epstein, S E; Leon, M B
1999-12-01
Direct myocardial injection of therapeutic agents has been explored as a new method for myocardial revascularization. The integration of a 3D electromechanical mapping catheter with a retractable injection needle should allow for intramyocardial injection to identified sites, obviating the need for open heart surgery. This study assessed the procedural safety and performance characteristics of a novel guided catheter-based transendocardial injection system. The electromagnetic guidance system was coupled with a retrievable 27G needle for left ventricular endocardial injection. Using this system, we injected, transendocardially, methylene-blue (MB) dye tracer at a volume of 0.1 or 0.2 ml per injection in eight normal pigs. Animals were sacrificed acutely, at 1, 3, and 7 days (two animal in each time). Three animals served as controls. The injections were followed by coronary angiography and echocardiogram to assess possible ventricular or coronary perforation and wall motion abnormalities. CK-MB levels were measured up to 24 hr following the procedure. The animals were sacrificed at the assigned time for gross and histopathology evaluation. A total of 101 injections were made in all regions of the heart except the apex and the mitral valve. No animal died as a result of the mapping or injection procedures. Vital signs did not change relative to baseline after the mapping and injection procedures. CK-MB values did not increase over time and there was no evidence of sustained arrhythmia or hemodynamic compromise. There was no evidence of left ventricular or coronary perforation, global or regional wall motion abnormalities, or hemopericardium. On histologic evaluation, the estimated volume of tissue staining was greater than the volume of the injected MB dye due to dispersion of the injectate in the interstitial and intracellular fluid compartments. It is concluded that using this magnetic guidance catheter-based navigational system, it is feasible and safe to perform the transendocardial injection procedure. Thus, if it is determined that direct intramyocardial injection of drugs is a valid therapeutic strategy, this approach offers a clear advantage over surgically based transepicardial injection procedures. Cathet. Cardiovasc. Intervent. 48:447-453, 1999. Copyright 1999 Wiley-Liss, Inc.
NASA Technical Reports Server (NTRS)
Mcdaniel, James C.; Fletcher, Douglas G.; Hartfield, Roy J.; Hollo, Steven D.
1991-01-01
A spatially-complete data set of the important primitive flow variables is presented for the complex, nonreacting, 3D unit combustor flow field employing transverse injection into a Mach 2 flow behind a rearward-facing step. A unique wind tunnel facility providing the capability for iodine seeding was built specifically for these measurements. Two optical techniques based on laser-induced-iodine fluorescence were developed and utilized for nonintrusive, in situ flow field measurements. LDA provided both mean and fluctuating velocity component measurements. A thermographic phosphor wall temperature measurement technique was developed and employed. Data from the 2D flow over a rearward-facing step and the complex 3D mixing flow with injection are reported.
NASA Technical Reports Server (NTRS)
Hartfield, Roy J.; Hollo, Steven D.; Mcdaniel, James C.
1990-01-01
Planar measurements of injectant mole fraction and temperature have been conducted in a nonreacting supersonic combustor configured with underexpanded injection in the base of a swept ramp. The temperature measurements were conducted with a Mach 2 test section inlet in streamwise planes perpendicular to the test section wall on which the ramp was mounted. Injection concentration measurements, conducted in cross flow planes with both Mach 2 and Mach 2.9 free stream conditions, dramatically illustrate the domination of the mixing process by streamwise vorticity generated by the ramp. These measurements, conducted using a nonintrusive optical technique (laser-induced iodine fluorescence), provide an accurate and extensive experimental data base for the validation of computation fluid dynamic codes for the calculation of highly three-dimensional supersonic combustor flow fields.
Sentinel Lymph Node (SLN) laparoscopic assessment early stage in endometrial cancer.
Gargiulo, T; Giusti, M; Bottero, A; Leo, L; Brokaj, L; Armellino, F; Palladin, L
2003-06-01
The aim of the study was to demonstrate the validity of sentinel lymph node (SLN) detection after injection of radioactive isotope and patent blue dye in patients affected by early stage endometrial cancer. The second purpose was to compare radioactive isotope and patent blue dye migration. Between September 2000 and May 2001, 11 patients with endometrial cancer FIGO stage Ib (n=10) and IIa (n=1) underwent laparoscopic SLN detection during laparoscopic assisted vaginal hysterectomy with bilateral salpingo-oophorectomy and pelvic bilateral systematic lymphadenectomy. Radioactive isotope injection was performed 24 ours before surgery and blue dye injection was performed just before surgery in the cervix at 3, 6, 9 and 12 hours. A 350 mm laparoscopic gamma-scintiprobe MR 100 type 11, (99m)Tc setted (Pol.Hi.Tech.), was used intraoperatively for detecting SLN. Seventeen SLN were detected at lymphoscintigraphy (6 bilateral and 5 monolateral). At laparoscopic surgery the same locations were found belonging at internal iliac lymph nodes (the so called "Leveuf-Godard" area, lateral to the inferior vescical artery, ventral to the origin of uterine artery and medial or caudal to the external iliac vein). Fourteen SLN were negative at histological analysis and only 3 positive for micrometastasis (mean SLN sections = 60. All the other pelvic lymph nodes were negative at histological analysis. The same SLN locations detected with g-scintiprobe were observed during laparoscopy after patent blue dye injection. If the sensitivity of the assessment of SLN is confirmed to be 100%, this laparoscopic approach could change the management of early stage endometrial cancer. The clinical validity of this technique must be evaluated prospectively.
Aydogan, Dogu Baran; Jacobs, Russell; Dulawa, Stephanie; Thompson, Summer L; Francois, Maite Christi; Toga, Arthur W; Dong, Hongwei; Knowles, James A; Shi, Yonggang
2018-04-16
Tractography is a powerful technique capable of non-invasively reconstructing the structural connections in the brain using diffusion MRI images, but the validation of tractograms is challenging due to lack of ground truth. Owing to recent developments in mapping the mouse brain connectome, high-resolution tracer injection-based axonal projection maps have been created and quickly adopted for the validation of tractography. Previous studies using tracer injections mainly focused on investigating the match in projections and optimal tractography protocols. Being a complicated technique, however, tractography relies on multiple stages of operations and parameters. These factors introduce large variabilities in tractograms, hindering the optimization of protocols and making the interpretation of results difficult. Based on this observation, in contrast to previous studies, in this work we focused on quantifying and ranking the amount of performance variation introduced by these factors. For this purpose, we performed over a million tractography experiments and studied the variability across different subjects, injections, anatomical constraints and tractography parameters. By using N-way ANOVA analysis, we show that all tractography parameters are significant and importantly performance variations with respect to the differences in subjects are comparable to the variations due to tractography parameters, which strongly underlines the importance of fully documenting the tractography protocols in scientific experiments. We also quantitatively show that inclusion of anatomical constraints is the most significant factor for improving tractography performance. Although this critical factor helps reduce false positives, our analysis indicates that anatomy-informed tractography still fails to capture a large portion of axonal projections.
Performance Measurements and Mapping of a R-407C Vapor Injection Scroll Compressor
NASA Astrophysics Data System (ADS)
Lumpkin, Domenique; Spielbauer, Niklas; Groll, Eckhard
2017-08-01
Environmental conditions significantly define the performance of HVAC&R systems. Vapor compression systems in hot climates tend to operate at higher pressure ratios, leading to increased discharge temperatures. Higher discharge temperatures can lead to higher irreversibilities in the compression process, lower specific enthalpies differences across the evaporator, and possibly a reduction in the compressor life due to the breakdown of the oil used for lubrication. To counter these effects, the use of economized, vapor injection compressors is proposed for vapor compression systems in high temperature climates. Such compressors are commercially available for refrigeration applications, in particular, supermarket refrigeration systems. However, compressor maps for vapor injection compressors are limited and none exist for R-407C. Through calorimeter testing, a compressor map for a single-port vapor injection compressor using R-407C was developed. A standard correlation for mapping single-port vapor injection compressors is proposed and validated using the compressor test results. The system and compressor performance with and without vapor injection was considered. As expected, with vapor injection there was a reduction in compressor discharge temperatures and an increase in the system coefficient of performance. The proposed dimensionless correlation is more accurate than the AHRI polynomial for mapping the injection ratio, discharge temperature, and compressor heat loss. The predicted volumetric efficiency values from the dimensionless correlation is within 1% of the measured valued. Similarly, the predicted isentropic efficiency values are within 2% of the measured values.
PhyloExplorer: a web server to validate, explore and query phylogenetic trees
Ranwez, Vincent; Clairon, Nicolas; Delsuc, Frédéric; Pourali, Saeed; Auberval, Nicolas; Diser, Sorel; Berry, Vincent
2009-01-01
Background Many important problems in evolutionary biology require molecular phylogenies to be reconstructed. Phylogenetic trees must then be manipulated for subsequent inclusion in publications or analyses such as supertree inference and tree comparisons. However, no tool is currently available to facilitate the management of tree collections providing, for instance: standardisation of taxon names among trees with respect to a reference taxonomy; selection of relevant subsets of trees or sub-trees according to a taxonomic query; or simply computation of descriptive statistics on the collection. Moreover, although several databases of phylogenetic trees exist, there is currently no easy way to find trees that are both relevant and complementary to a given collection of trees. Results We propose a tool to facilitate assessment and management of phylogenetic tree collections. Given an input collection of rooted trees, PhyloExplorer provides facilities for obtaining statistics describing the collection, correcting invalid taxon names, extracting taxonomically relevant parts of the collection using a dedicated query language, and identifying related trees in the TreeBASE database. Conclusion PhyloExplorer is a simple and interactive website implemented through underlying Python libraries and MySQL databases. It is available at: and the source code can be downloaded from: . PMID:19450253
Rigorous quantum limits on monitoring free masses and harmonic oscillators
NASA Astrophysics Data System (ADS)
Roy, S. M.
2018-03-01
There are heuristic arguments proposing that the accuracy of monitoring position of a free mass m is limited by the standard quantum limit (SQL): σ2( X (t ) ) ≥σ2( X (0 ) ) +(t2/m2) σ2( P (0 ) ) ≥ℏ t /m , where σ2( X (t ) ) and σ2( P (t ) ) denote variances of the Heisenberg representation position and momentum operators. Yuen [Phys. Rev. Lett. 51, 719 (1983), 10.1103/PhysRevLett.51.719] discovered that there are contractive states for which this result is incorrect. Here I prove universally valid rigorous quantum limits (RQL), viz. rigorous upper and lower bounds on σ2( X (t ) ) in terms of σ2( X (0 ) ) and σ2( P (0 ) ) , given by Eq. (12) for a free mass and by Eq. (36) for an oscillator. I also obtain the maximally contractive and maximally expanding states which saturate the RQL, and use the contractive states to set up an Ozawa-type measurement theory with accuracies respecting the RQL but beating the standard quantum limit. The contractive states for oscillators improve on the Schrödinger coherent states of constant variance and may be useful for gravitational wave detection and optical communication.
Zaman, Babar; Khandekar, Rajiv; Al Shahwan, Sami; Song, Jonathan; Al Jadaan, Ibrahim; Al Jiasim, Leyla; Owaydha, Ohood; Asghar, Nasira; Hijazi, Amar; Edward, Deepak P.
2014-01-01
In this brief communication, we present the steps used to establish a web-based congenital glaucoma registry at our institution. The contents of a case report form (CRF) were developed by a group of glaucoma subspecialists. Information Technology (IT) specialists used Lime Survey softwareTM to create an electronic CRF. A MY Structured Query Language (MySQL) server was used as a database with a virtual machine operating system. Two ophthalmologists and 2 IT specialists worked for 7 hours, and a biostatistician and a data registrar worked for 24 hours each to establish the electronic CRF. Using the CRF which was transferred to the Lime survey tool, and the MYSQL server application, data could be directly stored in spreadsheet programs that included Microsoft Excel, SPSS, and R-Language and queried in real-time. In a pilot test, clinical data from 80 patients with congenital glaucoma were entered into the registry and successful descriptive analysis and data entry validation was performed. A web-based disease registry was established in a short period of time in a cost-efficient manner using available resources and a team-based approach. PMID:24791112
Zaman, Babar; Khandekar, Rajiv; Al Shahwan, Sami; Song, Jonathan; Al Jadaan, Ibrahim; Al Jiasim, Leyla; Owaydha, Ohood; Asghar, Nasira; Hijazi, Amar; Edward, Deepak P
2014-01-01
In this brief communication, we present the steps used to establish a web-based congenital glaucoma registry at our institution. The contents of a case report form (CRF) were developed by a group of glaucoma subspecialists. Information Technology (IT) specialists used Lime Survey softwareTM to create an electronic CRF. A MY Structured Query Language (MySQL) server was used as a database with a virtual machine operating system. Two ophthalmologists and 2 IT specialists worked for 7 hours, and a biostatistician and a data registrar worked for 24 hours each to establish the electronic CRF. Using the CRF which was transferred to the Lime survey tool, and the MYSQL server application, data could be directly stored in spreadsheet programs that included Microsoft Excel, SPSS, and R-Language and queried in real-time. In a pilot test, clinical data from 80 patients with congenital glaucoma were entered into the registry and successful descriptive analysis and data entry validation was performed. A web-based disease registry was established in a short period of time in a cost-efficient manner using available resources and a team-based approach.
S&MPO - An information system for ozone spectroscopy on the WEB
NASA Astrophysics Data System (ADS)
Babikov, Yurii L.; Mikhailenko, Semen N.; Barbe, Alain; Tyuterev, Vladimir G.
2014-09-01
Spectroscopy and Molecular Properties of Ozone ("S&MPO") is an Internet accessible information system devoted to high resolution spectroscopy of the ozone molecule, related properties and data sources. S&MPO contains information on original spectroscopic data (line positions, line intensities, energies, transition moments, spectroscopic parameters) recovered from comprehensive analyses and modeling of experimental spectra as well as associated software for data representation written in PHP Java Script, C++ and FORTRAN. The line-by-line list of vibration-rotation transitions and other information is organized as a relational database under control of MySQL database tools. The main S&MPO goal is to provide access to all available information on vibration-rotation molecular states and transitions under extended conditions based on extrapolations of laboratory measurements using validated theoretical models. Applications for the S&MPO may include: education/training in molecular physics, radiative processes, laser physics; spectroscopic applications (analysis, Fourier transform spectroscopy, atmospheric optics, optical standards, spectroscopic atlases); applications to environment studies and atmospheric physics (remote sensing); data supply for specific databases; and to photochemistry (laser excitation, multiphoton processes). The system is accessible via Internet on two sites: http://smpo.iao.ru and http://smpo.univ-reims.fr.
Master Metadata Repository and Metadata-Management System
NASA Technical Reports Server (NTRS)
Armstrong, Edward; Reed, Nate; Zhang, Wen
2007-01-01
A master metadata repository (MMR) software system manages the storage and searching of metadata pertaining to data from national and international satellite sources of the Global Ocean Data Assimilation Experiment (GODAE) High Resolution Sea Surface Temperature Pilot Project [GHRSSTPP]. These sources produce a total of hundreds of data files daily, each file classified as one of more than ten data products representing global sea-surface temperatures. The MMR is a relational database wherein the metadata are divided into granulelevel records [denoted file records (FRs)] for individual satellite files and collection-level records [denoted data set descriptions (DSDs)] that describe metadata common to all the files from a specific data product. FRs and DSDs adhere to the NASA Directory Interchange Format (DIF). The FRs and DSDs are contained in separate subdatabases linked by a common field. The MMR is configured in MySQL database software with custom Practical Extraction and Reporting Language (PERL) programs to validate and ingest the metadata records. The database contents are converted into the Federal Geographic Data Committee (FGDC) standard format by use of the Extensible Markup Language (XML). A Web interface enables users to search for availability of data from all sources.